From 83bfae2145b9661cc684c95da040792ce6af2478 Mon Sep 17 00:00:00 2001 From: Azure SDK for Python bot Date: Wed, 28 Mar 2018 09:46:54 -0700 Subject: [PATCH 01/54] Generated from 2e36516bed7132327c0e1664747aa438ad744797 (#2292) [ADF] Swagger Change for Responsys Source --- .../azure/mgmt/datafactory/models/__init__.py | 965 ++++++++++++------ .../azure/mgmt/datafactory/models/activity.py | 18 +- .../datafactory/models/activity_dependency.py | 17 +- .../models/activity_dependency_py3.py | 46 + .../datafactory/models/activity_policy.py | 14 +- .../datafactory/models/activity_policy_py3.py | 54 + .../mgmt/datafactory/models/activity_py3.py | 59 ++ .../mgmt/datafactory/models/activity_run.py | 6 +- .../datafactory/models/activity_run_py3.py | 102 ++ .../models/amazon_mws_linked_service.py | 40 +- .../models/amazon_mws_linked_service_py3.py | 106 ++ .../models/amazon_mws_object_dataset.py | 20 +- .../models/amazon_mws_object_dataset_py3.py | 58 ++ .../datafactory/models/amazon_mws_source.py | 10 +- .../models/amazon_mws_source_py3.py | 52 + .../models/amazon_redshift_linked_service.py | 28 +- .../amazon_redshift_linked_service_py3.py | 86 ++ .../models/amazon_redshift_source.py | 12 +- .../models/amazon_redshift_source_py3.py | 60 ++ .../datafactory/models/amazon_s3_dataset.py | 26 +- .../models/amazon_s3_dataset_py3.py | 88 ++ .../models/amazon_s3_linked_service.py | 14 +- .../models/amazon_s3_linked_service_py3.py | 70 ++ .../mgmt/datafactory/models/avro_format.py | 15 +- .../datafactory/models/avro_format_py3.py | 46 + .../models/azure_batch_linked_service.py | 35 +- .../models/azure_batch_linked_service_py3.py | 88 ++ .../datafactory/models/azure_blob_dataset.py | 20 +- .../models/azure_blob_dataset_py3.py | 81 ++ ...zure_data_lake_analytics_linked_service.py | 32 +- ..._data_lake_analytics_linked_service_py3.py | 99 ++ .../models/azure_data_lake_store_dataset.py | 22 +- .../azure_data_lake_store_dataset_py3.py | 78 ++ .../azure_data_lake_store_linked_service.py | 28 +- ...zure_data_lake_store_linked_service_py3.py | 98 ++ .../models/azure_data_lake_store_sink.py | 10 +- .../models/azure_data_lake_store_sink_py3.py | 62 ++ .../models/azure_data_lake_store_source.py | 10 +- .../azure_data_lake_store_source_py3.py | 53 + .../models/azure_databricks_linked_service.py | 30 +- .../azure_databricks_linked_service_py3.py | 101 ++ .../models/azure_key_vault_linked_service.py | 12 +- .../azure_key_vault_linked_service_py3.py | 60 ++ .../azure_key_vault_secret_reference.py | 20 +- .../azure_key_vault_secret_reference_py3.py | 51 + .../azure_ml_batch_execution_activity.py | 16 +- .../azure_ml_batch_execution_activity_py3.py | 79 ++ .../models/azure_ml_linked_service.py | 29 +- .../models/azure_ml_linked_service_py3.py | 94 ++ .../azure_ml_update_resource_activity.py | 28 +- .../azure_ml_update_resource_activity_py3.py | 78 ++ .../models/azure_ml_web_service_file.py | 20 +- .../models/azure_ml_web_service_file_py3.py | 43 + .../models/azure_my_sql_linked_service.py | 14 +- .../models/azure_my_sql_linked_service_py3.py | 64 ++ .../datafactory/models/azure_my_sql_source.py | 10 +- .../models/azure_my_sql_source_py3.py | 52 + .../models/azure_my_sql_table_dataset.py | 12 +- .../models/azure_my_sql_table_dataset_py3.py | 63 ++ .../azure_postgre_sql_linked_service.py | 12 +- .../azure_postgre_sql_linked_service_py3.py | 63 ++ .../models/azure_postgre_sql_source.py | 10 +- .../models/azure_postgre_sql_source_py3.py | 52 + .../models/azure_postgre_sql_table_dataset.py | 20 +- .../azure_postgre_sql_table_dataset_py3.py | 58 ++ .../datafactory/models/azure_queue_sink.py | 17 +- .../models/azure_queue_sink_py3.py | 56 + .../models/azure_search_index_dataset.py | 16 +- .../models/azure_search_index_dataset_py3.py | 64 ++ .../models/azure_search_index_sink.py | 10 +- .../models/azure_search_index_sink_py3.py | 62 ++ .../models/azure_search_linked_service.py | 18 +- .../models/azure_search_linked_service_py3.py | 69 ++ .../azure_sql_database_linked_service.py | 20 +- .../azure_sql_database_linked_service_py3.py | 80 ++ .../models/azure_sql_dw_linked_service.py | 20 +- .../models/azure_sql_dw_linked_service_py3.py | 80 ++ .../models/azure_sql_dw_table_dataset.py | 16 +- .../models/azure_sql_dw_table_dataset_py3.py | 64 ++ .../models/azure_sql_table_dataset.py | 16 +- .../models/azure_sql_table_dataset_py3.py | 64 ++ .../models/azure_storage_linked_service.py | 14 +- .../azure_storage_linked_service_py3.py | 69 ++ .../datafactory/models/azure_table_dataset.py | 16 +- .../models/azure_table_dataset_py3.py | 64 ++ .../datafactory/models/azure_table_sink.py | 16 +- .../models/azure_table_sink_py3.py | 76 ++ .../datafactory/models/azure_table_source.py | 12 +- .../models/azure_table_source_py3.py | 58 ++ .../mgmt/datafactory/models/blob_sink.py | 16 +- .../mgmt/datafactory/models/blob_sink_py3.py | 77 ++ .../mgmt/datafactory/models/blob_source.py | 14 +- .../datafactory/models/blob_source_py3.py | 63 ++ .../mgmt/datafactory/models/blob_trigger.py | 25 +- .../datafactory/models/blob_trigger_py3.py | 74 ++ .../models/cassandra_linked_service.py | 24 +- .../models/cassandra_linked_service_py3.py | 84 ++ .../datafactory/models/cassandra_source.py | 12 +- .../models/cassandra_source_py3.py | 65 ++ .../models/cassandra_table_dataset.py | 14 +- .../models/cassandra_table_dataset_py3.py | 68 ++ .../models/concur_linked_service.py | 28 +- .../models/concur_linked_service_py3.py | 92 ++ .../models/concur_object_dataset.py | 20 +- .../models/concur_object_dataset_py3.py | 58 ++ .../mgmt/datafactory/models/concur_source.py | 10 +- .../datafactory/models/concur_source_py3.py | 52 + .../datafactory/models/control_activity.py | 18 +- .../models/control_activity_py3.py | 56 + .../mgmt/datafactory/models/copy_activity.py | 36 +- .../datafactory/models/copy_activity_py3.py | 113 ++ .../mgmt/datafactory/models/copy_sink.py | 18 +- .../mgmt/datafactory/models/copy_sink_py3.py | 71 ++ .../mgmt/datafactory/models/copy_source.py | 41 +- .../datafactory/models/copy_source_py3.py | 69 ++ .../datafactory/models/copy_translator.py | 10 +- .../datafactory/models/copy_translator_py3.py | 46 + .../models/cosmos_db_linked_service.py | 14 +- .../models/cosmos_db_linked_service_py3.py | 64 ++ .../models/couchbase_linked_service.py | 12 +- .../models/couchbase_linked_service_py3.py | 63 ++ .../datafactory/models/couchbase_source.py | 10 +- .../models/couchbase_source_py3.py | 52 + .../models/couchbase_table_dataset.py | 20 +- .../models/couchbase_table_dataset_py3.py | 58 ++ .../datafactory/models/create_run_response.py | 10 +- .../models/create_run_response_py3.py | 34 + .../datafactory/models/custom_activity.py | 24 +- .../datafactory/models/custom_activity_py3.py | 83 ++ .../custom_activity_reference_object.py | 8 +- .../custom_activity_reference_object_py3.py | 33 + .../custom_data_source_linked_service.py | 12 +- .../custom_data_source_linked_service_py3.py | 58 ++ .../mgmt/datafactory/models/custom_dataset.py | 14 +- .../datafactory/models/custom_dataset_py3.py | 63 ++ .../data_factory_management_client_enums.py | 114 +-- .../data_lake_analytics_usql_activity.py | 30 +- .../data_lake_analytics_usql_activity_py3.py | 95 ++ .../models/databricks_notebook_activity.py | 20 +- .../databricks_notebook_activity_py3.py | 68 ++ .../azure/mgmt/datafactory/models/dataset.py | 49 +- .../models/dataset_bzip2_compression.py | 13 +- .../models/dataset_bzip2_compression_py3.py | 38 + .../datafactory/models/dataset_compression.py | 10 +- .../models/dataset_compression_py3.py | 47 + .../models/dataset_deflate_compression.py | 10 +- .../models/dataset_deflate_compression_py3.py | 43 + .../models/dataset_gzip_compression.py | 10 +- .../models/dataset_gzip_compression_py3.py | 43 + .../mgmt/datafactory/models/dataset_py3.py | 91 ++ .../datafactory/models/dataset_reference.py | 15 +- .../models/dataset_reference_py3.py | 48 + .../datafactory/models/dataset_resource.py | 10 +- .../models/dataset_resource_py3.py | 53 + .../models/dataset_storage_format.py | 14 +- .../models/dataset_storage_format_py3.py | 57 ++ .../models/dataset_zip_deflate_compression.py | 10 +- .../dataset_zip_deflate_compression_py3.py | 43 + .../datafactory/models/db2_linked_service.py | 30 +- .../models/db2_linked_service_py3.py | 91 ++ .../datafactory/models/distcp_settings.py | 25 +- .../datafactory/models/distcp_settings_py3.py | 49 + .../models/document_db_collection_dataset.py | 16 +- .../document_db_collection_dataset_py3.py | 64 ++ .../models/document_db_collection_sink.py | 10 +- .../models/document_db_collection_sink_py3.py | 61 ++ .../models/document_db_collection_source.py | 12 +- .../document_db_collection_source_py3.py | 57 ++ .../models/drill_linked_service.py | 12 +- .../models/drill_linked_service_py3.py | 63 ++ .../mgmt/datafactory/models/drill_source.py | 10 +- .../datafactory/models/drill_source_py3.py | 52 + .../datafactory/models/drill_table_dataset.py | 20 +- .../models/drill_table_dataset_py3.py | 58 ++ .../models/dynamics_entity_dataset.py | 12 +- .../models/dynamics_entity_dataset_py3.py | 63 ++ .../models/dynamics_linked_service.py | 43 +- .../models/dynamics_linked_service_py3.py | 109 ++ .../mgmt/datafactory/models/dynamics_sink.py | 14 +- .../datafactory/models/dynamics_sink_py3.py | 72 ++ .../datafactory/models/dynamics_source.py | 10 +- .../datafactory/models/dynamics_source_py3.py | 53 + .../models/eloqua_linked_service.py | 28 +- .../models/eloqua_linked_service_py3.py | 91 ++ .../models/eloqua_object_dataset.py | 20 +- .../models/eloqua_object_dataset_py3.py | 58 ++ .../mgmt/datafactory/models/eloqua_source.py | 10 +- .../datafactory/models/eloqua_source_py3.py | 52 + .../mgmt/datafactory/models/error_response.py | 18 +- .../datafactory/models/error_response_py3.py | 60 ++ .../models/execute_pipeline_activity.py | 18 +- .../models/execute_pipeline_activity_py3.py | 62 ++ .../models/execute_ssis_package_activity.py | 24 +- .../execute_ssis_package_activity_py3.py | 80 ++ .../datafactory/models/execution_activity.py | 14 +- .../models/execution_activity_py3.py | 70 ++ .../mgmt/datafactory/models/expression.py | 12 +- .../mgmt/datafactory/models/expression_py3.py | 43 + .../azure/mgmt/datafactory/models/factory.py | 8 +- .../datafactory/models/factory_identity.py | 10 +- .../models/factory_identity_py3.py | 49 + .../mgmt/datafactory/models/factory_py3.py | 72 ++ .../models/factory_update_parameters.py | 8 +- .../models/factory_update_parameters_py3.py | 32 + .../models/file_server_linked_service.py | 20 +- .../models/file_server_linked_service_py3.py | 74 ++ .../datafactory/models/file_share_dataset.py | 20 +- .../models/file_share_dataset_py3.py | 82 ++ .../datafactory/models/file_system_sink.py | 10 +- .../models/file_system_sink_py3.py | 62 ++ .../datafactory/models/file_system_source.py | 10 +- .../models/file_system_source_py3.py | 53 + .../datafactory/models/filter_activity.py | 18 +- .../datafactory/models/filter_activity_py3.py | 58 ++ .../datafactory/models/for_each_activity.py | 22 +- .../models/for_each_activity_py3.py | 70 ++ .../models/ftp_server_linked_service.py | 28 +- .../models/ftp_server_linked_service_py3.py | 98 ++ .../models/get_metadata_activity.py | 16 +- .../models/get_metadata_activity_py3.py | 64 ++ .../models/google_big_query_linked_service.py | 41 +- .../google_big_query_linked_service_py3.py | 124 +++ .../models/google_big_query_object_dataset.py | 20 +- .../google_big_query_object_dataset_py3.py | 58 ++ .../models/google_big_query_source.py | 10 +- .../models/google_big_query_source_py3.py | 52 + .../models/greenplum_linked_service.py | 12 +- .../models/greenplum_linked_service_py3.py | 63 ++ .../datafactory/models/greenplum_source.py | 10 +- .../models/greenplum_source_py3.py | 52 + .../models/greenplum_table_dataset.py | 20 +- .../models/greenplum_table_dataset_py3.py | 58 ++ .../models/hbase_linked_service.py | 39 +- .../models/hbase_linked_service_py3.py | 114 +++ .../models/hbase_object_dataset.py | 20 +- .../models/hbase_object_dataset_py3.py | 58 ++ .../mgmt/datafactory/models/hbase_source.py | 10 +- .../datafactory/models/hbase_source_py3.py | 52 + .../models/hd_insight_hive_activity.py | 22 +- .../models/hd_insight_hive_activity_py3.py | 84 ++ .../models/hd_insight_linked_service.py | 24 +- .../models/hd_insight_linked_service_py3.py | 85 ++ .../models/hd_insight_map_reduce_activity.py | 32 +- .../hd_insight_map_reduce_activity_py3.py | 96 ++ .../hd_insight_on_demand_linked_service.py | 100 +- ...hd_insight_on_demand_linked_service_py3.py | 219 ++++ .../models/hd_insight_pig_activity.py | 22 +- .../models/hd_insight_pig_activity_py3.py | 84 ++ .../models/hd_insight_spark_activity.py | 36 +- .../models/hd_insight_spark_activity_py3.py | 97 ++ .../models/hd_insight_streaming_activity.py | 53 +- .../hd_insight_streaming_activity_py3.py | 119 +++ .../datafactory/models/hdfs_linked_service.py | 20 +- .../models/hdfs_linked_service_py3.py | 81 ++ .../mgmt/datafactory/models/hdfs_source.py | 12 +- .../datafactory/models/hdfs_source_py3.py | 57 ++ .../datafactory/models/hive_linked_service.py | 51 +- .../models/hive_linked_service_py3.py | 147 +++ .../datafactory/models/hive_object_dataset.py | 20 +- .../models/hive_object_dataset_py3.py | 58 ++ .../mgmt/datafactory/models/hive_source.py | 10 +- .../datafactory/models/hive_source_py3.py | 52 + .../mgmt/datafactory/models/http_dataset.py | 22 +- .../datafactory/models/http_dataset_py3.py | 90 ++ .../datafactory/models/http_linked_service.py | 26 +- .../models/http_linked_service_py3.py | 105 ++ .../mgmt/datafactory/models/http_source.py | 10 +- .../datafactory/models/http_source_py3.py | 55 + .../models/hubspot_linked_service.py | 27 +- .../models/hubspot_linked_service_py3.py | 96 ++ .../models/hubspot_object_dataset.py | 20 +- .../models/hubspot_object_dataset_py3.py | 58 ++ .../mgmt/datafactory/models/hubspot_source.py | 10 +- .../datafactory/models/hubspot_source_py3.py | 52 + .../models/if_condition_activity.py | 20 +- .../models/if_condition_activity_py3.py | 69 ++ .../models/impala_linked_service.py | 39 +- .../models/impala_linked_service_py3.py | 117 +++ .../models/impala_object_dataset.py | 20 +- .../models/impala_object_dataset_py3.py | 58 ++ .../mgmt/datafactory/models/impala_source.py | 10 +- .../datafactory/models/impala_source_py3.py | 52 + .../datafactory/models/integration_runtime.py | 12 +- .../models/integration_runtime_auth_keys.py | 8 +- .../integration_runtime_auth_keys_py3.py | 32 + .../integration_runtime_compute_properties.py | 16 +- ...egration_runtime_compute_properties_py3.py | 60 ++ .../integration_runtime_connection_info.py | 6 +- ...integration_runtime_connection_info_py3.py | 70 ++ ..._runtime_custom_setup_script_properties.py | 8 +- ...time_custom_setup_script_properties_py3.py | 33 + .../integration_runtime_monitoring_data.py | 8 +- ...integration_runtime_monitoring_data_py3.py | 33 + .../integration_runtime_node_ip_address.py | 4 +- ...integration_runtime_node_ip_address_py3.py | 35 + ...ntegration_runtime_node_monitoring_data.py | 6 +- ...ration_runtime_node_monitoring_data_py3.py | 79 ++ .../models/integration_runtime_py3.py | 51 + .../models/integration_runtime_reference.py | 14 +- .../integration_runtime_reference_py3.py | 48 + ...ation_runtime_regenerate_key_parameters.py | 6 +- ...n_runtime_regenerate_key_parameters_py3.py | 30 + ...integration_runtime_remove_node_request.py | 8 +- ...gration_runtime_remove_node_request_py3.py | 33 + .../models/integration_runtime_resource.py | 10 +- .../integration_runtime_resource_py3.py | 53 + .../integration_runtime_ssis_catalog_info.py | 14 +- ...tegration_runtime_ssis_catalog_info_py3.py | 55 + .../integration_runtime_ssis_properties.py | 14 +- ...integration_runtime_ssis_properties_py3.py | 53 + .../models/integration_runtime_status.py | 10 +- ...ntegration_runtime_status_list_response.py | 12 +- ...ration_runtime_status_list_response_py3.py | 40 + .../models/integration_runtime_status_py3.py | 64 ++ .../integration_runtime_status_response.py | 10 +- ...integration_runtime_status_response_py3.py | 42 + .../integration_runtime_vnet_properties.py | 10 +- ...integration_runtime_vnet_properties_py3.py | 38 + .../datafactory/models/jira_linked_service.py | 31 +- .../models/jira_linked_service_py3.py | 98 ++ .../datafactory/models/jira_object_dataset.py | 20 +- .../models/jira_object_dataset_py3.py | 58 ++ .../mgmt/datafactory/models/jira_source.py | 10 +- .../datafactory/models/jira_source_py3.py | 52 + .../mgmt/datafactory/models/json_format.py | 18 +- .../datafactory/models/json_format_py3.py | 84 ++ .../models/linked_integration_runtime.py | 4 +- .../models/linked_integration_runtime_key.py | 12 +- .../linked_integration_runtime_key_py3.py | 39 + .../linked_integration_runtime_properties.py | 8 +- ...nked_integration_runtime_properties_py3.py | 41 + .../models/linked_integration_runtime_py3.py | 58 ++ .../models/linked_integration_runtime_rbac.py | 14 +- .../linked_integration_runtime_rbac_py3.py | 40 + .../mgmt/datafactory/models/linked_service.py | 22 +- .../datafactory/models/linked_service_py3.py | 92 ++ .../models/linked_service_reference.py | 14 +- .../models/linked_service_reference_py3.py | 48 + .../models/linked_service_resource.py | 10 +- .../models/linked_service_resource_py3.py | 53 + .../datafactory/models/lookup_activity.py | 22 +- .../datafactory/models/lookup_activity_py3.py | 71 ++ .../models/magento_linked_service.py | 22 +- .../models/magento_linked_service_py3.py | 85 ++ .../models/magento_object_dataset.py | 20 +- .../models/magento_object_dataset_py3.py | 58 ++ .../mgmt/datafactory/models/magento_source.py | 10 +- .../datafactory/models/magento_source_py3.py | 52 + .../models/managed_integration_runtime.py | 12 +- .../managed_integration_runtime_error.py | 6 +- .../managed_integration_runtime_error_py3.py | 55 + .../managed_integration_runtime_node.py | 8 +- .../managed_integration_runtime_node_py3.py | 52 + ...ed_integration_runtime_operation_result.py | 6 +- ...ntegration_runtime_operation_result_py3.py | 65 ++ .../models/managed_integration_runtime_py3.py | 65 ++ .../managed_integration_runtime_status.py | 8 +- .../managed_integration_runtime_status_py3.py | 78 ++ .../models/maria_db_linked_service.py | 12 +- .../models/maria_db_linked_service_py3.py | 63 ++ .../datafactory/models/maria_db_source.py | 10 +- .../datafactory/models/maria_db_source_py3.py | 52 + .../models/maria_db_table_dataset.py | 20 +- .../models/maria_db_table_dataset_py3.py | 58 ++ .../models/marketo_linked_service.py | 26 +- .../models/marketo_linked_service_py3.py | 90 ++ .../models/marketo_object_dataset.py | 20 +- .../models/marketo_object_dataset_py3.py | 58 ++ .../mgmt/datafactory/models/marketo_source.py | 10 +- .../datafactory/models/marketo_source_py3.py | 52 + .../models/mongo_db_collection_dataset.py | 16 +- .../models/mongo_db_collection_dataset_py3.py | 64 ++ .../models/mongo_db_linked_service.py | 36 +- .../models/mongo_db_linked_service_py3.py | 109 ++ .../datafactory/models/mongo_db_source.py | 10 +- .../datafactory/models/mongo_db_source_py3.py | 52 + .../models/multiple_pipeline_trigger.py | 10 +- .../models/multiple_pipeline_trigger_py3.py | 64 ++ .../models/my_sql_linked_service.py | 28 +- .../models/my_sql_linked_service_py3.py | 85 ++ .../models/netezza_linked_service.py | 12 +- .../models/netezza_linked_service_py3.py | 63 ++ .../mgmt/datafactory/models/netezza_source.py | 10 +- .../datafactory/models/netezza_source_py3.py | 52 + .../models/netezza_table_dataset.py | 20 +- .../models/netezza_table_dataset_py3.py | 58 ++ .../models/odata_linked_service.py | 22 +- .../models/odata_linked_service_py3.py | 80 ++ .../models/odata_resource_dataset.py | 12 +- .../models/odata_resource_dataset_py3.py | 63 ++ .../datafactory/models/odbc_linked_service.py | 24 +- .../models/odbc_linked_service_py3.py | 85 ++ .../mgmt/datafactory/models/odbc_sink.py | 10 +- .../mgmt/datafactory/models/odbc_sink_py3.py | 61 ++ .../mgmt/datafactory/models/operation.py | 12 +- .../datafactory/models/operation_display.py | 12 +- .../models/operation_display_py3.py | 41 + .../models/operation_list_response.py | 8 +- .../models/operation_list_response_py3.py | 34 + .../models/operation_log_specification.py | 10 +- .../models/operation_log_specification_py3.py | 37 + .../models/operation_metric_availability.py | 8 +- .../operation_metric_availability_py3.py | 33 + .../models/operation_metric_specification.py | 22 +- .../operation_metric_specification_py3.py | 63 ++ .../mgmt/datafactory/models/operation_py3.py | 41 + .../models/operation_service_specification.py | 8 +- .../operation_service_specification_py3.py | 34 + .../models/oracle_linked_service.py | 14 +- .../models/oracle_linked_service_py3.py | 64 ++ .../mgmt/datafactory/models/oracle_sink.py | 10 +- .../datafactory/models/oracle_sink_py3.py | 61 ++ .../mgmt/datafactory/models/oracle_source.py | 12 +- .../datafactory/models/oracle_source_py3.py | 58 ++ .../models/oracle_table_dataset.py | 16 +- .../models/oracle_table_dataset_py3.py | 64 ++ .../mgmt/datafactory/models/orc_format.py | 15 +- .../mgmt/datafactory/models/orc_format_py3.py | 46 + .../models/parameter_specification.py | 14 +- .../models/parameter_specification_py3.py | 39 + .../mgmt/datafactory/models/parquet_format.py | 15 +- .../datafactory/models/parquet_format_py3.py | 46 + .../models/paypal_linked_service.py | 28 +- .../models/paypal_linked_service_py3.py | 92 ++ .../models/paypal_object_dataset.py | 20 +- .../models/paypal_object_dataset_py3.py | 58 ++ .../mgmt/datafactory/models/paypal_source.py | 10 +- .../datafactory/models/paypal_source_py3.py | 52 + .../models/phoenix_linked_service.py | 40 +- .../models/phoenix_linked_service_py3.py | 121 +++ .../models/phoenix_object_dataset.py | 20 +- .../models/phoenix_object_dataset_py3.py | 58 ++ .../mgmt/datafactory/models/phoenix_source.py | 10 +- .../datafactory/models/phoenix_source_py3.py | 52 + .../datafactory/models/pipeline_reference.py | 15 +- .../models/pipeline_reference_py3.py | 48 + .../datafactory/models/pipeline_resource.py | 16 +- .../models/pipeline_resource_py3.py | 74 ++ .../mgmt/datafactory/models/pipeline_run.py | 6 +- .../models/pipeline_run_filter_parameters.py | 24 +- .../pipeline_run_filter_parameters_py3.py | 55 + .../models/pipeline_run_invoked_by.py | 4 +- .../models/pipeline_run_invoked_by_py3.py | 40 + .../datafactory/models/pipeline_run_py3.py | 87 ++ .../models/pipeline_run_query_filter.py | 22 +- .../models/pipeline_run_query_filter_py3.py | 48 + .../models/pipeline_run_query_order_by.py | 18 +- .../models/pipeline_run_query_order_by_py3.py | 42 + .../models/pipeline_run_query_response.py | 12 +- .../models/pipeline_run_query_response_py3.py | 39 + .../datafactory/models/polybase_settings.py | 14 +- .../models/polybase_settings_py3.py | 53 + .../models/postgre_sql_linked_service.py | 28 +- .../models/postgre_sql_linked_service_py3.py | 85 ++ .../models/presto_linked_service.py | 50 +- .../models/presto_linked_service_py3.py | 132 +++ .../models/presto_object_dataset.py | 20 +- .../models/presto_object_dataset_py3.py | 58 ++ .../mgmt/datafactory/models/presto_source.py | 10 +- .../datafactory/models/presto_source_py3.py | 52 + .../models/quick_books_linked_service.py | 25 +- .../models/quick_books_linked_service_py3.py | 85 ++ .../models/quick_books_object_dataset.py | 20 +- .../models/quick_books_object_dataset_py3.py | 58 ++ .../datafactory/models/quick_books_source.py | 10 +- .../models/quick_books_source_py3.py | 52 + .../datafactory/models/recurrence_schedule.py | 16 +- .../models/recurrence_schedule_occurrence.py | 10 +- .../recurrence_schedule_occurrence_py3.py | 38 + .../models/recurrence_schedule_py3.py | 50 + .../redirect_incompatible_row_settings.py | 20 +- .../redirect_incompatible_row_settings_py3.py | 47 + .../models/redshift_unload_settings.py | 24 +- .../models/redshift_unload_settings_py3.py | 48 + .../datafactory/models/relational_source.py | 10 +- .../models/relational_source_py3.py | 52 + .../models/relational_table_dataset.py | 12 +- .../models/relational_table_dataset_py3.py | 63 ++ .../azure/mgmt/datafactory/models/resource.py | 8 +- .../mgmt/datafactory/models/resource_py3.py | 53 + .../models/responsys_linked_service.py | 94 ++ .../models/responsys_linked_service_py3.py | 94 ++ .../models/responsys_object_dataset.py | 58 ++ .../models/responsys_object_dataset_py3.py | 58 ++ .../datafactory/models/responsys_source.py | 52 + .../models/responsys_source_py3.py | 52 + .../mgmt/datafactory/models/retry_policy.py | 8 +- .../datafactory/models/retry_policy_py3.py | 38 + .../models/salesforce_linked_service.py | 18 +- .../models/salesforce_linked_service_py3.py | 82 ++ ...lesforce_marketing_cloud_linked_service.py | 25 +- ...orce_marketing_cloud_linked_service_py3.py | 91 ++ ...lesforce_marketing_cloud_object_dataset.py | 20 +- ...orce_marketing_cloud_object_dataset_py3.py | 58 ++ .../salesforce_marketing_cloud_source.py | 10 +- .../salesforce_marketing_cloud_source_py3.py | 52 + .../models/salesforce_object_dataset.py | 12 +- .../models/salesforce_object_dataset_py3.py | 63 ++ .../datafactory/models/salesforce_sink.py | 14 +- .../datafactory/models/salesforce_sink_py3.py | 79 ++ .../datafactory/models/salesforce_source.py | 12 +- .../models/salesforce_source_py3.py | 58 ++ .../models/sap_bw_linked_service.py | 36 +- .../models/sap_bw_linked_service_py3.py | 88 ++ .../sap_cloud_for_customer_linked_service.py | 22 +- ...p_cloud_for_customer_linked_service_py3.py | 76 ++ ...sap_cloud_for_customer_resource_dataset.py | 16 +- ...cloud_for_customer_resource_dataset_py3.py | 64 ++ .../models/sap_cloud_for_customer_sink.py | 10 +- .../models/sap_cloud_for_customer_sink_py3.py | 62 ++ .../models/sap_cloud_for_customer_source.py | 10 +- .../sap_cloud_for_customer_source_py3.py | 52 + .../models/sap_ecc_linked_service.py | 18 +- .../models/sap_ecc_linked_service_py3.py | 76 ++ .../models/sap_ecc_resource_dataset.py | 16 +- .../models/sap_ecc_resource_dataset_py3.py | 64 ++ .../mgmt/datafactory/models/sap_ecc_source.py | 10 +- .../datafactory/models/sap_ecc_source_py3.py | 52 + .../models/sap_hana_linked_service.py | 22 +- .../models/sap_hana_linked_service_py3.py | 80 ++ .../datafactory/models/schedule_trigger.py | 12 +- .../models/schedule_trigger_py3.py | 60 ++ .../models/schedule_trigger_recurrence.py | 18 +- .../models/schedule_trigger_recurrence_py3.py | 54 + .../mgmt/datafactory/models/secret_base.py | 8 +- .../datafactory/models/secret_base_py3.py | 41 + .../mgmt/datafactory/models/secure_string.py | 12 +- .../datafactory/models/secure_string_py3.py | 40 + .../models/self_hosted_integration_runtime.py | 10 +- .../self_hosted_integration_runtime_node.py | 6 +- ...elf_hosted_integration_runtime_node_py3.py | 139 +++ .../self_hosted_integration_runtime_py3.py | 46 + .../self_hosted_integration_runtime_status.py | 12 +- ...f_hosted_integration_runtime_status_py3.py | 129 +++ .../models/service_now_linked_service.py | 34 +- .../models/service_now_linked_service_py3.py | 106 ++ .../models/service_now_object_dataset.py | 20 +- .../models/service_now_object_dataset_py3.py | 58 ++ .../datafactory/models/service_now_source.py | 10 +- .../models/service_now_source_py3.py | 52 + .../models/sftp_server_linked_service.py | 34 +- .../models/sftp_server_linked_service_py3.py | 119 +++ .../models/shopify_linked_service.py | 22 +- .../models/shopify_linked_service_py3.py | 86 ++ .../models/shopify_object_dataset.py | 20 +- .../models/shopify_object_dataset_py3.py | 58 ++ .../mgmt/datafactory/models/shopify_source.py | 10 +- .../datafactory/models/shopify_source_py3.py | 52 + .../models/spark_linked_service.py | 46 +- .../models/spark_linked_service_py3.py | 131 +++ .../models/spark_object_dataset.py | 20 +- .../models/spark_object_dataset_py3.py | 58 ++ .../mgmt/datafactory/models/spark_source.py | 10 +- .../datafactory/models/spark_source_py3.py | 52 + .../mgmt/datafactory/models/sql_dw_sink.py | 14 +- .../datafactory/models/sql_dw_sink_py3.py | 72 ++ .../mgmt/datafactory/models/sql_dw_source.py | 14 +- .../datafactory/models/sql_dw_source_py3.py | 65 ++ .../models/sql_server_linked_service.py | 18 +- .../models/sql_server_linked_service_py3.py | 73 ++ .../sql_server_stored_procedure_activity.py | 18 +- ...ql_server_stored_procedure_activity_py3.py | 67 ++ .../models/sql_server_table_dataset.py | 16 +- .../models/sql_server_table_dataset_py3.py | 64 ++ .../azure/mgmt/datafactory/models/sql_sink.py | 16 +- .../mgmt/datafactory/models/sql_sink_py3.py | 76 ++ .../mgmt/datafactory/models/sql_source.py | 14 +- .../mgmt/datafactory/models/sql_source_py3.py | 64 ++ .../models/square_linked_service.py | 34 +- .../models/square_linked_service_py3.py | 98 ++ .../models/square_object_dataset.py | 20 +- .../models/square_object_dataset_py3.py | 58 ++ .../mgmt/datafactory/models/square_source.py | 10 +- .../datafactory/models/square_source_py3.py | 52 + .../models/ssis_package_location.py | 10 +- .../models/ssis_package_location_py3.py | 34 + .../datafactory/models/staging_settings.py | 16 +- .../models/staging_settings_py3.py | 51 + .../models/stored_procedure_parameter.py | 12 +- .../models/stored_procedure_parameter_py3.py | 41 + .../mgmt/datafactory/models/sub_resource.py | 4 +- .../datafactory/models/sub_resource_py3.py | 50 + .../models/sybase_linked_service.py | 30 +- .../models/sybase_linked_service_py3.py | 91 ++ .../datafactory/models/tabular_translator.py | 10 +- .../models/tabular_translator_py3.py | 43 + .../models/teradata_linked_service.py | 24 +- .../models/teradata_linked_service_py3.py | 85 ++ .../mgmt/datafactory/models/text_format.py | 26 +- .../datafactory/models/text_format_py3.py | 99 ++ .../azure/mgmt/datafactory/models/trigger.py | 12 +- .../models/trigger_pipeline_reference.py | 8 +- .../models/trigger_pipeline_reference_py3.py | 32 + .../mgmt/datafactory/models/trigger_py3.py | 62 ++ .../datafactory/models/trigger_resource.py | 10 +- .../models/trigger_resource_py3.py | 53 + .../mgmt/datafactory/models/trigger_run.py | 6 +- .../datafactory/models/trigger_run_py3.py | 78 ++ .../models/tumbling_window_trigger.py | 46 +- .../models/tumbling_window_trigger_py3.py | 103 ++ .../mgmt/datafactory/models/until_activity.py | 22 +- .../datafactory/models/until_activity_py3.py | 69 ++ ...update_integration_runtime_node_request.py | 6 +- ...te_integration_runtime_node_request_py3.py | 34 + .../update_integration_runtime_request.py | 8 +- .../update_integration_runtime_request_py3.py | 38 + .../models/vertica_linked_service.py | 12 +- .../models/vertica_linked_service_py3.py | 63 ++ .../mgmt/datafactory/models/vertica_source.py | 10 +- .../datafactory/models/vertica_source_py3.py | 52 + .../models/vertica_table_dataset.py | 20 +- .../models/vertica_table_dataset_py3.py | 58 ++ .../mgmt/datafactory/models/wait_activity.py | 14 +- .../datafactory/models/wait_activity_py3.py | 53 + .../mgmt/datafactory/models/web_activity.py | 32 +- .../models/web_activity_authentication.py | 19 +- .../models/web_activity_authentication_py3.py | 53 + .../datafactory/models/web_activity_py3.py | 95 ++ .../models/web_anonymous_authentication.py | 15 +- .../web_anonymous_authentication_py3.py | 41 + .../models/web_basic_authentication.py | 20 +- .../models/web_basic_authentication_py3.py | 52 + .../web_client_certificate_authentication.py | 18 +- ...b_client_certificate_authentication_py3.py | 53 + .../datafactory/models/web_linked_service.py | 12 +- .../models/web_linked_service_py3.py | 59 ++ .../web_linked_service_type_properties.py | 12 +- .../web_linked_service_type_properties_py3.py | 50 + .../mgmt/datafactory/models/web_source.py | 15 +- .../mgmt/datafactory/models/web_source_py3.py | 47 + .../datafactory/models/web_table_dataset.py | 18 +- .../models/web_table_dataset_py3.py | 69 ++ .../datafactory/models/xero_linked_service.py | 25 +- .../models/xero_linked_service_py3.py | 93 ++ .../datafactory/models/xero_object_dataset.py | 20 +- .../models/xero_object_dataset_py3.py | 58 ++ .../mgmt/datafactory/models/xero_source.py | 10 +- .../datafactory/models/xero_source_py3.py | 52 + .../datafactory/models/zoho_linked_service.py | 22 +- .../models/zoho_linked_service_py3.py | 85 ++ .../datafactory/models/zoho_object_dataset.py | 20 +- .../models/zoho_object_dataset_py3.py | 58 ++ .../mgmt/datafactory/models/zoho_source.py | 10 +- .../datafactory/models/zoho_source_py3.py | 52 + .../integration_runtimes_operations.py | 92 +- .../operations/triggers_operations.py | 90 +- 646 files changed, 25108 insertions(+), 2791 deletions(-) create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/activity_dependency_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/activity_policy_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/activity_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/activity_run_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_linked_service_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_object_dataset_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_source_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_redshift_linked_service_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_redshift_source_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_s3_dataset_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_s3_linked_service_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/avro_format_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_batch_linked_service_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_dataset_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_analytics_linked_service_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_dataset_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_linked_service_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_sink_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_source_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_databricks_linked_service_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_key_vault_linked_service_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_key_vault_secret_reference_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_batch_execution_activity_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_linked_service_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_update_resource_activity_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_web_service_file_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_linked_service_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_source_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_table_dataset_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_linked_service_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_source_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_table_dataset_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_queue_sink_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_search_index_dataset_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_search_index_sink_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_search_linked_service_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_database_linked_service_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_dw_linked_service_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_dw_table_dataset_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_table_dataset_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_storage_linked_service_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_table_dataset_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_table_sink_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_table_source_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/blob_sink_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/blob_source_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/blob_trigger_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/cassandra_linked_service_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/cassandra_source_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/cassandra_table_dataset_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_linked_service_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_object_dataset_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_source_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/control_activity_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_activity_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_sink_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_source_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_translator_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/cosmos_db_linked_service_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_linked_service_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_source_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_table_dataset_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/create_run_response_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_activity_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_activity_reference_object_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_data_source_linked_service_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_dataset_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_lake_analytics_usql_activity_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_notebook_activity_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_bzip2_compression_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_compression_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_deflate_compression_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_gzip_compression_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_reference_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_resource_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_storage_format_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_zip_deflate_compression_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/db2_linked_service_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/distcp_settings_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/document_db_collection_dataset_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/document_db_collection_sink_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/document_db_collection_source_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_linked_service_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_source_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_table_dataset_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_entity_dataset_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_linked_service_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_sink_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_source_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/eloqua_linked_service_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/eloqua_object_dataset_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/eloqua_source_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/error_response_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_pipeline_activity_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_ssis_package_activity_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/execution_activity_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/expression_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_identity_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_update_parameters_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_server_linked_service_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_share_dataset_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_system_sink_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_system_source_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/filter_activity_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/for_each_activity_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/ftp_server_linked_service_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/get_metadata_activity_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_big_query_linked_service_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_big_query_object_dataset_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_big_query_source_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_linked_service_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_source_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_table_dataset_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/hbase_linked_service_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/hbase_object_dataset_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/hbase_source_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_hive_activity_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_linked_service_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_map_reduce_activity_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_on_demand_linked_service_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_pig_activity_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_spark_activity_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_streaming_activity_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/hdfs_linked_service_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/hdfs_source_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/hive_linked_service_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/hive_object_dataset_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/hive_source_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/http_dataset_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/http_linked_service_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/http_source_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_linked_service_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_object_dataset_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_source_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/if_condition_activity_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/impala_linked_service_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/impala_object_dataset_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/impala_source_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_auth_keys_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_compute_properties_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_connection_info_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_custom_setup_script_properties_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_monitoring_data_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_node_ip_address_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_node_monitoring_data_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_reference_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_regenerate_key_parameters_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_remove_node_request_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_resource_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_ssis_catalog_info_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_ssis_properties_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_status_list_response_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_status_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_status_response_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_vnet_properties_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_linked_service_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_object_dataset_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_source_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/json_format_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_key_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_properties_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_rbac_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service_reference_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service_resource_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/lookup_activity_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/magento_linked_service_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/magento_object_dataset_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/magento_source_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/managed_integration_runtime_error_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/managed_integration_runtime_node_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/managed_integration_runtime_operation_result_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/managed_integration_runtime_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/managed_integration_runtime_status_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_linked_service_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_source_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_table_dataset_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/marketo_linked_service_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/marketo_object_dataset_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/marketo_source_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/mongo_db_collection_dataset_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/mongo_db_linked_service_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/mongo_db_source_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/multiple_pipeline_trigger_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/my_sql_linked_service_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/netezza_linked_service_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/netezza_source_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/netezza_table_dataset_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/odata_linked_service_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/odata_resource_dataset_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/odbc_linked_service_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/odbc_sink_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_display_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_list_response_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_log_specification_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_metric_availability_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_metric_specification_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_service_specification_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_linked_service_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_sink_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_source_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_table_dataset_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/orc_format_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/parameter_specification_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/parquet_format_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_linked_service_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_object_dataset_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_source_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/phoenix_linked_service_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/phoenix_object_dataset_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/phoenix_source_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_reference_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_resource_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run_filter_parameters_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run_invoked_by_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run_query_filter_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run_query_order_by_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run_query_response_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/polybase_settings_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/postgre_sql_linked_service_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/presto_linked_service_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/presto_object_dataset_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/presto_source_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_linked_service_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_object_dataset_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_source_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/recurrence_schedule_occurrence_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/recurrence_schedule_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/redirect_incompatible_row_settings_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/redshift_unload_settings_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/relational_source_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/relational_table_dataset_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/resource_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_linked_service.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_linked_service_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_object_dataset.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_object_dataset_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_source.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_source_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/retry_policy_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_linked_service_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_marketing_cloud_linked_service_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_marketing_cloud_object_dataset_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_marketing_cloud_source_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_object_dataset_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_sink_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_source_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_bw_linked_service_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_linked_service_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_resource_dataset_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_sink_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_source_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_ecc_linked_service_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_ecc_resource_dataset_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_ecc_source_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_hana_linked_service_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/schedule_trigger_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/schedule_trigger_recurrence_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/secret_base_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/secure_string_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime_node_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime_status_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_linked_service_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_object_dataset_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_source_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/sftp_server_linked_service_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_linked_service_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_object_dataset_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_source_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/spark_linked_service_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/spark_object_dataset_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/spark_source_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_dw_sink_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_dw_source_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_linked_service_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_stored_procedure_activity_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_table_dataset_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_sink_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_source_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_linked_service_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_object_dataset_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_source_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/ssis_package_location_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/staging_settings_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/stored_procedure_parameter_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/sub_resource_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/sybase_linked_service_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/tabular_translator_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/teradata_linked_service_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/text_format_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger_pipeline_reference_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger_resource_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger_run_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/tumbling_window_trigger_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/until_activity_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/update_integration_runtime_node_request_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/update_integration_runtime_request_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/vertica_linked_service_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/vertica_source_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/vertica_table_dataset_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/wait_activity_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_activity_authentication_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_activity_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_anonymous_authentication_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_basic_authentication_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_client_certificate_authentication_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_linked_service_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_linked_service_type_properties_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_source_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_table_dataset_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_linked_service_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_object_dataset_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_source_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/zoho_linked_service_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/zoho_object_dataset_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/zoho_source_py3.py diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py index e505c7cae283..23607a9e47f1 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py @@ -9,324 +9,650 @@ # regenerated. # -------------------------------------------------------------------------- -from .resource import Resource -from .sub_resource import SubResource -from .expression import Expression -from .secure_string import SecureString -from .linked_service_reference import LinkedServiceReference -from .azure_key_vault_secret_reference import AzureKeyVaultSecretReference -from .secret_base import SecretBase -from .factory_identity import FactoryIdentity -from .factory import Factory -from .integration_runtime import IntegrationRuntime -from .integration_runtime_resource import IntegrationRuntimeResource -from .integration_runtime_reference import IntegrationRuntimeReference -from .integration_runtime_status import IntegrationRuntimeStatus -from .integration_runtime_status_response import IntegrationRuntimeStatusResponse -from .integration_runtime_status_list_response import IntegrationRuntimeStatusListResponse -from .update_integration_runtime_request import UpdateIntegrationRuntimeRequest -from .update_integration_runtime_node_request import UpdateIntegrationRuntimeNodeRequest -from .parameter_specification import ParameterSpecification -from .linked_service import LinkedService -from .linked_service_resource import LinkedServiceResource -from .dataset import Dataset -from .dataset_resource import DatasetResource -from .activity_dependency import ActivityDependency -from .activity import Activity -from .pipeline_resource import PipelineResource -from .trigger import Trigger -from .trigger_resource import TriggerResource -from .create_run_response import CreateRunResponse -from .error_response import ErrorResponse, ErrorResponseException -from .pipeline_reference import PipelineReference -from .trigger_pipeline_reference import TriggerPipelineReference -from .factory_update_parameters import FactoryUpdateParameters -from .dataset_reference import DatasetReference -from .pipeline_run_query_filter import PipelineRunQueryFilter -from .pipeline_run_query_order_by import PipelineRunQueryOrderBy -from .pipeline_run_filter_parameters import PipelineRunFilterParameters -from .pipeline_run_invoked_by import PipelineRunInvokedBy -from .pipeline_run import PipelineRun -from .pipeline_run_query_response import PipelineRunQueryResponse -from .activity_run import ActivityRun -from .trigger_run import TriggerRun -from .operation_display import OperationDisplay -from .operation_log_specification import OperationLogSpecification -from .operation_metric_availability import OperationMetricAvailability -from .operation_metric_specification import OperationMetricSpecification -from .operation_service_specification import OperationServiceSpecification -from .operation import Operation -from .operation_list_response import OperationListResponse -from .azure_databricks_linked_service import AzureDatabricksLinkedService -from .azure_data_lake_analytics_linked_service import AzureDataLakeAnalyticsLinkedService -from .hd_insight_on_demand_linked_service import HDInsightOnDemandLinkedService -from .salesforce_marketing_cloud_linked_service import SalesforceMarketingCloudLinkedService -from .netezza_linked_service import NetezzaLinkedService -from .vertica_linked_service import VerticaLinkedService -from .zoho_linked_service import ZohoLinkedService -from .xero_linked_service import XeroLinkedService -from .square_linked_service import SquareLinkedService -from .spark_linked_service import SparkLinkedService -from .shopify_linked_service import ShopifyLinkedService -from .service_now_linked_service import ServiceNowLinkedService -from .quick_books_linked_service import QuickBooksLinkedService -from .presto_linked_service import PrestoLinkedService -from .phoenix_linked_service import PhoenixLinkedService -from .paypal_linked_service import PaypalLinkedService -from .marketo_linked_service import MarketoLinkedService -from .maria_db_linked_service import MariaDBLinkedService -from .magento_linked_service import MagentoLinkedService -from .jira_linked_service import JiraLinkedService -from .impala_linked_service import ImpalaLinkedService -from .hubspot_linked_service import HubspotLinkedService -from .hive_linked_service import HiveLinkedService -from .hbase_linked_service import HBaseLinkedService -from .greenplum_linked_service import GreenplumLinkedService -from .google_big_query_linked_service import GoogleBigQueryLinkedService -from .eloqua_linked_service import EloquaLinkedService -from .drill_linked_service import DrillLinkedService -from .couchbase_linked_service import CouchbaseLinkedService -from .concur_linked_service import ConcurLinkedService -from .azure_postgre_sql_linked_service import AzurePostgreSqlLinkedService -from .amazon_mws_linked_service import AmazonMWSLinkedService -from .sap_hana_linked_service import SapHanaLinkedService -from .sap_bw_linked_service import SapBWLinkedService -from .sftp_server_linked_service import SftpServerLinkedService -from .ftp_server_linked_service import FtpServerLinkedService -from .http_linked_service import HttpLinkedService -from .azure_search_linked_service import AzureSearchLinkedService -from .custom_data_source_linked_service import CustomDataSourceLinkedService -from .amazon_redshift_linked_service import AmazonRedshiftLinkedService -from .amazon_s3_linked_service import AmazonS3LinkedService -from .sap_ecc_linked_service import SapEccLinkedService -from .sap_cloud_for_customer_linked_service import SapCloudForCustomerLinkedService -from .salesforce_linked_service import SalesforceLinkedService -from .azure_data_lake_store_linked_service import AzureDataLakeStoreLinkedService -from .mongo_db_linked_service import MongoDbLinkedService -from .cassandra_linked_service import CassandraLinkedService -from .web_client_certificate_authentication import WebClientCertificateAuthentication -from .web_basic_authentication import WebBasicAuthentication -from .web_anonymous_authentication import WebAnonymousAuthentication -from .web_linked_service_type_properties import WebLinkedServiceTypeProperties -from .web_linked_service import WebLinkedService -from .odata_linked_service import ODataLinkedService -from .hdfs_linked_service import HdfsLinkedService -from .odbc_linked_service import OdbcLinkedService -from .azure_ml_linked_service import AzureMLLinkedService -from .teradata_linked_service import TeradataLinkedService -from .db2_linked_service import Db2LinkedService -from .sybase_linked_service import SybaseLinkedService -from .postgre_sql_linked_service import PostgreSqlLinkedService -from .my_sql_linked_service import MySqlLinkedService -from .azure_my_sql_linked_service import AzureMySqlLinkedService -from .oracle_linked_service import OracleLinkedService -from .file_server_linked_service import FileServerLinkedService -from .hd_insight_linked_service import HDInsightLinkedService -from .dynamics_linked_service import DynamicsLinkedService -from .cosmos_db_linked_service import CosmosDbLinkedService -from .azure_key_vault_linked_service import AzureKeyVaultLinkedService -from .azure_batch_linked_service import AzureBatchLinkedService -from .azure_sql_database_linked_service import AzureSqlDatabaseLinkedService -from .sql_server_linked_service import SqlServerLinkedService -from .azure_sql_dw_linked_service import AzureSqlDWLinkedService -from .azure_storage_linked_service import AzureStorageLinkedService -from .salesforce_marketing_cloud_object_dataset import SalesforceMarketingCloudObjectDataset -from .vertica_table_dataset import VerticaTableDataset -from .netezza_table_dataset import NetezzaTableDataset -from .zoho_object_dataset import ZohoObjectDataset -from .xero_object_dataset import XeroObjectDataset -from .square_object_dataset import SquareObjectDataset -from .spark_object_dataset import SparkObjectDataset -from .shopify_object_dataset import ShopifyObjectDataset -from .service_now_object_dataset import ServiceNowObjectDataset -from .quick_books_object_dataset import QuickBooksObjectDataset -from .presto_object_dataset import PrestoObjectDataset -from .phoenix_object_dataset import PhoenixObjectDataset -from .paypal_object_dataset import PaypalObjectDataset -from .marketo_object_dataset import MarketoObjectDataset -from .maria_db_table_dataset import MariaDBTableDataset -from .magento_object_dataset import MagentoObjectDataset -from .jira_object_dataset import JiraObjectDataset -from .impala_object_dataset import ImpalaObjectDataset -from .hubspot_object_dataset import HubspotObjectDataset -from .hive_object_dataset import HiveObjectDataset -from .hbase_object_dataset import HBaseObjectDataset -from .greenplum_table_dataset import GreenplumTableDataset -from .google_big_query_object_dataset import GoogleBigQueryObjectDataset -from .eloqua_object_dataset import EloquaObjectDataset -from .drill_table_dataset import DrillTableDataset -from .couchbase_table_dataset import CouchbaseTableDataset -from .concur_object_dataset import ConcurObjectDataset -from .azure_postgre_sql_table_dataset import AzurePostgreSqlTableDataset -from .amazon_mws_object_dataset import AmazonMWSObjectDataset -from .dataset_zip_deflate_compression import DatasetZipDeflateCompression -from .dataset_deflate_compression import DatasetDeflateCompression -from .dataset_gzip_compression import DatasetGZipCompression -from .dataset_bzip2_compression import DatasetBZip2Compression -from .dataset_compression import DatasetCompression -from .parquet_format import ParquetFormat -from .orc_format import OrcFormat -from .avro_format import AvroFormat -from .json_format import JsonFormat -from .text_format import TextFormat -from .dataset_storage_format import DatasetStorageFormat -from .http_dataset import HttpDataset -from .azure_search_index_dataset import AzureSearchIndexDataset -from .web_table_dataset import WebTableDataset -from .sql_server_table_dataset import SqlServerTableDataset -from .sap_ecc_resource_dataset import SapEccResourceDataset -from .sap_cloud_for_customer_resource_dataset import SapCloudForCustomerResourceDataset -from .salesforce_object_dataset import SalesforceObjectDataset -from .relational_table_dataset import RelationalTableDataset -from .azure_my_sql_table_dataset import AzureMySqlTableDataset -from .oracle_table_dataset import OracleTableDataset -from .odata_resource_dataset import ODataResourceDataset -from .mongo_db_collection_dataset import MongoDbCollectionDataset -from .file_share_dataset import FileShareDataset -from .azure_data_lake_store_dataset import AzureDataLakeStoreDataset -from .dynamics_entity_dataset import DynamicsEntityDataset -from .document_db_collection_dataset import DocumentDbCollectionDataset -from .custom_dataset import CustomDataset -from .cassandra_table_dataset import CassandraTableDataset -from .azure_sql_dw_table_dataset import AzureSqlDWTableDataset -from .azure_sql_table_dataset import AzureSqlTableDataset -from .azure_table_dataset import AzureTableDataset -from .azure_blob_dataset import AzureBlobDataset -from .amazon_s3_dataset import AmazonS3Dataset -from .retry_policy import RetryPolicy -from .tumbling_window_trigger import TumblingWindowTrigger -from .blob_trigger import BlobTrigger -from .recurrence_schedule_occurrence import RecurrenceScheduleOccurrence -from .recurrence_schedule import RecurrenceSchedule -from .schedule_trigger_recurrence import ScheduleTriggerRecurrence -from .schedule_trigger import ScheduleTrigger -from .multiple_pipeline_trigger import MultiplePipelineTrigger -from .activity_policy import ActivityPolicy -from .databricks_notebook_activity import DatabricksNotebookActivity -from .data_lake_analytics_usql_activity import DataLakeAnalyticsUSQLActivity -from .azure_ml_update_resource_activity import AzureMLUpdateResourceActivity -from .azure_ml_web_service_file import AzureMLWebServiceFile -from .azure_ml_batch_execution_activity import AzureMLBatchExecutionActivity -from .get_metadata_activity import GetMetadataActivity -from .web_activity_authentication import WebActivityAuthentication -from .web_activity import WebActivity -from .redshift_unload_settings import RedshiftUnloadSettings -from .amazon_redshift_source import AmazonRedshiftSource -from .salesforce_marketing_cloud_source import SalesforceMarketingCloudSource -from .vertica_source import VerticaSource -from .netezza_source import NetezzaSource -from .zoho_source import ZohoSource -from .xero_source import XeroSource -from .square_source import SquareSource -from .spark_source import SparkSource -from .shopify_source import ShopifySource -from .service_now_source import ServiceNowSource -from .quick_books_source import QuickBooksSource -from .presto_source import PrestoSource -from .phoenix_source import PhoenixSource -from .paypal_source import PaypalSource -from .marketo_source import MarketoSource -from .maria_db_source import MariaDBSource -from .magento_source import MagentoSource -from .jira_source import JiraSource -from .impala_source import ImpalaSource -from .hubspot_source import HubspotSource -from .hive_source import HiveSource -from .hbase_source import HBaseSource -from .greenplum_source import GreenplumSource -from .google_big_query_source import GoogleBigQuerySource -from .eloqua_source import EloquaSource -from .drill_source import DrillSource -from .couchbase_source import CouchbaseSource -from .concur_source import ConcurSource -from .azure_postgre_sql_source import AzurePostgreSqlSource -from .amazon_mws_source import AmazonMWSSource -from .http_source import HttpSource -from .azure_data_lake_store_source import AzureDataLakeStoreSource -from .mongo_db_source import MongoDbSource -from .cassandra_source import CassandraSource -from .web_source import WebSource -from .oracle_source import OracleSource -from .azure_my_sql_source import AzureMySqlSource -from .distcp_settings import DistcpSettings -from .hdfs_source import HdfsSource -from .file_system_source import FileSystemSource -from .sql_dw_source import SqlDWSource -from .stored_procedure_parameter import StoredProcedureParameter -from .sql_source import SqlSource -from .sap_ecc_source import SapEccSource -from .sap_cloud_for_customer_source import SapCloudForCustomerSource -from .salesforce_source import SalesforceSource -from .relational_source import RelationalSource -from .dynamics_source import DynamicsSource -from .document_db_collection_source import DocumentDbCollectionSource -from .blob_source import BlobSource -from .azure_table_source import AzureTableSource -from .copy_source import CopySource -from .lookup_activity import LookupActivity -from .sql_server_stored_procedure_activity import SqlServerStoredProcedureActivity -from .custom_activity_reference_object import CustomActivityReferenceObject -from .custom_activity import CustomActivity -from .ssis_package_location import SSISPackageLocation -from .execute_ssis_package_activity import ExecuteSSISPackageActivity -from .hd_insight_spark_activity import HDInsightSparkActivity -from .hd_insight_streaming_activity import HDInsightStreamingActivity -from .hd_insight_map_reduce_activity import HDInsightMapReduceActivity -from .hd_insight_pig_activity import HDInsightPigActivity -from .hd_insight_hive_activity import HDInsightHiveActivity -from .redirect_incompatible_row_settings import RedirectIncompatibleRowSettings -from .staging_settings import StagingSettings -from .tabular_translator import TabularTranslator -from .copy_translator import CopyTranslator -from .salesforce_sink import SalesforceSink -from .dynamics_sink import DynamicsSink -from .odbc_sink import OdbcSink -from .azure_search_index_sink import AzureSearchIndexSink -from .azure_data_lake_store_sink import AzureDataLakeStoreSink -from .oracle_sink import OracleSink -from .polybase_settings import PolybaseSettings -from .sql_dw_sink import SqlDWSink -from .sql_sink import SqlSink -from .document_db_collection_sink import DocumentDbCollectionSink -from .file_system_sink import FileSystemSink -from .blob_sink import BlobSink -from .azure_table_sink import AzureTableSink -from .azure_queue_sink import AzureQueueSink -from .sap_cloud_for_customer_sink import SapCloudForCustomerSink -from .copy_sink import CopySink -from .copy_activity import CopyActivity -from .execution_activity import ExecutionActivity -from .filter_activity import FilterActivity -from .until_activity import UntilActivity -from .wait_activity import WaitActivity -from .for_each_activity import ForEachActivity -from .if_condition_activity import IfConditionActivity -from .execute_pipeline_activity import ExecutePipelineActivity -from .control_activity import ControlActivity -from .linked_integration_runtime import LinkedIntegrationRuntime -from .self_hosted_integration_runtime_node import SelfHostedIntegrationRuntimeNode -from .self_hosted_integration_runtime_status import SelfHostedIntegrationRuntimeStatus -from .managed_integration_runtime_operation_result import ManagedIntegrationRuntimeOperationResult -from .managed_integration_runtime_error import ManagedIntegrationRuntimeError -from .managed_integration_runtime_node import ManagedIntegrationRuntimeNode -from .managed_integration_runtime_status import ManagedIntegrationRuntimeStatus -from .linked_integration_runtime_rbac import LinkedIntegrationRuntimeRbac -from .linked_integration_runtime_key import LinkedIntegrationRuntimeKey -from .linked_integration_runtime_properties import LinkedIntegrationRuntimeProperties -from .self_hosted_integration_runtime import SelfHostedIntegrationRuntime -from .integration_runtime_custom_setup_script_properties import IntegrationRuntimeCustomSetupScriptProperties -from .integration_runtime_ssis_catalog_info import IntegrationRuntimeSsisCatalogInfo -from .integration_runtime_ssis_properties import IntegrationRuntimeSsisProperties -from .integration_runtime_vnet_properties import IntegrationRuntimeVNetProperties -from .integration_runtime_compute_properties import IntegrationRuntimeComputeProperties -from .managed_integration_runtime import ManagedIntegrationRuntime -from .integration_runtime_node_ip_address import IntegrationRuntimeNodeIpAddress -from .integration_runtime_node_monitoring_data import IntegrationRuntimeNodeMonitoringData -from .integration_runtime_monitoring_data import IntegrationRuntimeMonitoringData -from .integration_runtime_remove_node_request import IntegrationRuntimeRemoveNodeRequest -from .integration_runtime_auth_keys import IntegrationRuntimeAuthKeys -from .integration_runtime_regenerate_key_parameters import IntegrationRuntimeRegenerateKeyParameters -from .integration_runtime_connection_info import IntegrationRuntimeConnectionInfo +try: + from .resource_py3 import Resource + from .sub_resource_py3 import SubResource + from .expression_py3 import Expression + from .secure_string_py3 import SecureString + from .linked_service_reference_py3 import LinkedServiceReference + from .azure_key_vault_secret_reference_py3 import AzureKeyVaultSecretReference + from .secret_base_py3 import SecretBase + from .factory_identity_py3 import FactoryIdentity + from .factory_py3 import Factory + from .integration_runtime_py3 import IntegrationRuntime + from .integration_runtime_resource_py3 import IntegrationRuntimeResource + from .integration_runtime_reference_py3 import IntegrationRuntimeReference + from .integration_runtime_status_py3 import IntegrationRuntimeStatus + from .integration_runtime_status_response_py3 import IntegrationRuntimeStatusResponse + from .integration_runtime_status_list_response_py3 import IntegrationRuntimeStatusListResponse + from .update_integration_runtime_request_py3 import UpdateIntegrationRuntimeRequest + from .update_integration_runtime_node_request_py3 import UpdateIntegrationRuntimeNodeRequest + from .parameter_specification_py3 import ParameterSpecification + from .linked_service_py3 import LinkedService + from .linked_service_resource_py3 import LinkedServiceResource + from .dataset_py3 import Dataset + from .dataset_resource_py3 import DatasetResource + from .activity_dependency_py3 import ActivityDependency + from .activity_py3 import Activity + from .pipeline_resource_py3 import PipelineResource + from .trigger_py3 import Trigger + from .trigger_resource_py3 import TriggerResource + from .create_run_response_py3 import CreateRunResponse + from .error_response_py3 import ErrorResponse, ErrorResponseException + from .pipeline_reference_py3 import PipelineReference + from .trigger_pipeline_reference_py3 import TriggerPipelineReference + from .factory_update_parameters_py3 import FactoryUpdateParameters + from .dataset_reference_py3 import DatasetReference + from .pipeline_run_query_filter_py3 import PipelineRunQueryFilter + from .pipeline_run_query_order_by_py3 import PipelineRunQueryOrderBy + from .pipeline_run_filter_parameters_py3 import PipelineRunFilterParameters + from .pipeline_run_invoked_by_py3 import PipelineRunInvokedBy + from .pipeline_run_py3 import PipelineRun + from .pipeline_run_query_response_py3 import PipelineRunQueryResponse + from .activity_run_py3 import ActivityRun + from .trigger_run_py3 import TriggerRun + from .operation_display_py3 import OperationDisplay + from .operation_log_specification_py3 import OperationLogSpecification + from .operation_metric_availability_py3 import OperationMetricAvailability + from .operation_metric_specification_py3 import OperationMetricSpecification + from .operation_service_specification_py3 import OperationServiceSpecification + from .operation_py3 import Operation + from .operation_list_response_py3 import OperationListResponse + from .responsys_linked_service_py3 import ResponsysLinkedService + from .azure_databricks_linked_service_py3 import AzureDatabricksLinkedService + from .azure_data_lake_analytics_linked_service_py3 import AzureDataLakeAnalyticsLinkedService + from .hd_insight_on_demand_linked_service_py3 import HDInsightOnDemandLinkedService + from .salesforce_marketing_cloud_linked_service_py3 import SalesforceMarketingCloudLinkedService + from .netezza_linked_service_py3 import NetezzaLinkedService + from .vertica_linked_service_py3 import VerticaLinkedService + from .zoho_linked_service_py3 import ZohoLinkedService + from .xero_linked_service_py3 import XeroLinkedService + from .square_linked_service_py3 import SquareLinkedService + from .spark_linked_service_py3 import SparkLinkedService + from .shopify_linked_service_py3 import ShopifyLinkedService + from .service_now_linked_service_py3 import ServiceNowLinkedService + from .quick_books_linked_service_py3 import QuickBooksLinkedService + from .presto_linked_service_py3 import PrestoLinkedService + from .phoenix_linked_service_py3 import PhoenixLinkedService + from .paypal_linked_service_py3 import PaypalLinkedService + from .marketo_linked_service_py3 import MarketoLinkedService + from .maria_db_linked_service_py3 import MariaDBLinkedService + from .magento_linked_service_py3 import MagentoLinkedService + from .jira_linked_service_py3 import JiraLinkedService + from .impala_linked_service_py3 import ImpalaLinkedService + from .hubspot_linked_service_py3 import HubspotLinkedService + from .hive_linked_service_py3 import HiveLinkedService + from .hbase_linked_service_py3 import HBaseLinkedService + from .greenplum_linked_service_py3 import GreenplumLinkedService + from .google_big_query_linked_service_py3 import GoogleBigQueryLinkedService + from .eloqua_linked_service_py3 import EloquaLinkedService + from .drill_linked_service_py3 import DrillLinkedService + from .couchbase_linked_service_py3 import CouchbaseLinkedService + from .concur_linked_service_py3 import ConcurLinkedService + from .azure_postgre_sql_linked_service_py3 import AzurePostgreSqlLinkedService + from .amazon_mws_linked_service_py3 import AmazonMWSLinkedService + from .sap_hana_linked_service_py3 import SapHanaLinkedService + from .sap_bw_linked_service_py3 import SapBWLinkedService + from .sftp_server_linked_service_py3 import SftpServerLinkedService + from .ftp_server_linked_service_py3 import FtpServerLinkedService + from .http_linked_service_py3 import HttpLinkedService + from .azure_search_linked_service_py3 import AzureSearchLinkedService + from .custom_data_source_linked_service_py3 import CustomDataSourceLinkedService + from .amazon_redshift_linked_service_py3 import AmazonRedshiftLinkedService + from .amazon_s3_linked_service_py3 import AmazonS3LinkedService + from .sap_ecc_linked_service_py3 import SapEccLinkedService + from .sap_cloud_for_customer_linked_service_py3 import SapCloudForCustomerLinkedService + from .salesforce_linked_service_py3 import SalesforceLinkedService + from .azure_data_lake_store_linked_service_py3 import AzureDataLakeStoreLinkedService + from .mongo_db_linked_service_py3 import MongoDbLinkedService + from .cassandra_linked_service_py3 import CassandraLinkedService + from .web_client_certificate_authentication_py3 import WebClientCertificateAuthentication + from .web_basic_authentication_py3 import WebBasicAuthentication + from .web_anonymous_authentication_py3 import WebAnonymousAuthentication + from .web_linked_service_type_properties_py3 import WebLinkedServiceTypeProperties + from .web_linked_service_py3 import WebLinkedService + from .odata_linked_service_py3 import ODataLinkedService + from .hdfs_linked_service_py3 import HdfsLinkedService + from .odbc_linked_service_py3 import OdbcLinkedService + from .azure_ml_linked_service_py3 import AzureMLLinkedService + from .teradata_linked_service_py3 import TeradataLinkedService + from .db2_linked_service_py3 import Db2LinkedService + from .sybase_linked_service_py3 import SybaseLinkedService + from .postgre_sql_linked_service_py3 import PostgreSqlLinkedService + from .my_sql_linked_service_py3 import MySqlLinkedService + from .azure_my_sql_linked_service_py3 import AzureMySqlLinkedService + from .oracle_linked_service_py3 import OracleLinkedService + from .file_server_linked_service_py3 import FileServerLinkedService + from .hd_insight_linked_service_py3 import HDInsightLinkedService + from .dynamics_linked_service_py3 import DynamicsLinkedService + from .cosmos_db_linked_service_py3 import CosmosDbLinkedService + from .azure_key_vault_linked_service_py3 import AzureKeyVaultLinkedService + from .azure_batch_linked_service_py3 import AzureBatchLinkedService + from .azure_sql_database_linked_service_py3 import AzureSqlDatabaseLinkedService + from .sql_server_linked_service_py3 import SqlServerLinkedService + from .azure_sql_dw_linked_service_py3 import AzureSqlDWLinkedService + from .azure_storage_linked_service_py3 import AzureStorageLinkedService + from .responsys_object_dataset_py3 import ResponsysObjectDataset + from .salesforce_marketing_cloud_object_dataset_py3 import SalesforceMarketingCloudObjectDataset + from .vertica_table_dataset_py3 import VerticaTableDataset + from .netezza_table_dataset_py3 import NetezzaTableDataset + from .zoho_object_dataset_py3 import ZohoObjectDataset + from .xero_object_dataset_py3 import XeroObjectDataset + from .square_object_dataset_py3 import SquareObjectDataset + from .spark_object_dataset_py3 import SparkObjectDataset + from .shopify_object_dataset_py3 import ShopifyObjectDataset + from .service_now_object_dataset_py3 import ServiceNowObjectDataset + from .quick_books_object_dataset_py3 import QuickBooksObjectDataset + from .presto_object_dataset_py3 import PrestoObjectDataset + from .phoenix_object_dataset_py3 import PhoenixObjectDataset + from .paypal_object_dataset_py3 import PaypalObjectDataset + from .marketo_object_dataset_py3 import MarketoObjectDataset + from .maria_db_table_dataset_py3 import MariaDBTableDataset + from .magento_object_dataset_py3 import MagentoObjectDataset + from .jira_object_dataset_py3 import JiraObjectDataset + from .impala_object_dataset_py3 import ImpalaObjectDataset + from .hubspot_object_dataset_py3 import HubspotObjectDataset + from .hive_object_dataset_py3 import HiveObjectDataset + from .hbase_object_dataset_py3 import HBaseObjectDataset + from .greenplum_table_dataset_py3 import GreenplumTableDataset + from .google_big_query_object_dataset_py3 import GoogleBigQueryObjectDataset + from .eloqua_object_dataset_py3 import EloquaObjectDataset + from .drill_table_dataset_py3 import DrillTableDataset + from .couchbase_table_dataset_py3 import CouchbaseTableDataset + from .concur_object_dataset_py3 import ConcurObjectDataset + from .azure_postgre_sql_table_dataset_py3 import AzurePostgreSqlTableDataset + from .amazon_mws_object_dataset_py3 import AmazonMWSObjectDataset + from .dataset_zip_deflate_compression_py3 import DatasetZipDeflateCompression + from .dataset_deflate_compression_py3 import DatasetDeflateCompression + from .dataset_gzip_compression_py3 import DatasetGZipCompression + from .dataset_bzip2_compression_py3 import DatasetBZip2Compression + from .dataset_compression_py3 import DatasetCompression + from .parquet_format_py3 import ParquetFormat + from .orc_format_py3 import OrcFormat + from .avro_format_py3 import AvroFormat + from .json_format_py3 import JsonFormat + from .text_format_py3 import TextFormat + from .dataset_storage_format_py3 import DatasetStorageFormat + from .http_dataset_py3 import HttpDataset + from .azure_search_index_dataset_py3 import AzureSearchIndexDataset + from .web_table_dataset_py3 import WebTableDataset + from .sql_server_table_dataset_py3 import SqlServerTableDataset + from .sap_ecc_resource_dataset_py3 import SapEccResourceDataset + from .sap_cloud_for_customer_resource_dataset_py3 import SapCloudForCustomerResourceDataset + from .salesforce_object_dataset_py3 import SalesforceObjectDataset + from .relational_table_dataset_py3 import RelationalTableDataset + from .azure_my_sql_table_dataset_py3 import AzureMySqlTableDataset + from .oracle_table_dataset_py3 import OracleTableDataset + from .odata_resource_dataset_py3 import ODataResourceDataset + from .mongo_db_collection_dataset_py3 import MongoDbCollectionDataset + from .file_share_dataset_py3 import FileShareDataset + from .azure_data_lake_store_dataset_py3 import AzureDataLakeStoreDataset + from .dynamics_entity_dataset_py3 import DynamicsEntityDataset + from .document_db_collection_dataset_py3 import DocumentDbCollectionDataset + from .custom_dataset_py3 import CustomDataset + from .cassandra_table_dataset_py3 import CassandraTableDataset + from .azure_sql_dw_table_dataset_py3 import AzureSqlDWTableDataset + from .azure_sql_table_dataset_py3 import AzureSqlTableDataset + from .azure_table_dataset_py3 import AzureTableDataset + from .azure_blob_dataset_py3 import AzureBlobDataset + from .amazon_s3_dataset_py3 import AmazonS3Dataset + from .retry_policy_py3 import RetryPolicy + from .tumbling_window_trigger_py3 import TumblingWindowTrigger + from .blob_trigger_py3 import BlobTrigger + from .recurrence_schedule_occurrence_py3 import RecurrenceScheduleOccurrence + from .recurrence_schedule_py3 import RecurrenceSchedule + from .schedule_trigger_recurrence_py3 import ScheduleTriggerRecurrence + from .schedule_trigger_py3 import ScheduleTrigger + from .multiple_pipeline_trigger_py3 import MultiplePipelineTrigger + from .activity_policy_py3 import ActivityPolicy + from .databricks_notebook_activity_py3 import DatabricksNotebookActivity + from .data_lake_analytics_usql_activity_py3 import DataLakeAnalyticsUSQLActivity + from .azure_ml_update_resource_activity_py3 import AzureMLUpdateResourceActivity + from .azure_ml_web_service_file_py3 import AzureMLWebServiceFile + from .azure_ml_batch_execution_activity_py3 import AzureMLBatchExecutionActivity + from .get_metadata_activity_py3 import GetMetadataActivity + from .web_activity_authentication_py3 import WebActivityAuthentication + from .web_activity_py3 import WebActivity + from .redshift_unload_settings_py3 import RedshiftUnloadSettings + from .amazon_redshift_source_py3 import AmazonRedshiftSource + from .responsys_source_py3 import ResponsysSource + from .salesforce_marketing_cloud_source_py3 import SalesforceMarketingCloudSource + from .vertica_source_py3 import VerticaSource + from .netezza_source_py3 import NetezzaSource + from .zoho_source_py3 import ZohoSource + from .xero_source_py3 import XeroSource + from .square_source_py3 import SquareSource + from .spark_source_py3 import SparkSource + from .shopify_source_py3 import ShopifySource + from .service_now_source_py3 import ServiceNowSource + from .quick_books_source_py3 import QuickBooksSource + from .presto_source_py3 import PrestoSource + from .phoenix_source_py3 import PhoenixSource + from .paypal_source_py3 import PaypalSource + from .marketo_source_py3 import MarketoSource + from .maria_db_source_py3 import MariaDBSource + from .magento_source_py3 import MagentoSource + from .jira_source_py3 import JiraSource + from .impala_source_py3 import ImpalaSource + from .hubspot_source_py3 import HubspotSource + from .hive_source_py3 import HiveSource + from .hbase_source_py3 import HBaseSource + from .greenplum_source_py3 import GreenplumSource + from .google_big_query_source_py3 import GoogleBigQuerySource + from .eloqua_source_py3 import EloquaSource + from .drill_source_py3 import DrillSource + from .couchbase_source_py3 import CouchbaseSource + from .concur_source_py3 import ConcurSource + from .azure_postgre_sql_source_py3 import AzurePostgreSqlSource + from .amazon_mws_source_py3 import AmazonMWSSource + from .http_source_py3 import HttpSource + from .azure_data_lake_store_source_py3 import AzureDataLakeStoreSource + from .mongo_db_source_py3 import MongoDbSource + from .cassandra_source_py3 import CassandraSource + from .web_source_py3 import WebSource + from .oracle_source_py3 import OracleSource + from .azure_my_sql_source_py3 import AzureMySqlSource + from .distcp_settings_py3 import DistcpSettings + from .hdfs_source_py3 import HdfsSource + from .file_system_source_py3 import FileSystemSource + from .sql_dw_source_py3 import SqlDWSource + from .stored_procedure_parameter_py3 import StoredProcedureParameter + from .sql_source_py3 import SqlSource + from .sap_ecc_source_py3 import SapEccSource + from .sap_cloud_for_customer_source_py3 import SapCloudForCustomerSource + from .salesforce_source_py3 import SalesforceSource + from .relational_source_py3 import RelationalSource + from .dynamics_source_py3 import DynamicsSource + from .document_db_collection_source_py3 import DocumentDbCollectionSource + from .blob_source_py3 import BlobSource + from .azure_table_source_py3 import AzureTableSource + from .copy_source_py3 import CopySource + from .lookup_activity_py3 import LookupActivity + from .sql_server_stored_procedure_activity_py3 import SqlServerStoredProcedureActivity + from .custom_activity_reference_object_py3 import CustomActivityReferenceObject + from .custom_activity_py3 import CustomActivity + from .ssis_package_location_py3 import SSISPackageLocation + from .execute_ssis_package_activity_py3 import ExecuteSSISPackageActivity + from .hd_insight_spark_activity_py3 import HDInsightSparkActivity + from .hd_insight_streaming_activity_py3 import HDInsightStreamingActivity + from .hd_insight_map_reduce_activity_py3 import HDInsightMapReduceActivity + from .hd_insight_pig_activity_py3 import HDInsightPigActivity + from .hd_insight_hive_activity_py3 import HDInsightHiveActivity + from .redirect_incompatible_row_settings_py3 import RedirectIncompatibleRowSettings + from .staging_settings_py3 import StagingSettings + from .tabular_translator_py3 import TabularTranslator + from .copy_translator_py3 import CopyTranslator + from .salesforce_sink_py3 import SalesforceSink + from .dynamics_sink_py3 import DynamicsSink + from .odbc_sink_py3 import OdbcSink + from .azure_search_index_sink_py3 import AzureSearchIndexSink + from .azure_data_lake_store_sink_py3 import AzureDataLakeStoreSink + from .oracle_sink_py3 import OracleSink + from .polybase_settings_py3 import PolybaseSettings + from .sql_dw_sink_py3 import SqlDWSink + from .sql_sink_py3 import SqlSink + from .document_db_collection_sink_py3 import DocumentDbCollectionSink + from .file_system_sink_py3 import FileSystemSink + from .blob_sink_py3 import BlobSink + from .azure_table_sink_py3 import AzureTableSink + from .azure_queue_sink_py3 import AzureQueueSink + from .sap_cloud_for_customer_sink_py3 import SapCloudForCustomerSink + from .copy_sink_py3 import CopySink + from .copy_activity_py3 import CopyActivity + from .execution_activity_py3 import ExecutionActivity + from .filter_activity_py3 import FilterActivity + from .until_activity_py3 import UntilActivity + from .wait_activity_py3 import WaitActivity + from .for_each_activity_py3 import ForEachActivity + from .if_condition_activity_py3 import IfConditionActivity + from .execute_pipeline_activity_py3 import ExecutePipelineActivity + from .control_activity_py3 import ControlActivity + from .linked_integration_runtime_py3 import LinkedIntegrationRuntime + from .self_hosted_integration_runtime_node_py3 import SelfHostedIntegrationRuntimeNode + from .self_hosted_integration_runtime_status_py3 import SelfHostedIntegrationRuntimeStatus + from .managed_integration_runtime_operation_result_py3 import ManagedIntegrationRuntimeOperationResult + from .managed_integration_runtime_error_py3 import ManagedIntegrationRuntimeError + from .managed_integration_runtime_node_py3 import ManagedIntegrationRuntimeNode + from .managed_integration_runtime_status_py3 import ManagedIntegrationRuntimeStatus + from .linked_integration_runtime_rbac_py3 import LinkedIntegrationRuntimeRbac + from .linked_integration_runtime_key_py3 import LinkedIntegrationRuntimeKey + from .linked_integration_runtime_properties_py3 import LinkedIntegrationRuntimeProperties + from .self_hosted_integration_runtime_py3 import SelfHostedIntegrationRuntime + from .integration_runtime_custom_setup_script_properties_py3 import IntegrationRuntimeCustomSetupScriptProperties + from .integration_runtime_ssis_catalog_info_py3 import IntegrationRuntimeSsisCatalogInfo + from .integration_runtime_ssis_properties_py3 import IntegrationRuntimeSsisProperties + from .integration_runtime_vnet_properties_py3 import IntegrationRuntimeVNetProperties + from .integration_runtime_compute_properties_py3 import IntegrationRuntimeComputeProperties + from .managed_integration_runtime_py3 import ManagedIntegrationRuntime + from .integration_runtime_node_ip_address_py3 import IntegrationRuntimeNodeIpAddress + from .integration_runtime_node_monitoring_data_py3 import IntegrationRuntimeNodeMonitoringData + from .integration_runtime_monitoring_data_py3 import IntegrationRuntimeMonitoringData + from .integration_runtime_remove_node_request_py3 import IntegrationRuntimeRemoveNodeRequest + from .integration_runtime_auth_keys_py3 import IntegrationRuntimeAuthKeys + from .integration_runtime_regenerate_key_parameters_py3 import IntegrationRuntimeRegenerateKeyParameters + from .integration_runtime_connection_info_py3 import IntegrationRuntimeConnectionInfo +except (SyntaxError, ImportError): + from .resource import Resource + from .sub_resource import SubResource + from .expression import Expression + from .secure_string import SecureString + from .linked_service_reference import LinkedServiceReference + from .azure_key_vault_secret_reference import AzureKeyVaultSecretReference + from .secret_base import SecretBase + from .factory_identity import FactoryIdentity + from .factory import Factory + from .integration_runtime import IntegrationRuntime + from .integration_runtime_resource import IntegrationRuntimeResource + from .integration_runtime_reference import IntegrationRuntimeReference + from .integration_runtime_status import IntegrationRuntimeStatus + from .integration_runtime_status_response import IntegrationRuntimeStatusResponse + from .integration_runtime_status_list_response import IntegrationRuntimeStatusListResponse + from .update_integration_runtime_request import UpdateIntegrationRuntimeRequest + from .update_integration_runtime_node_request import UpdateIntegrationRuntimeNodeRequest + from .parameter_specification import ParameterSpecification + from .linked_service import LinkedService + from .linked_service_resource import LinkedServiceResource + from .dataset import Dataset + from .dataset_resource import DatasetResource + from .activity_dependency import ActivityDependency + from .activity import Activity + from .pipeline_resource import PipelineResource + from .trigger import Trigger + from .trigger_resource import TriggerResource + from .create_run_response import CreateRunResponse + from .error_response import ErrorResponse, ErrorResponseException + from .pipeline_reference import PipelineReference + from .trigger_pipeline_reference import TriggerPipelineReference + from .factory_update_parameters import FactoryUpdateParameters + from .dataset_reference import DatasetReference + from .pipeline_run_query_filter import PipelineRunQueryFilter + from .pipeline_run_query_order_by import PipelineRunQueryOrderBy + from .pipeline_run_filter_parameters import PipelineRunFilterParameters + from .pipeline_run_invoked_by import PipelineRunInvokedBy + from .pipeline_run import PipelineRun + from .pipeline_run_query_response import PipelineRunQueryResponse + from .activity_run import ActivityRun + from .trigger_run import TriggerRun + from .operation_display import OperationDisplay + from .operation_log_specification import OperationLogSpecification + from .operation_metric_availability import OperationMetricAvailability + from .operation_metric_specification import OperationMetricSpecification + from .operation_service_specification import OperationServiceSpecification + from .operation import Operation + from .operation_list_response import OperationListResponse + from .responsys_linked_service import ResponsysLinkedService + from .azure_databricks_linked_service import AzureDatabricksLinkedService + from .azure_data_lake_analytics_linked_service import AzureDataLakeAnalyticsLinkedService + from .hd_insight_on_demand_linked_service import HDInsightOnDemandLinkedService + from .salesforce_marketing_cloud_linked_service import SalesforceMarketingCloudLinkedService + from .netezza_linked_service import NetezzaLinkedService + from .vertica_linked_service import VerticaLinkedService + from .zoho_linked_service import ZohoLinkedService + from .xero_linked_service import XeroLinkedService + from .square_linked_service import SquareLinkedService + from .spark_linked_service import SparkLinkedService + from .shopify_linked_service import ShopifyLinkedService + from .service_now_linked_service import ServiceNowLinkedService + from .quick_books_linked_service import QuickBooksLinkedService + from .presto_linked_service import PrestoLinkedService + from .phoenix_linked_service import PhoenixLinkedService + from .paypal_linked_service import PaypalLinkedService + from .marketo_linked_service import MarketoLinkedService + from .maria_db_linked_service import MariaDBLinkedService + from .magento_linked_service import MagentoLinkedService + from .jira_linked_service import JiraLinkedService + from .impala_linked_service import ImpalaLinkedService + from .hubspot_linked_service import HubspotLinkedService + from .hive_linked_service import HiveLinkedService + from .hbase_linked_service import HBaseLinkedService + from .greenplum_linked_service import GreenplumLinkedService + from .google_big_query_linked_service import GoogleBigQueryLinkedService + from .eloqua_linked_service import EloquaLinkedService + from .drill_linked_service import DrillLinkedService + from .couchbase_linked_service import CouchbaseLinkedService + from .concur_linked_service import ConcurLinkedService + from .azure_postgre_sql_linked_service import AzurePostgreSqlLinkedService + from .amazon_mws_linked_service import AmazonMWSLinkedService + from .sap_hana_linked_service import SapHanaLinkedService + from .sap_bw_linked_service import SapBWLinkedService + from .sftp_server_linked_service import SftpServerLinkedService + from .ftp_server_linked_service import FtpServerLinkedService + from .http_linked_service import HttpLinkedService + from .azure_search_linked_service import AzureSearchLinkedService + from .custom_data_source_linked_service import CustomDataSourceLinkedService + from .amazon_redshift_linked_service import AmazonRedshiftLinkedService + from .amazon_s3_linked_service import AmazonS3LinkedService + from .sap_ecc_linked_service import SapEccLinkedService + from .sap_cloud_for_customer_linked_service import SapCloudForCustomerLinkedService + from .salesforce_linked_service import SalesforceLinkedService + from .azure_data_lake_store_linked_service import AzureDataLakeStoreLinkedService + from .mongo_db_linked_service import MongoDbLinkedService + from .cassandra_linked_service import CassandraLinkedService + from .web_client_certificate_authentication import WebClientCertificateAuthentication + from .web_basic_authentication import WebBasicAuthentication + from .web_anonymous_authentication import WebAnonymousAuthentication + from .web_linked_service_type_properties import WebLinkedServiceTypeProperties + from .web_linked_service import WebLinkedService + from .odata_linked_service import ODataLinkedService + from .hdfs_linked_service import HdfsLinkedService + from .odbc_linked_service import OdbcLinkedService + from .azure_ml_linked_service import AzureMLLinkedService + from .teradata_linked_service import TeradataLinkedService + from .db2_linked_service import Db2LinkedService + from .sybase_linked_service import SybaseLinkedService + from .postgre_sql_linked_service import PostgreSqlLinkedService + from .my_sql_linked_service import MySqlLinkedService + from .azure_my_sql_linked_service import AzureMySqlLinkedService + from .oracle_linked_service import OracleLinkedService + from .file_server_linked_service import FileServerLinkedService + from .hd_insight_linked_service import HDInsightLinkedService + from .dynamics_linked_service import DynamicsLinkedService + from .cosmos_db_linked_service import CosmosDbLinkedService + from .azure_key_vault_linked_service import AzureKeyVaultLinkedService + from .azure_batch_linked_service import AzureBatchLinkedService + from .azure_sql_database_linked_service import AzureSqlDatabaseLinkedService + from .sql_server_linked_service import SqlServerLinkedService + from .azure_sql_dw_linked_service import AzureSqlDWLinkedService + from .azure_storage_linked_service import AzureStorageLinkedService + from .responsys_object_dataset import ResponsysObjectDataset + from .salesforce_marketing_cloud_object_dataset import SalesforceMarketingCloudObjectDataset + from .vertica_table_dataset import VerticaTableDataset + from .netezza_table_dataset import NetezzaTableDataset + from .zoho_object_dataset import ZohoObjectDataset + from .xero_object_dataset import XeroObjectDataset + from .square_object_dataset import SquareObjectDataset + from .spark_object_dataset import SparkObjectDataset + from .shopify_object_dataset import ShopifyObjectDataset + from .service_now_object_dataset import ServiceNowObjectDataset + from .quick_books_object_dataset import QuickBooksObjectDataset + from .presto_object_dataset import PrestoObjectDataset + from .phoenix_object_dataset import PhoenixObjectDataset + from .paypal_object_dataset import PaypalObjectDataset + from .marketo_object_dataset import MarketoObjectDataset + from .maria_db_table_dataset import MariaDBTableDataset + from .magento_object_dataset import MagentoObjectDataset + from .jira_object_dataset import JiraObjectDataset + from .impala_object_dataset import ImpalaObjectDataset + from .hubspot_object_dataset import HubspotObjectDataset + from .hive_object_dataset import HiveObjectDataset + from .hbase_object_dataset import HBaseObjectDataset + from .greenplum_table_dataset import GreenplumTableDataset + from .google_big_query_object_dataset import GoogleBigQueryObjectDataset + from .eloqua_object_dataset import EloquaObjectDataset + from .drill_table_dataset import DrillTableDataset + from .couchbase_table_dataset import CouchbaseTableDataset + from .concur_object_dataset import ConcurObjectDataset + from .azure_postgre_sql_table_dataset import AzurePostgreSqlTableDataset + from .amazon_mws_object_dataset import AmazonMWSObjectDataset + from .dataset_zip_deflate_compression import DatasetZipDeflateCompression + from .dataset_deflate_compression import DatasetDeflateCompression + from .dataset_gzip_compression import DatasetGZipCompression + from .dataset_bzip2_compression import DatasetBZip2Compression + from .dataset_compression import DatasetCompression + from .parquet_format import ParquetFormat + from .orc_format import OrcFormat + from .avro_format import AvroFormat + from .json_format import JsonFormat + from .text_format import TextFormat + from .dataset_storage_format import DatasetStorageFormat + from .http_dataset import HttpDataset + from .azure_search_index_dataset import AzureSearchIndexDataset + from .web_table_dataset import WebTableDataset + from .sql_server_table_dataset import SqlServerTableDataset + from .sap_ecc_resource_dataset import SapEccResourceDataset + from .sap_cloud_for_customer_resource_dataset import SapCloudForCustomerResourceDataset + from .salesforce_object_dataset import SalesforceObjectDataset + from .relational_table_dataset import RelationalTableDataset + from .azure_my_sql_table_dataset import AzureMySqlTableDataset + from .oracle_table_dataset import OracleTableDataset + from .odata_resource_dataset import ODataResourceDataset + from .mongo_db_collection_dataset import MongoDbCollectionDataset + from .file_share_dataset import FileShareDataset + from .azure_data_lake_store_dataset import AzureDataLakeStoreDataset + from .dynamics_entity_dataset import DynamicsEntityDataset + from .document_db_collection_dataset import DocumentDbCollectionDataset + from .custom_dataset import CustomDataset + from .cassandra_table_dataset import CassandraTableDataset + from .azure_sql_dw_table_dataset import AzureSqlDWTableDataset + from .azure_sql_table_dataset import AzureSqlTableDataset + from .azure_table_dataset import AzureTableDataset + from .azure_blob_dataset import AzureBlobDataset + from .amazon_s3_dataset import AmazonS3Dataset + from .retry_policy import RetryPolicy + from .tumbling_window_trigger import TumblingWindowTrigger + from .blob_trigger import BlobTrigger + from .recurrence_schedule_occurrence import RecurrenceScheduleOccurrence + from .recurrence_schedule import RecurrenceSchedule + from .schedule_trigger_recurrence import ScheduleTriggerRecurrence + from .schedule_trigger import ScheduleTrigger + from .multiple_pipeline_trigger import MultiplePipelineTrigger + from .activity_policy import ActivityPolicy + from .databricks_notebook_activity import DatabricksNotebookActivity + from .data_lake_analytics_usql_activity import DataLakeAnalyticsUSQLActivity + from .azure_ml_update_resource_activity import AzureMLUpdateResourceActivity + from .azure_ml_web_service_file import AzureMLWebServiceFile + from .azure_ml_batch_execution_activity import AzureMLBatchExecutionActivity + from .get_metadata_activity import GetMetadataActivity + from .web_activity_authentication import WebActivityAuthentication + from .web_activity import WebActivity + from .redshift_unload_settings import RedshiftUnloadSettings + from .amazon_redshift_source import AmazonRedshiftSource + from .responsys_source import ResponsysSource + from .salesforce_marketing_cloud_source import SalesforceMarketingCloudSource + from .vertica_source import VerticaSource + from .netezza_source import NetezzaSource + from .zoho_source import ZohoSource + from .xero_source import XeroSource + from .square_source import SquareSource + from .spark_source import SparkSource + from .shopify_source import ShopifySource + from .service_now_source import ServiceNowSource + from .quick_books_source import QuickBooksSource + from .presto_source import PrestoSource + from .phoenix_source import PhoenixSource + from .paypal_source import PaypalSource + from .marketo_source import MarketoSource + from .maria_db_source import MariaDBSource + from .magento_source import MagentoSource + from .jira_source import JiraSource + from .impala_source import ImpalaSource + from .hubspot_source import HubspotSource + from .hive_source import HiveSource + from .hbase_source import HBaseSource + from .greenplum_source import GreenplumSource + from .google_big_query_source import GoogleBigQuerySource + from .eloqua_source import EloquaSource + from .drill_source import DrillSource + from .couchbase_source import CouchbaseSource + from .concur_source import ConcurSource + from .azure_postgre_sql_source import AzurePostgreSqlSource + from .amazon_mws_source import AmazonMWSSource + from .http_source import HttpSource + from .azure_data_lake_store_source import AzureDataLakeStoreSource + from .mongo_db_source import MongoDbSource + from .cassandra_source import CassandraSource + from .web_source import WebSource + from .oracle_source import OracleSource + from .azure_my_sql_source import AzureMySqlSource + from .distcp_settings import DistcpSettings + from .hdfs_source import HdfsSource + from .file_system_source import FileSystemSource + from .sql_dw_source import SqlDWSource + from .stored_procedure_parameter import StoredProcedureParameter + from .sql_source import SqlSource + from .sap_ecc_source import SapEccSource + from .sap_cloud_for_customer_source import SapCloudForCustomerSource + from .salesforce_source import SalesforceSource + from .relational_source import RelationalSource + from .dynamics_source import DynamicsSource + from .document_db_collection_source import DocumentDbCollectionSource + from .blob_source import BlobSource + from .azure_table_source import AzureTableSource + from .copy_source import CopySource + from .lookup_activity import LookupActivity + from .sql_server_stored_procedure_activity import SqlServerStoredProcedureActivity + from .custom_activity_reference_object import CustomActivityReferenceObject + from .custom_activity import CustomActivity + from .ssis_package_location import SSISPackageLocation + from .execute_ssis_package_activity import ExecuteSSISPackageActivity + from .hd_insight_spark_activity import HDInsightSparkActivity + from .hd_insight_streaming_activity import HDInsightStreamingActivity + from .hd_insight_map_reduce_activity import HDInsightMapReduceActivity + from .hd_insight_pig_activity import HDInsightPigActivity + from .hd_insight_hive_activity import HDInsightHiveActivity + from .redirect_incompatible_row_settings import RedirectIncompatibleRowSettings + from .staging_settings import StagingSettings + from .tabular_translator import TabularTranslator + from .copy_translator import CopyTranslator + from .salesforce_sink import SalesforceSink + from .dynamics_sink import DynamicsSink + from .odbc_sink import OdbcSink + from .azure_search_index_sink import AzureSearchIndexSink + from .azure_data_lake_store_sink import AzureDataLakeStoreSink + from .oracle_sink import OracleSink + from .polybase_settings import PolybaseSettings + from .sql_dw_sink import SqlDWSink + from .sql_sink import SqlSink + from .document_db_collection_sink import DocumentDbCollectionSink + from .file_system_sink import FileSystemSink + from .blob_sink import BlobSink + from .azure_table_sink import AzureTableSink + from .azure_queue_sink import AzureQueueSink + from .sap_cloud_for_customer_sink import SapCloudForCustomerSink + from .copy_sink import CopySink + from .copy_activity import CopyActivity + from .execution_activity import ExecutionActivity + from .filter_activity import FilterActivity + from .until_activity import UntilActivity + from .wait_activity import WaitActivity + from .for_each_activity import ForEachActivity + from .if_condition_activity import IfConditionActivity + from .execute_pipeline_activity import ExecutePipelineActivity + from .control_activity import ControlActivity + from .linked_integration_runtime import LinkedIntegrationRuntime + from .self_hosted_integration_runtime_node import SelfHostedIntegrationRuntimeNode + from .self_hosted_integration_runtime_status import SelfHostedIntegrationRuntimeStatus + from .managed_integration_runtime_operation_result import ManagedIntegrationRuntimeOperationResult + from .managed_integration_runtime_error import ManagedIntegrationRuntimeError + from .managed_integration_runtime_node import ManagedIntegrationRuntimeNode + from .managed_integration_runtime_status import ManagedIntegrationRuntimeStatus + from .linked_integration_runtime_rbac import LinkedIntegrationRuntimeRbac + from .linked_integration_runtime_key import LinkedIntegrationRuntimeKey + from .linked_integration_runtime_properties import LinkedIntegrationRuntimeProperties + from .self_hosted_integration_runtime import SelfHostedIntegrationRuntime + from .integration_runtime_custom_setup_script_properties import IntegrationRuntimeCustomSetupScriptProperties + from .integration_runtime_ssis_catalog_info import IntegrationRuntimeSsisCatalogInfo + from .integration_runtime_ssis_properties import IntegrationRuntimeSsisProperties + from .integration_runtime_vnet_properties import IntegrationRuntimeVNetProperties + from .integration_runtime_compute_properties import IntegrationRuntimeComputeProperties + from .managed_integration_runtime import ManagedIntegrationRuntime + from .integration_runtime_node_ip_address import IntegrationRuntimeNodeIpAddress + from .integration_runtime_node_monitoring_data import IntegrationRuntimeNodeMonitoringData + from .integration_runtime_monitoring_data import IntegrationRuntimeMonitoringData + from .integration_runtime_remove_node_request import IntegrationRuntimeRemoveNodeRequest + from .integration_runtime_auth_keys import IntegrationRuntimeAuthKeys + from .integration_runtime_regenerate_key_parameters import IntegrationRuntimeRegenerateKeyParameters + from .integration_runtime_connection_info import IntegrationRuntimeConnectionInfo from .factory_paged import FactoryPaged from .integration_runtime_resource_paged import IntegrationRuntimeResourcePaged from .linked_service_resource_paged import LinkedServiceResourcePaged @@ -444,6 +770,7 @@ 'OperationServiceSpecification', 'Operation', 'OperationListResponse', + 'ResponsysLinkedService', 'AzureDatabricksLinkedService', 'AzureDataLakeAnalyticsLinkedService', 'HDInsightOnDemandLinkedService', @@ -517,6 +844,7 @@ 'SqlServerLinkedService', 'AzureSqlDWLinkedService', 'AzureStorageLinkedService', + 'ResponsysObjectDataset', 'SalesforceMarketingCloudObjectDataset', 'VerticaTableDataset', 'NetezzaTableDataset', @@ -599,6 +927,7 @@ 'WebActivity', 'RedshiftUnloadSettings', 'AmazonRedshiftSource', + 'ResponsysSource', 'SalesforceMarketingCloudSource', 'VerticaSource', 'NetezzaSource', diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/activity.py index 600b800ce112..602b8d52a38b 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/activity.py @@ -18,16 +18,18 @@ class Activity(Model): You probably want to use the sub-classes and not this class directly. Known sub-classes are: ExecutionActivity, ControlActivity + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] - :param name: Activity name. + :param name: Required. Activity name. :type name: str :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str """ @@ -48,10 +50,10 @@ class Activity(Model): 'type': {'Execution': 'ExecutionActivity', 'Container': 'ControlActivity'} } - def __init__(self, name, additional_properties=None, description=None, depends_on=None): - super(Activity, self).__init__() - self.additional_properties = additional_properties - self.name = name - self.description = description - self.depends_on = depends_on + def __init__(self, **kwargs): + super(Activity, self).__init__(**kwargs) + self.additional_properties = kwargs.get('additional_properties', None) + self.name = kwargs.get('name', None) + self.description = kwargs.get('description', None) + self.depends_on = kwargs.get('depends_on', None) self.type = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/activity_dependency.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/activity_dependency.py index ab346ecbe635..a15b34acc24f 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/activity_dependency.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/activity_dependency.py @@ -15,12 +15,15 @@ class ActivityDependency(Model): """Activity dependency information. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] - :param activity: Activity name. + :param activity: Required. Activity name. :type activity: str - :param dependency_conditions: Match-Condition for the dependency. + :param dependency_conditions: Required. Match-Condition for the + dependency. :type dependency_conditions: list[str or ~azure.mgmt.datafactory.models.DependencyCondition] """ @@ -36,8 +39,8 @@ class ActivityDependency(Model): 'dependency_conditions': {'key': 'dependencyConditions', 'type': '[str]'}, } - def __init__(self, activity, dependency_conditions, additional_properties=None): - super(ActivityDependency, self).__init__() - self.additional_properties = additional_properties - self.activity = activity - self.dependency_conditions = dependency_conditions + def __init__(self, **kwargs): + super(ActivityDependency, self).__init__(**kwargs) + self.additional_properties = kwargs.get('additional_properties', None) + self.activity = kwargs.get('activity', None) + self.dependency_conditions = kwargs.get('dependency_conditions', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/activity_dependency_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/activity_dependency_py3.py new file mode 100644 index 000000000000..2883a81a0adc --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/activity_dependency_py3.py @@ -0,0 +1,46 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class ActivityDependency(Model): + """Activity dependency information. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param activity: Required. Activity name. + :type activity: str + :param dependency_conditions: Required. Match-Condition for the + dependency. + :type dependency_conditions: list[str or + ~azure.mgmt.datafactory.models.DependencyCondition] + """ + + _validation = { + 'activity': {'required': True}, + 'dependency_conditions': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'activity': {'key': 'activity', 'type': 'str'}, + 'dependency_conditions': {'key': 'dependencyConditions', 'type': '[str]'}, + } + + def __init__(self, *, activity: str, dependency_conditions, additional_properties=None, **kwargs) -> None: + super(ActivityDependency, self).__init__(**kwargs) + self.additional_properties = additional_properties + self.activity = activity + self.dependency_conditions = dependency_conditions diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/activity_policy.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/activity_policy.py index c5fd3c861df1..76cb83407fb8 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/activity_policy.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/activity_policy.py @@ -45,10 +45,10 @@ class ActivityPolicy(Model): 'secure_output': {'key': 'secureOutput', 'type': 'bool'}, } - def __init__(self, additional_properties=None, timeout=None, retry=None, retry_interval_in_seconds=None, secure_output=None): - super(ActivityPolicy, self).__init__() - self.additional_properties = additional_properties - self.timeout = timeout - self.retry = retry - self.retry_interval_in_seconds = retry_interval_in_seconds - self.secure_output = secure_output + def __init__(self, **kwargs): + super(ActivityPolicy, self).__init__(**kwargs) + self.additional_properties = kwargs.get('additional_properties', None) + self.timeout = kwargs.get('timeout', None) + self.retry = kwargs.get('retry', None) + self.retry_interval_in_seconds = kwargs.get('retry_interval_in_seconds', None) + self.secure_output = kwargs.get('secure_output', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/activity_policy_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/activity_policy_py3.py new file mode 100644 index 000000000000..0a99fdb4d92d --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/activity_policy_py3.py @@ -0,0 +1,54 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class ActivityPolicy(Model): + """Execution policy for an activity. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param timeout: Specifies the timeout for the activity to run. The default + timeout is 7 days. Type: string (or Expression with resultType string), + pattern: ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type timeout: object + :param retry: Maximum ordinary retry attempts. Default is 0. Type: integer + (or Expression with resultType integer), minimum: 0. + :type retry: object + :param retry_interval_in_seconds: Interval between each retry attempt (in + seconds). The default is 30 sec. + :type retry_interval_in_seconds: int + :param secure_output: When set to true, Output from activity is considered + as secure and will not be logged to monitoring. + :type secure_output: bool + """ + + _validation = { + 'retry_interval_in_seconds': {'maximum': 86400, 'minimum': 30}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'timeout': {'key': 'timeout', 'type': 'object'}, + 'retry': {'key': 'retry', 'type': 'object'}, + 'retry_interval_in_seconds': {'key': 'retryIntervalInSeconds', 'type': 'int'}, + 'secure_output': {'key': 'secureOutput', 'type': 'bool'}, + } + + def __init__(self, *, additional_properties=None, timeout=None, retry=None, retry_interval_in_seconds: int=None, secure_output: bool=None, **kwargs) -> None: + super(ActivityPolicy, self).__init__(**kwargs) + self.additional_properties = additional_properties + self.timeout = timeout + self.retry = retry + self.retry_interval_in_seconds = retry_interval_in_seconds + self.secure_output = secure_output diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/activity_py3.py new file mode 100644 index 000000000000..c1283c260908 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/activity_py3.py @@ -0,0 +1,59 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class Activity(Model): + """A pipeline activity. + + You probably want to use the sub-classes and not this class directly. Known + sub-classes are: ExecutionActivity, ControlActivity + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param name: Required. Activity name. + :type name: str + :param description: Activity description. + :type description: str + :param depends_on: Activity depends on condition. + :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param type: Required. Constant filled by server. + :type type: str + """ + + _validation = { + 'name': {'required': True}, + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'name': {'key': 'name', 'type': 'str'}, + 'description': {'key': 'description', 'type': 'str'}, + 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + _subtype_map = { + 'type': {'Execution': 'ExecutionActivity', 'Container': 'ControlActivity'} + } + + def __init__(self, *, name: str, additional_properties=None, description: str=None, depends_on=None, **kwargs) -> None: + super(Activity, self).__init__(**kwargs) + self.additional_properties = additional_properties + self.name = name + self.description = description + self.depends_on = depends_on + self.type = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/activity_run.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/activity_run.py index 3492b892ef7f..901ffe23cd4e 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/activity_run.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/activity_run.py @@ -84,9 +84,9 @@ class ActivityRun(Model): 'error': {'key': 'error', 'type': 'object'}, } - def __init__(self, additional_properties=None): - super(ActivityRun, self).__init__() - self.additional_properties = additional_properties + def __init__(self, **kwargs): + super(ActivityRun, self).__init__(**kwargs) + self.additional_properties = kwargs.get('additional_properties', None) self.pipeline_name = None self.pipeline_run_id = None self.activity_name = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/activity_run_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/activity_run_py3.py new file mode 100644 index 000000000000..488e822de957 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/activity_run_py3.py @@ -0,0 +1,102 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class ActivityRun(Model): + """Information about an activity run in a pipeline. + + Variables are only populated by the server, and will be ignored when + sending a request. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :ivar pipeline_name: The name of the pipeline. + :vartype pipeline_name: str + :ivar pipeline_run_id: The id of the pipeline run. + :vartype pipeline_run_id: str + :ivar activity_name: The name of the activity. + :vartype activity_name: str + :ivar activity_type: The type of the activity. + :vartype activity_type: str + :ivar activity_run_id: The id of the activity run. + :vartype activity_run_id: str + :ivar linked_service_name: The name of the compute linked service. + :vartype linked_service_name: str + :ivar status: The status of the activity run. + :vartype status: str + :ivar activity_run_start: The start time of the activity run in 'ISO 8601' + format. + :vartype activity_run_start: datetime + :ivar activity_run_end: The end time of the activity run in 'ISO 8601' + format. + :vartype activity_run_end: datetime + :ivar duration_in_ms: The duration of the activity run. + :vartype duration_in_ms: int + :ivar input: The input for the activity. + :vartype input: object + :ivar output: The output for the activity. + :vartype output: object + :ivar error: The error if any from the activity run. + :vartype error: object + """ + + _validation = { + 'pipeline_name': {'readonly': True}, + 'pipeline_run_id': {'readonly': True}, + 'activity_name': {'readonly': True}, + 'activity_type': {'readonly': True}, + 'activity_run_id': {'readonly': True}, + 'linked_service_name': {'readonly': True}, + 'status': {'readonly': True}, + 'activity_run_start': {'readonly': True}, + 'activity_run_end': {'readonly': True}, + 'duration_in_ms': {'readonly': True}, + 'input': {'readonly': True}, + 'output': {'readonly': True}, + 'error': {'readonly': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'pipeline_name': {'key': 'pipelineName', 'type': 'str'}, + 'pipeline_run_id': {'key': 'pipelineRunId', 'type': 'str'}, + 'activity_name': {'key': 'activityName', 'type': 'str'}, + 'activity_type': {'key': 'activityType', 'type': 'str'}, + 'activity_run_id': {'key': 'activityRunId', 'type': 'str'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'str'}, + 'status': {'key': 'status', 'type': 'str'}, + 'activity_run_start': {'key': 'activityRunStart', 'type': 'iso-8601'}, + 'activity_run_end': {'key': 'activityRunEnd', 'type': 'iso-8601'}, + 'duration_in_ms': {'key': 'durationInMs', 'type': 'int'}, + 'input': {'key': 'input', 'type': 'object'}, + 'output': {'key': 'output', 'type': 'object'}, + 'error': {'key': 'error', 'type': 'object'}, + } + + def __init__(self, *, additional_properties=None, **kwargs) -> None: + super(ActivityRun, self).__init__(**kwargs) + self.additional_properties = additional_properties + self.pipeline_name = None + self.pipeline_run_id = None + self.activity_name = None + self.activity_type = None + self.activity_run_id = None + self.linked_service_name = None + self.status = None + self.activity_run_start = None + self.activity_run_end = None + self.duration_in_ms = None + self.input = None + self.output = None + self.error = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_linked_service.py index 30427d0ebbc8..f83ccbfa9568 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_linked_service.py @@ -15,6 +15,8 @@ class AmazonMWSLinkedService(LinkedService): """Amazon Marketplace Web Service linked service. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -29,20 +31,20 @@ class AmazonMWSLinkedService(LinkedService): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param endpoint: The endpoint of the Amazon MWS server, (i.e. + :param endpoint: Required. The endpoint of the Amazon MWS server, (i.e. mws.amazonservices.com) :type endpoint: object - :param marketplace_id: The Amazon Marketplace ID you want to retrieve data - from. To retrive data from multiple Marketplace IDs, seperate them with a - comma (,). (i.e. A2EUQ1WTGCTBG2) + :param marketplace_id: Required. The Amazon Marketplace ID you want to + retrieve data from. To retrive data from multiple Marketplace IDs, + seperate them with a comma (,). (i.e. A2EUQ1WTGCTBG2) :type marketplace_id: object - :param seller_id: The Amazon seller ID. + :param seller_id: Required. The Amazon seller ID. :type seller_id: object :param mws_auth_token: The Amazon MWS authentication token. :type mws_auth_token: ~azure.mgmt.datafactory.models.SecretBase - :param access_key_id: The access key id used to access data. + :param access_key_id: Required. The access key id used to access data. :type access_key_id: object :param secret_key: The secret key used to access data. :type secret_key: ~azure.mgmt.datafactory.models.SecretBase @@ -89,16 +91,16 @@ class AmazonMWSLinkedService(LinkedService): 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, endpoint, marketplace_id, seller_id, access_key_id, additional_properties=None, connect_via=None, description=None, parameters=None, annotations=None, mws_auth_token=None, secret_key=None, use_encrypted_endpoints=None, use_host_verification=None, use_peer_verification=None, encrypted_credential=None): - super(AmazonMWSLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations) - self.endpoint = endpoint - self.marketplace_id = marketplace_id - self.seller_id = seller_id - self.mws_auth_token = mws_auth_token - self.access_key_id = access_key_id - self.secret_key = secret_key - self.use_encrypted_endpoints = use_encrypted_endpoints - self.use_host_verification = use_host_verification - self.use_peer_verification = use_peer_verification - self.encrypted_credential = encrypted_credential + def __init__(self, **kwargs): + super(AmazonMWSLinkedService, self).__init__(**kwargs) + self.endpoint = kwargs.get('endpoint', None) + self.marketplace_id = kwargs.get('marketplace_id', None) + self.seller_id = kwargs.get('seller_id', None) + self.mws_auth_token = kwargs.get('mws_auth_token', None) + self.access_key_id = kwargs.get('access_key_id', None) + self.secret_key = kwargs.get('secret_key', None) + self.use_encrypted_endpoints = kwargs.get('use_encrypted_endpoints', None) + self.use_host_verification = kwargs.get('use_host_verification', None) + self.use_peer_verification = kwargs.get('use_peer_verification', None) + self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'AmazonMWS' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_linked_service_py3.py new file mode 100644 index 000000000000..e55cd710ca7c --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_linked_service_py3.py @@ -0,0 +1,106 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class AmazonMWSLinkedService(LinkedService): + """Amazon Marketplace Web Service linked service. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param endpoint: Required. The endpoint of the Amazon MWS server, (i.e. + mws.amazonservices.com) + :type endpoint: object + :param marketplace_id: Required. The Amazon Marketplace ID you want to + retrieve data from. To retrive data from multiple Marketplace IDs, + seperate them with a comma (,). (i.e. A2EUQ1WTGCTBG2) + :type marketplace_id: object + :param seller_id: Required. The Amazon seller ID. + :type seller_id: object + :param mws_auth_token: The Amazon MWS authentication token. + :type mws_auth_token: ~azure.mgmt.datafactory.models.SecretBase + :param access_key_id: Required. The access key id used to access data. + :type access_key_id: object + :param secret_key: The secret key used to access data. + :type secret_key: ~azure.mgmt.datafactory.models.SecretBase + :param use_encrypted_endpoints: Specifies whether the data source + endpoints are encrypted using HTTPS. The default value is true. + :type use_encrypted_endpoints: object + :param use_host_verification: Specifies whether to require the host name + in the server's certificate to match the host name of the server when + connecting over SSL. The default value is true. + :type use_host_verification: object + :param use_peer_verification: Specifies whether to verify the identity of + the server when connecting over SSL. The default value is true. + :type use_peer_verification: object + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'endpoint': {'required': True}, + 'marketplace_id': {'required': True}, + 'seller_id': {'required': True}, + 'access_key_id': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'endpoint': {'key': 'typeProperties.endpoint', 'type': 'object'}, + 'marketplace_id': {'key': 'typeProperties.marketplaceID', 'type': 'object'}, + 'seller_id': {'key': 'typeProperties.sellerID', 'type': 'object'}, + 'mws_auth_token': {'key': 'typeProperties.mwsAuthToken', 'type': 'SecretBase'}, + 'access_key_id': {'key': 'typeProperties.accessKeyId', 'type': 'object'}, + 'secret_key': {'key': 'typeProperties.secretKey', 'type': 'SecretBase'}, + 'use_encrypted_endpoints': {'key': 'typeProperties.useEncryptedEndpoints', 'type': 'object'}, + 'use_host_verification': {'key': 'typeProperties.useHostVerification', 'type': 'object'}, + 'use_peer_verification': {'key': 'typeProperties.usePeerVerification', 'type': 'object'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, *, endpoint, marketplace_id, seller_id, access_key_id, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, mws_auth_token=None, secret_key=None, use_encrypted_endpoints=None, use_host_verification=None, use_peer_verification=None, encrypted_credential=None, **kwargs) -> None: + super(AmazonMWSLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.endpoint = endpoint + self.marketplace_id = marketplace_id + self.seller_id = seller_id + self.mws_auth_token = mws_auth_token + self.access_key_id = access_key_id + self.secret_key = secret_key + self.use_encrypted_endpoints = use_encrypted_endpoints + self.use_host_verification = use_host_verification + self.use_peer_verification = use_peer_verification + self.encrypted_credential = encrypted_credential + self.type = 'AmazonMWS' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_object_dataset.py index 357fa0a5fdf5..7062a5534fca 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_object_dataset.py @@ -15,6 +15,8 @@ class AmazonMWSObjectDataset(Dataset): """Amazon Marketplace Web Service dataset. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -23,7 +25,7 @@ class AmazonMWSObjectDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object - :param linked_service_name: Linked service reference. + :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference :param parameters: Parameters for dataset. @@ -32,7 +34,7 @@ class AmazonMWSObjectDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str """ @@ -41,6 +43,16 @@ class AmazonMWSObjectDataset(Dataset): 'type': {'required': True}, } - def __init__(self, linked_service_name, additional_properties=None, description=None, structure=None, parameters=None, annotations=None): - super(AmazonMWSObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations) + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, **kwargs): + super(AmazonMWSObjectDataset, self).__init__(**kwargs) self.type = 'AmazonMWSObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_object_dataset_py3.py new file mode 100644 index 000000000000..980bc9445650 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_object_dataset_py3.py @@ -0,0 +1,58 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class AmazonMWSObjectDataset(Dataset): + """Amazon Marketplace Web Service dataset. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Required. Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: + super(AmazonMWSObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + self.type = 'AmazonMWSObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_source.py index 3b27d7b15b61..1cabba2201c7 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_source.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_source.py @@ -15,6 +15,8 @@ class AmazonMWSSource(CopySource): """A copy activity Amazon Marketplace Web Service source. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -25,7 +27,7 @@ class AmazonMWSSource(CopySource): with resultType string), pattern: ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type source_retry_wait: object - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). @@ -44,7 +46,7 @@ class AmazonMWSSource(CopySource): 'query': {'key': 'query', 'type': 'object'}, } - def __init__(self, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None): - super(AmazonMWSSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait) - self.query = query + def __init__(self, **kwargs): + super(AmazonMWSSource, self).__init__(**kwargs) + self.query = kwargs.get('query', None) self.type = 'AmazonMWSSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_source_py3.py new file mode 100644 index 000000000000..524a047cc6fc --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_source_py3.py @@ -0,0 +1,52 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class AmazonMWSSource(CopySource): + """A copy activity Amazon Marketplace Web Service source. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Required. Constant filled by server. + :type type: str + :param query: A query to retrieve data from source. Type: string (or + Expression with resultType string). + :type query: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'query': {'key': 'query', 'type': 'object'}, + } + + def __init__(self, *, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None, **kwargs) -> None: + super(AmazonMWSSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, **kwargs) + self.query = query + self.type = 'AmazonMWSSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_redshift_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_redshift_linked_service.py index 8040c2f3f095..a85e73b458ae 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_redshift_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_redshift_linked_service.py @@ -15,6 +15,8 @@ class AmazonRedshiftLinkedService(LinkedService): """Linked service for Amazon Redshift. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -29,18 +31,18 @@ class AmazonRedshiftLinkedService(LinkedService): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param server: The name of the Amazon Redshift server. Type: string (or - Expression with resultType string). + :param server: Required. The name of the Amazon Redshift server. Type: + string (or Expression with resultType string). :type server: object :param username: The username of the Amazon Redshift source. Type: string (or Expression with resultType string). :type username: object :param password: The password of the Amazon Redshift source. :type password: ~azure.mgmt.datafactory.models.SecretBase - :param database: The database name of the Amazon Redshift source. Type: - string (or Expression with resultType string). + :param database: Required. The database name of the Amazon Redshift + source. Type: string (or Expression with resultType string). :type database: object :param port: The TCP port number that the Amazon Redshift server uses to listen for client connections. The default value is 5439. Type: integer @@ -73,12 +75,12 @@ class AmazonRedshiftLinkedService(LinkedService): 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, server, database, additional_properties=None, connect_via=None, description=None, parameters=None, annotations=None, username=None, password=None, port=None, encrypted_credential=None): - super(AmazonRedshiftLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations) - self.server = server - self.username = username - self.password = password - self.database = database - self.port = port - self.encrypted_credential = encrypted_credential + def __init__(self, **kwargs): + super(AmazonRedshiftLinkedService, self).__init__(**kwargs) + self.server = kwargs.get('server', None) + self.username = kwargs.get('username', None) + self.password = kwargs.get('password', None) + self.database = kwargs.get('database', None) + self.port = kwargs.get('port', None) + self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'AmazonRedshift' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_redshift_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_redshift_linked_service_py3.py new file mode 100644 index 000000000000..1d6e7a098786 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_redshift_linked_service_py3.py @@ -0,0 +1,86 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class AmazonRedshiftLinkedService(LinkedService): + """Linked service for Amazon Redshift. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param server: Required. The name of the Amazon Redshift server. Type: + string (or Expression with resultType string). + :type server: object + :param username: The username of the Amazon Redshift source. Type: string + (or Expression with resultType string). + :type username: object + :param password: The password of the Amazon Redshift source. + :type password: ~azure.mgmt.datafactory.models.SecretBase + :param database: Required. The database name of the Amazon Redshift + source. Type: string (or Expression with resultType string). + :type database: object + :param port: The TCP port number that the Amazon Redshift server uses to + listen for client connections. The default value is 5439. Type: integer + (or Expression with resultType integer). + :type port: object + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'server': {'required': True}, + 'database': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'server': {'key': 'typeProperties.server', 'type': 'object'}, + 'username': {'key': 'typeProperties.username', 'type': 'object'}, + 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, + 'database': {'key': 'typeProperties.database', 'type': 'object'}, + 'port': {'key': 'typeProperties.port', 'type': 'object'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, *, server, database, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, username=None, password=None, port=None, encrypted_credential=None, **kwargs) -> None: + super(AmazonRedshiftLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.server = server + self.username = username + self.password = password + self.database = database + self.port = port + self.encrypted_credential = encrypted_credential + self.type = 'AmazonRedshift' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_redshift_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_redshift_source.py index d866d7ca5b6a..0fa9a82ff9db 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_redshift_source.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_redshift_source.py @@ -15,6 +15,8 @@ class AmazonRedshiftSource(CopySource): """A copy activity source for Amazon Redshift Source. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -25,7 +27,7 @@ class AmazonRedshiftSource(CopySource): with resultType string), pattern: ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type source_retry_wait: object - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param query: Database query. Type: string (or Expression with resultType string). @@ -51,8 +53,8 @@ class AmazonRedshiftSource(CopySource): 'redshift_unload_settings': {'key': 'redshiftUnloadSettings', 'type': 'RedshiftUnloadSettings'}, } - def __init__(self, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None, redshift_unload_settings=None): - super(AmazonRedshiftSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait) - self.query = query - self.redshift_unload_settings = redshift_unload_settings + def __init__(self, **kwargs): + super(AmazonRedshiftSource, self).__init__(**kwargs) + self.query = kwargs.get('query', None) + self.redshift_unload_settings = kwargs.get('redshift_unload_settings', None) self.type = 'AmazonRedshiftSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_redshift_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_redshift_source_py3.py new file mode 100644 index 000000000000..de43ec3bdff0 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_redshift_source_py3.py @@ -0,0 +1,60 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class AmazonRedshiftSource(CopySource): + """A copy activity source for Amazon Redshift Source. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Required. Constant filled by server. + :type type: str + :param query: Database query. Type: string (or Expression with resultType + string). + :type query: object + :param redshift_unload_settings: The Amazon S3 settings needed for the + interim Amazon S3 when copying from Amazon Redshift with unload. With + this, data from Amazon Redshift source will be unloaded into S3 first and + then copied into the targeted sink from the interim S3. + :type redshift_unload_settings: + ~azure.mgmt.datafactory.models.RedshiftUnloadSettings + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'query': {'key': 'query', 'type': 'object'}, + 'redshift_unload_settings': {'key': 'redshiftUnloadSettings', 'type': 'RedshiftUnloadSettings'}, + } + + def __init__(self, *, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None, redshift_unload_settings=None, **kwargs) -> None: + super(AmazonRedshiftSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, **kwargs) + self.query = query + self.redshift_unload_settings = redshift_unload_settings + self.type = 'AmazonRedshiftSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_s3_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_s3_dataset.py index 97c33747b546..f2bd2a93dff8 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_s3_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_s3_dataset.py @@ -15,6 +15,8 @@ class AmazonS3Dataset(Dataset): """A single Amazon Simple Storage Service (S3) object or a set of S3 objects. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -23,7 +25,7 @@ class AmazonS3Dataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object - :param linked_service_name: Linked service reference. + :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference :param parameters: Parameters for dataset. @@ -32,10 +34,10 @@ class AmazonS3Dataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param bucket_name: The name of the Amazon S3 bucket. Type: string (or - Expression with resultType string). + :param bucket_name: Required. The name of the Amazon S3 bucket. Type: + string (or Expression with resultType string). :type bucket_name: object :param key: The key of the Amazon S3 object. Type: string (or Expression with resultType string). @@ -75,12 +77,12 @@ class AmazonS3Dataset(Dataset): 'compression': {'key': 'typeProperties.compression', 'type': 'DatasetCompression'}, } - def __init__(self, linked_service_name, bucket_name, additional_properties=None, description=None, structure=None, parameters=None, annotations=None, key=None, prefix=None, version=None, format=None, compression=None): - super(AmazonS3Dataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations) - self.bucket_name = bucket_name - self.key = key - self.prefix = prefix - self.version = version - self.format = format - self.compression = compression + def __init__(self, **kwargs): + super(AmazonS3Dataset, self).__init__(**kwargs) + self.bucket_name = kwargs.get('bucket_name', None) + self.key = kwargs.get('key', None) + self.prefix = kwargs.get('prefix', None) + self.version = kwargs.get('version', None) + self.format = kwargs.get('format', None) + self.compression = kwargs.get('compression', None) self.type = 'AmazonS3Object' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_s3_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_s3_dataset_py3.py new file mode 100644 index 000000000000..3ca9f3ca5d86 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_s3_dataset_py3.py @@ -0,0 +1,88 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class AmazonS3Dataset(Dataset): + """A single Amazon Simple Storage Service (S3) object or a set of S3 objects. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Required. Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param bucket_name: Required. The name of the Amazon S3 bucket. Type: + string (or Expression with resultType string). + :type bucket_name: object + :param key: The key of the Amazon S3 object. Type: string (or Expression + with resultType string). + :type key: object + :param prefix: The prefix filter for the S3 object name. Type: string (or + Expression with resultType string). + :type prefix: object + :param version: The version for the S3 object. Type: string (or Expression + with resultType string). + :type version: object + :param format: The format of files. + :type format: ~azure.mgmt.datafactory.models.DatasetStorageFormat + :param compression: The data compression method used for the Amazon S3 + object. + :type compression: ~azure.mgmt.datafactory.models.DatasetCompression + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + 'bucket_name': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'bucket_name': {'key': 'typeProperties.bucketName', 'type': 'object'}, + 'key': {'key': 'typeProperties.key', 'type': 'object'}, + 'prefix': {'key': 'typeProperties.prefix', 'type': 'object'}, + 'version': {'key': 'typeProperties.version', 'type': 'object'}, + 'format': {'key': 'typeProperties.format', 'type': 'DatasetStorageFormat'}, + 'compression': {'key': 'typeProperties.compression', 'type': 'DatasetCompression'}, + } + + def __init__(self, *, linked_service_name, bucket_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, key=None, prefix=None, version=None, format=None, compression=None, **kwargs) -> None: + super(AmazonS3Dataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + self.bucket_name = bucket_name + self.key = key + self.prefix = prefix + self.version = version + self.format = format + self.compression = compression + self.type = 'AmazonS3Object' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_s3_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_s3_linked_service.py index a3636590ab9b..c9ff7261d915 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_s3_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_s3_linked_service.py @@ -15,6 +15,8 @@ class AmazonS3LinkedService(LinkedService): """Linked service for Amazon S3. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -29,7 +31,7 @@ class AmazonS3LinkedService(LinkedService): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param access_key_id: The access key identifier of the Amazon S3 Identity and Access Management (IAM) user. Type: string (or Expression with @@ -60,9 +62,9 @@ class AmazonS3LinkedService(LinkedService): 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, additional_properties=None, connect_via=None, description=None, parameters=None, annotations=None, access_key_id=None, secret_access_key=None, encrypted_credential=None): - super(AmazonS3LinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations) - self.access_key_id = access_key_id - self.secret_access_key = secret_access_key - self.encrypted_credential = encrypted_credential + def __init__(self, **kwargs): + super(AmazonS3LinkedService, self).__init__(**kwargs) + self.access_key_id = kwargs.get('access_key_id', None) + self.secret_access_key = kwargs.get('secret_access_key', None) + self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'AmazonS3' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_s3_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_s3_linked_service_py3.py new file mode 100644 index 000000000000..fca34c82fb88 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_s3_linked_service_py3.py @@ -0,0 +1,70 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class AmazonS3LinkedService(LinkedService): + """Linked service for Amazon S3. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param access_key_id: The access key identifier of the Amazon S3 Identity + and Access Management (IAM) user. Type: string (or Expression with + resultType string). + :type access_key_id: object + :param secret_access_key: The secret access key of the Amazon S3 Identity + and Access Management (IAM) user. + :type secret_access_key: ~azure.mgmt.datafactory.models.SecretBase + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'access_key_id': {'key': 'typeProperties.accessKeyId', 'type': 'object'}, + 'secret_access_key': {'key': 'typeProperties.secretAccessKey', 'type': 'SecretBase'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, *, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, access_key_id=None, secret_access_key=None, encrypted_credential=None, **kwargs) -> None: + super(AmazonS3LinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.access_key_id = access_key_id + self.secret_access_key = secret_access_key + self.encrypted_credential = encrypted_credential + self.type = 'AmazonS3' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/avro_format.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/avro_format.py index 0a015516867e..f0346a76080c 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/avro_format.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/avro_format.py @@ -15,6 +15,8 @@ class AvroFormat(DatasetStorageFormat): """The data stored in Avro format. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -24,7 +26,7 @@ class AvroFormat(DatasetStorageFormat): :param deserializer: Deserializer. Type: string (or Expression with resultType string). :type deserializer: object - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str """ @@ -32,6 +34,13 @@ class AvroFormat(DatasetStorageFormat): 'type': {'required': True}, } - def __init__(self, additional_properties=None, serializer=None, deserializer=None): - super(AvroFormat, self).__init__(additional_properties=additional_properties, serializer=serializer, deserializer=deserializer) + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'serializer': {'key': 'serializer', 'type': 'object'}, + 'deserializer': {'key': 'deserializer', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, **kwargs): + super(AvroFormat, self).__init__(**kwargs) self.type = 'AvroFormat' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/avro_format_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/avro_format_py3.py new file mode 100644 index 000000000000..eb1f141852a9 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/avro_format_py3.py @@ -0,0 +1,46 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset_storage_format import DatasetStorageFormat + + +class AvroFormat(DatasetStorageFormat): + """The data stored in Avro format. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param serializer: Serializer. Type: string (or Expression with resultType + string). + :type serializer: object + :param deserializer: Deserializer. Type: string (or Expression with + resultType string). + :type deserializer: object + :param type: Required. Constant filled by server. + :type type: str + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'serializer': {'key': 'serializer', 'type': 'object'}, + 'deserializer': {'key': 'deserializer', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, *, additional_properties=None, serializer=None, deserializer=None, **kwargs) -> None: + super(AvroFormat, self).__init__(additional_properties=additional_properties, serializer=serializer, deserializer=deserializer, **kwargs) + self.type = 'AvroFormat' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_batch_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_batch_linked_service.py index e50696a857b2..2fcf33e8d0c8 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_batch_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_batch_linked_service.py @@ -15,6 +15,8 @@ class AzureBatchLinkedService(LinkedService): """Azure Batch linked service. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -29,20 +31,21 @@ class AzureBatchLinkedService(LinkedService): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param account_name: The Azure Batch account name. Type: string (or - Expression with resultType string). + :param account_name: Required. The Azure Batch account name. Type: string + (or Expression with resultType string). :type account_name: object :param access_key: The Azure Batch account access key. :type access_key: ~azure.mgmt.datafactory.models.SecretBase - :param batch_uri: The Azure Batch URI. Type: string (or Expression with - resultType string). + :param batch_uri: Required. The Azure Batch URI. Type: string (or + Expression with resultType string). :type batch_uri: object - :param pool_name: The Azure Batch pool name. Type: string (or Expression - with resultType string). + :param pool_name: Required. The Azure Batch pool name. Type: string (or + Expression with resultType string). :type pool_name: object - :param linked_service_name: The Azure Storage linked service reference. + :param linked_service_name: Required. The Azure Storage linked service + reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference :param encrypted_credential: The encrypted credential used for @@ -74,12 +77,12 @@ class AzureBatchLinkedService(LinkedService): 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, account_name, batch_uri, pool_name, linked_service_name, additional_properties=None, connect_via=None, description=None, parameters=None, annotations=None, access_key=None, encrypted_credential=None): - super(AzureBatchLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations) - self.account_name = account_name - self.access_key = access_key - self.batch_uri = batch_uri - self.pool_name = pool_name - self.linked_service_name = linked_service_name - self.encrypted_credential = encrypted_credential + def __init__(self, **kwargs): + super(AzureBatchLinkedService, self).__init__(**kwargs) + self.account_name = kwargs.get('account_name', None) + self.access_key = kwargs.get('access_key', None) + self.batch_uri = kwargs.get('batch_uri', None) + self.pool_name = kwargs.get('pool_name', None) + self.linked_service_name = kwargs.get('linked_service_name', None) + self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'AzureBatch' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_batch_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_batch_linked_service_py3.py new file mode 100644 index 000000000000..928dfecf2664 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_batch_linked_service_py3.py @@ -0,0 +1,88 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class AzureBatchLinkedService(LinkedService): + """Azure Batch linked service. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param account_name: Required. The Azure Batch account name. Type: string + (or Expression with resultType string). + :type account_name: object + :param access_key: The Azure Batch account access key. + :type access_key: ~azure.mgmt.datafactory.models.SecretBase + :param batch_uri: Required. The Azure Batch URI. Type: string (or + Expression with resultType string). + :type batch_uri: object + :param pool_name: Required. The Azure Batch pool name. Type: string (or + Expression with resultType string). + :type pool_name: object + :param linked_service_name: Required. The Azure Storage linked service + reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'account_name': {'required': True}, + 'batch_uri': {'required': True}, + 'pool_name': {'required': True}, + 'linked_service_name': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'account_name': {'key': 'typeProperties.accountName', 'type': 'object'}, + 'access_key': {'key': 'typeProperties.accessKey', 'type': 'SecretBase'}, + 'batch_uri': {'key': 'typeProperties.batchUri', 'type': 'object'}, + 'pool_name': {'key': 'typeProperties.poolName', 'type': 'object'}, + 'linked_service_name': {'key': 'typeProperties.linkedServiceName', 'type': 'LinkedServiceReference'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, *, account_name, batch_uri, pool_name, linked_service_name, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, access_key=None, encrypted_credential=None, **kwargs) -> None: + super(AzureBatchLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.account_name = account_name + self.access_key = access_key + self.batch_uri = batch_uri + self.pool_name = pool_name + self.linked_service_name = linked_service_name + self.encrypted_credential = encrypted_credential + self.type = 'AzureBatch' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_dataset.py index 99a2e5f0c7bc..fa15f8a9ad1e 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_dataset.py @@ -15,6 +15,8 @@ class AzureBlobDataset(Dataset): """The Azure Blob storage. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -23,7 +25,7 @@ class AzureBlobDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object - :param linked_service_name: Linked service reference. + :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference :param parameters: Parameters for dataset. @@ -32,7 +34,7 @@ class AzureBlobDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param folder_path: The path of the Azure Blob storage. Type: string (or Expression with resultType string). @@ -69,11 +71,11 @@ class AzureBlobDataset(Dataset): 'compression': {'key': 'typeProperties.compression', 'type': 'DatasetCompression'}, } - def __init__(self, linked_service_name, additional_properties=None, description=None, structure=None, parameters=None, annotations=None, folder_path=None, table_root_location=None, file_name=None, format=None, compression=None): - super(AzureBlobDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations) - self.folder_path = folder_path - self.table_root_location = table_root_location - self.file_name = file_name - self.format = format - self.compression = compression + def __init__(self, **kwargs): + super(AzureBlobDataset, self).__init__(**kwargs) + self.folder_path = kwargs.get('folder_path', None) + self.table_root_location = kwargs.get('table_root_location', None) + self.file_name = kwargs.get('file_name', None) + self.format = kwargs.get('format', None) + self.compression = kwargs.get('compression', None) self.type = 'AzureBlob' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_dataset_py3.py new file mode 100644 index 000000000000..fb3265a92c6f --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_dataset_py3.py @@ -0,0 +1,81 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class AzureBlobDataset(Dataset): + """The Azure Blob storage. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Required. Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param folder_path: The path of the Azure Blob storage. Type: string (or + Expression with resultType string). + :type folder_path: object + :param table_root_location: The root of blob path. Type: string (or + Expression with resultType string). + :type table_root_location: object + :param file_name: The name of the Azure Blob. Type: string (or Expression + with resultType string). + :type file_name: object + :param format: The format of the Azure Blob storage. + :type format: ~azure.mgmt.datafactory.models.DatasetStorageFormat + :param compression: The data compression method used for the blob storage. + :type compression: ~azure.mgmt.datafactory.models.DatasetCompression + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'folder_path': {'key': 'typeProperties.folderPath', 'type': 'object'}, + 'table_root_location': {'key': 'typeProperties.tableRootLocation', 'type': 'object'}, + 'file_name': {'key': 'typeProperties.fileName', 'type': 'object'}, + 'format': {'key': 'typeProperties.format', 'type': 'DatasetStorageFormat'}, + 'compression': {'key': 'typeProperties.compression', 'type': 'DatasetCompression'}, + } + + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder_path=None, table_root_location=None, file_name=None, format=None, compression=None, **kwargs) -> None: + super(AzureBlobDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + self.folder_path = folder_path + self.table_root_location = table_root_location + self.file_name = file_name + self.format = format + self.compression = compression + self.type = 'AzureBlob' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_analytics_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_analytics_linked_service.py index a25d07a2923a..73ec2b6f9de9 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_analytics_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_analytics_linked_service.py @@ -15,6 +15,8 @@ class AzureDataLakeAnalyticsLinkedService(LinkedService): """Azure Data Lake Analytics linked service. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -29,10 +31,10 @@ class AzureDataLakeAnalyticsLinkedService(LinkedService): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param account_name: The Azure Data Lake Analytics account name. Type: - string (or Expression with resultType string). + :param account_name: Required. The Azure Data Lake Analytics account name. + Type: string (or Expression with resultType string). :type account_name: object :param service_principal_id: The ID of the application used to authenticate against the Azure Data Lake Analytics account. Type: string @@ -41,8 +43,8 @@ class AzureDataLakeAnalyticsLinkedService(LinkedService): :param service_principal_key: The Key of the application used to authenticate against the Azure Data Lake Analytics account. :type service_principal_key: ~azure.mgmt.datafactory.models.SecretBase - :param tenant: The name or ID of the tenant to which the service principal - belongs. Type: string (or Expression with resultType string). + :param tenant: Required. The name or ID of the tenant to which the service + principal belongs. Type: string (or Expression with resultType string). :type tenant: object :param subscription_id: Data Lake Analytics account subscription ID (if different from Data Factory account). Type: string (or Expression with @@ -84,14 +86,14 @@ class AzureDataLakeAnalyticsLinkedService(LinkedService): 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, account_name, tenant, additional_properties=None, connect_via=None, description=None, parameters=None, annotations=None, service_principal_id=None, service_principal_key=None, subscription_id=None, resource_group_name=None, data_lake_analytics_uri=None, encrypted_credential=None): - super(AzureDataLakeAnalyticsLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations) - self.account_name = account_name - self.service_principal_id = service_principal_id - self.service_principal_key = service_principal_key - self.tenant = tenant - self.subscription_id = subscription_id - self.resource_group_name = resource_group_name - self.data_lake_analytics_uri = data_lake_analytics_uri - self.encrypted_credential = encrypted_credential + def __init__(self, **kwargs): + super(AzureDataLakeAnalyticsLinkedService, self).__init__(**kwargs) + self.account_name = kwargs.get('account_name', None) + self.service_principal_id = kwargs.get('service_principal_id', None) + self.service_principal_key = kwargs.get('service_principal_key', None) + self.tenant = kwargs.get('tenant', None) + self.subscription_id = kwargs.get('subscription_id', None) + self.resource_group_name = kwargs.get('resource_group_name', None) + self.data_lake_analytics_uri = kwargs.get('data_lake_analytics_uri', None) + self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'AzureDataLakeAnalytics' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_analytics_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_analytics_linked_service_py3.py new file mode 100644 index 000000000000..f5b32873f157 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_analytics_linked_service_py3.py @@ -0,0 +1,99 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class AzureDataLakeAnalyticsLinkedService(LinkedService): + """Azure Data Lake Analytics linked service. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param account_name: Required. The Azure Data Lake Analytics account name. + Type: string (or Expression with resultType string). + :type account_name: object + :param service_principal_id: The ID of the application used to + authenticate against the Azure Data Lake Analytics account. Type: string + (or Expression with resultType string). + :type service_principal_id: object + :param service_principal_key: The Key of the application used to + authenticate against the Azure Data Lake Analytics account. + :type service_principal_key: ~azure.mgmt.datafactory.models.SecretBase + :param tenant: Required. The name or ID of the tenant to which the service + principal belongs. Type: string (or Expression with resultType string). + :type tenant: object + :param subscription_id: Data Lake Analytics account subscription ID (if + different from Data Factory account). Type: string (or Expression with + resultType string). + :type subscription_id: object + :param resource_group_name: Data Lake Analytics account resource group + name (if different from Data Factory account). Type: string (or Expression + with resultType string). + :type resource_group_name: object + :param data_lake_analytics_uri: Azure Data Lake Analytics URI Type: string + (or Expression with resultType string). + :type data_lake_analytics_uri: object + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'account_name': {'required': True}, + 'tenant': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'account_name': {'key': 'typeProperties.accountName', 'type': 'object'}, + 'service_principal_id': {'key': 'typeProperties.servicePrincipalId', 'type': 'object'}, + 'service_principal_key': {'key': 'typeProperties.servicePrincipalKey', 'type': 'SecretBase'}, + 'tenant': {'key': 'typeProperties.tenant', 'type': 'object'}, + 'subscription_id': {'key': 'typeProperties.subscriptionId', 'type': 'object'}, + 'resource_group_name': {'key': 'typeProperties.resourceGroupName', 'type': 'object'}, + 'data_lake_analytics_uri': {'key': 'typeProperties.dataLakeAnalyticsUri', 'type': 'object'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, *, account_name, tenant, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, service_principal_id=None, service_principal_key=None, subscription_id=None, resource_group_name=None, data_lake_analytics_uri=None, encrypted_credential=None, **kwargs) -> None: + super(AzureDataLakeAnalyticsLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.account_name = account_name + self.service_principal_id = service_principal_id + self.service_principal_key = service_principal_key + self.tenant = tenant + self.subscription_id = subscription_id + self.resource_group_name = resource_group_name + self.data_lake_analytics_uri = data_lake_analytics_uri + self.encrypted_credential = encrypted_credential + self.type = 'AzureDataLakeAnalytics' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_dataset.py index 0df9746c81b9..be589a7803cc 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_dataset.py @@ -15,6 +15,8 @@ class AzureDataLakeStoreDataset(Dataset): """Azure Data Lake Store dataset. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -23,7 +25,7 @@ class AzureDataLakeStoreDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object - :param linked_service_name: Linked service reference. + :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference :param parameters: Parameters for dataset. @@ -32,10 +34,10 @@ class AzureDataLakeStoreDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param folder_path: Path to the folder in the Azure Data Lake Store. Type: - string (or Expression with resultType string). + :param folder_path: Required. Path to the folder in the Azure Data Lake + Store. Type: string (or Expression with resultType string). :type folder_path: object :param file_name: The name of the file in the Azure Data Lake Store. Type: string (or Expression with resultType string). @@ -67,10 +69,10 @@ class AzureDataLakeStoreDataset(Dataset): 'compression': {'key': 'typeProperties.compression', 'type': 'DatasetCompression'}, } - def __init__(self, linked_service_name, folder_path, additional_properties=None, description=None, structure=None, parameters=None, annotations=None, file_name=None, format=None, compression=None): - super(AzureDataLakeStoreDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations) - self.folder_path = folder_path - self.file_name = file_name - self.format = format - self.compression = compression + def __init__(self, **kwargs): + super(AzureDataLakeStoreDataset, self).__init__(**kwargs) + self.folder_path = kwargs.get('folder_path', None) + self.file_name = kwargs.get('file_name', None) + self.format = kwargs.get('format', None) + self.compression = kwargs.get('compression', None) self.type = 'AzureDataLakeStoreFile' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_dataset_py3.py new file mode 100644 index 000000000000..0fcf6571644f --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_dataset_py3.py @@ -0,0 +1,78 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class AzureDataLakeStoreDataset(Dataset): + """Azure Data Lake Store dataset. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Required. Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param folder_path: Required. Path to the folder in the Azure Data Lake + Store. Type: string (or Expression with resultType string). + :type folder_path: object + :param file_name: The name of the file in the Azure Data Lake Store. Type: + string (or Expression with resultType string). + :type file_name: object + :param format: The format of the Data Lake Store. + :type format: ~azure.mgmt.datafactory.models.DatasetStorageFormat + :param compression: The data compression method used for the item(s) in + the Azure Data Lake Store. + :type compression: ~azure.mgmt.datafactory.models.DatasetCompression + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + 'folder_path': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'folder_path': {'key': 'typeProperties.folderPath', 'type': 'object'}, + 'file_name': {'key': 'typeProperties.fileName', 'type': 'object'}, + 'format': {'key': 'typeProperties.format', 'type': 'DatasetStorageFormat'}, + 'compression': {'key': 'typeProperties.compression', 'type': 'DatasetCompression'}, + } + + def __init__(self, *, linked_service_name, folder_path, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, file_name=None, format=None, compression=None, **kwargs) -> None: + super(AzureDataLakeStoreDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + self.folder_path = folder_path + self.file_name = file_name + self.format = format + self.compression = compression + self.type = 'AzureDataLakeStoreFile' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_linked_service.py index d8d0d1c19765..0c39866887ef 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_linked_service.py @@ -15,6 +15,8 @@ class AzureDataLakeStoreLinkedService(LinkedService): """Azure Data Lake Store linked service. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -29,10 +31,10 @@ class AzureDataLakeStoreLinkedService(LinkedService): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param data_lake_store_uri: Data Lake Store service URI. Type: string (or - Expression with resultType string). + :param data_lake_store_uri: Required. Data Lake Store service URI. Type: + string (or Expression with resultType string). :type data_lake_store_uri: object :param service_principal_id: The ID of the application used to authenticate against the Azure Data Lake Store account. Type: string (or @@ -83,14 +85,14 @@ class AzureDataLakeStoreLinkedService(LinkedService): 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, data_lake_store_uri, additional_properties=None, connect_via=None, description=None, parameters=None, annotations=None, service_principal_id=None, service_principal_key=None, tenant=None, account_name=None, subscription_id=None, resource_group_name=None, encrypted_credential=None): - super(AzureDataLakeStoreLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations) - self.data_lake_store_uri = data_lake_store_uri - self.service_principal_id = service_principal_id - self.service_principal_key = service_principal_key - self.tenant = tenant - self.account_name = account_name - self.subscription_id = subscription_id - self.resource_group_name = resource_group_name - self.encrypted_credential = encrypted_credential + def __init__(self, **kwargs): + super(AzureDataLakeStoreLinkedService, self).__init__(**kwargs) + self.data_lake_store_uri = kwargs.get('data_lake_store_uri', None) + self.service_principal_id = kwargs.get('service_principal_id', None) + self.service_principal_key = kwargs.get('service_principal_key', None) + self.tenant = kwargs.get('tenant', None) + self.account_name = kwargs.get('account_name', None) + self.subscription_id = kwargs.get('subscription_id', None) + self.resource_group_name = kwargs.get('resource_group_name', None) + self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'AzureDataLakeStore' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_linked_service_py3.py new file mode 100644 index 000000000000..3244f5d35bb6 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_linked_service_py3.py @@ -0,0 +1,98 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class AzureDataLakeStoreLinkedService(LinkedService): + """Azure Data Lake Store linked service. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param data_lake_store_uri: Required. Data Lake Store service URI. Type: + string (or Expression with resultType string). + :type data_lake_store_uri: object + :param service_principal_id: The ID of the application used to + authenticate against the Azure Data Lake Store account. Type: string (or + Expression with resultType string). + :type service_principal_id: object + :param service_principal_key: The Key of the application used to + authenticate against the Azure Data Lake Store account. + :type service_principal_key: ~azure.mgmt.datafactory.models.SecretBase + :param tenant: The name or ID of the tenant to which the service principal + belongs. Type: string (or Expression with resultType string). + :type tenant: object + :param account_name: Data Lake Store account name. Type: string (or + Expression with resultType string). + :type account_name: object + :param subscription_id: Data Lake Store account subscription ID (if + different from Data Factory account). Type: string (or Expression with + resultType string). + :type subscription_id: object + :param resource_group_name: Data Lake Store account resource group name + (if different from Data Factory account). Type: string (or Expression with + resultType string). + :type resource_group_name: object + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'data_lake_store_uri': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'data_lake_store_uri': {'key': 'typeProperties.dataLakeStoreUri', 'type': 'object'}, + 'service_principal_id': {'key': 'typeProperties.servicePrincipalId', 'type': 'object'}, + 'service_principal_key': {'key': 'typeProperties.servicePrincipalKey', 'type': 'SecretBase'}, + 'tenant': {'key': 'typeProperties.tenant', 'type': 'object'}, + 'account_name': {'key': 'typeProperties.accountName', 'type': 'object'}, + 'subscription_id': {'key': 'typeProperties.subscriptionId', 'type': 'object'}, + 'resource_group_name': {'key': 'typeProperties.resourceGroupName', 'type': 'object'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, *, data_lake_store_uri, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, service_principal_id=None, service_principal_key=None, tenant=None, account_name=None, subscription_id=None, resource_group_name=None, encrypted_credential=None, **kwargs) -> None: + super(AzureDataLakeStoreLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.data_lake_store_uri = data_lake_store_uri + self.service_principal_id = service_principal_id + self.service_principal_key = service_principal_key + self.tenant = tenant + self.account_name = account_name + self.subscription_id = subscription_id + self.resource_group_name = resource_group_name + self.encrypted_credential = encrypted_credential + self.type = 'AzureDataLakeStore' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_sink.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_sink.py index 2dc6ff9f9f80..ceaabf438097 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_sink.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_sink.py @@ -15,6 +15,8 @@ class AzureDataLakeStoreSink(CopySink): """A copy activity Azure Data Lake Store sink. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -32,7 +34,7 @@ class AzureDataLakeStoreSink(CopySink): resultType string), pattern: ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type sink_retry_wait: object - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param copy_behavior: The type of copy behavior for copy sink. Possible values include: 'PreserveHierarchy', 'FlattenHierarchy', 'MergeFiles' @@ -54,7 +56,7 @@ class AzureDataLakeStoreSink(CopySink): 'copy_behavior': {'key': 'copyBehavior', 'type': 'str'}, } - def __init__(self, additional_properties=None, write_batch_size=None, write_batch_timeout=None, sink_retry_count=None, sink_retry_wait=None, copy_behavior=None): - super(AzureDataLakeStoreSink, self).__init__(additional_properties=additional_properties, write_batch_size=write_batch_size, write_batch_timeout=write_batch_timeout, sink_retry_count=sink_retry_count, sink_retry_wait=sink_retry_wait) - self.copy_behavior = copy_behavior + def __init__(self, **kwargs): + super(AzureDataLakeStoreSink, self).__init__(**kwargs) + self.copy_behavior = kwargs.get('copy_behavior', None) self.type = 'AzureDataLakeStoreSink' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_sink_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_sink_py3.py new file mode 100644 index 000000000000..1ea62762385f --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_sink_py3.py @@ -0,0 +1,62 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_sink import CopySink + + +class AzureDataLakeStoreSink(CopySink): + """A copy activity Azure Data Lake Store sink. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param write_batch_size: Write batch size. Type: integer (or Expression + with resultType integer), minimum: 0. + :type write_batch_size: object + :param write_batch_timeout: Write batch timeout. Type: string (or + Expression with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type write_batch_timeout: object + :param sink_retry_count: Sink retry count. Type: integer (or Expression + with resultType integer). + :type sink_retry_count: object + :param sink_retry_wait: Sink retry wait. Type: string (or Expression with + resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type sink_retry_wait: object + :param type: Required. Constant filled by server. + :type type: str + :param copy_behavior: The type of copy behavior for copy sink. Possible + values include: 'PreserveHierarchy', 'FlattenHierarchy', 'MergeFiles' + :type copy_behavior: str or + ~azure.mgmt.datafactory.models.CopyBehaviorType + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, + 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, + 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, + 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'copy_behavior': {'key': 'copyBehavior', 'type': 'str'}, + } + + def __init__(self, *, additional_properties=None, write_batch_size=None, write_batch_timeout=None, sink_retry_count=None, sink_retry_wait=None, copy_behavior=None, **kwargs) -> None: + super(AzureDataLakeStoreSink, self).__init__(additional_properties=additional_properties, write_batch_size=write_batch_size, write_batch_timeout=write_batch_timeout, sink_retry_count=sink_retry_count, sink_retry_wait=sink_retry_wait, **kwargs) + self.copy_behavior = copy_behavior + self.type = 'AzureDataLakeStoreSink' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_source.py index ddd9319b43d5..60a6599c8fbb 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_source.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_source.py @@ -15,6 +15,8 @@ class AzureDataLakeStoreSource(CopySource): """A copy activity Azure Data Lake source. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -25,7 +27,7 @@ class AzureDataLakeStoreSource(CopySource): with resultType string), pattern: ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type source_retry_wait: object - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param recursive: If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType @@ -45,7 +47,7 @@ class AzureDataLakeStoreSource(CopySource): 'recursive': {'key': 'recursive', 'type': 'object'}, } - def __init__(self, additional_properties=None, source_retry_count=None, source_retry_wait=None, recursive=None): - super(AzureDataLakeStoreSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait) - self.recursive = recursive + def __init__(self, **kwargs): + super(AzureDataLakeStoreSource, self).__init__(**kwargs) + self.recursive = kwargs.get('recursive', None) self.type = 'AzureDataLakeStoreSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_source_py3.py new file mode 100644 index 000000000000..36896f7f69e3 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_source_py3.py @@ -0,0 +1,53 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class AzureDataLakeStoreSource(CopySource): + """A copy activity Azure Data Lake source. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Required. Constant filled by server. + :type type: str + :param recursive: If true, files under the folder path will be read + recursively. Default is true. Type: boolean (or Expression with resultType + boolean). + :type recursive: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'recursive': {'key': 'recursive', 'type': 'object'}, + } + + def __init__(self, *, additional_properties=None, source_retry_count=None, source_retry_wait=None, recursive=None, **kwargs) -> None: + super(AzureDataLakeStoreSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, **kwargs) + self.recursive = recursive + self.type = 'AzureDataLakeStoreSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_databricks_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_databricks_linked_service.py index f5a4e5850739..6f683d9b6062 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_databricks_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_databricks_linked_service.py @@ -15,6 +15,8 @@ class AzureDatabricksLinkedService(LinkedService): """Azure Databricks linked service. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -29,14 +31,14 @@ class AzureDatabricksLinkedService(LinkedService): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param domain: .azuredatabricks.net, domain name of your + :param domain: Required. .azuredatabricks.net, domain name of your Databricks deployment. Type: string (or Expression with resultType string). :type domain: object - :param access_token: Access token for databricks REST API. Refer to - https://docs.azuredatabricks.net/api/latest/authentication.html. Type: + :param access_token: Required. Access token for databricks REST API. Refer + to https://docs.azuredatabricks.net/api/latest/authentication.html. Type: string (or Expression with resultType string). :type access_token: ~azure.mgmt.datafactory.models.SecretBase :param existing_cluster_id: The id of an existing cluster that will be @@ -86,14 +88,14 @@ class AzureDatabricksLinkedService(LinkedService): 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, domain, access_token, additional_properties=None, connect_via=None, description=None, parameters=None, annotations=None, existing_cluster_id=None, new_cluster_version=None, new_cluster_num_of_worker=None, new_cluster_node_type=None, new_cluster_spark_conf=None, encrypted_credential=None): - super(AzureDatabricksLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations) - self.domain = domain - self.access_token = access_token - self.existing_cluster_id = existing_cluster_id - self.new_cluster_version = new_cluster_version - self.new_cluster_num_of_worker = new_cluster_num_of_worker - self.new_cluster_node_type = new_cluster_node_type - self.new_cluster_spark_conf = new_cluster_spark_conf - self.encrypted_credential = encrypted_credential + def __init__(self, **kwargs): + super(AzureDatabricksLinkedService, self).__init__(**kwargs) + self.domain = kwargs.get('domain', None) + self.access_token = kwargs.get('access_token', None) + self.existing_cluster_id = kwargs.get('existing_cluster_id', None) + self.new_cluster_version = kwargs.get('new_cluster_version', None) + self.new_cluster_num_of_worker = kwargs.get('new_cluster_num_of_worker', None) + self.new_cluster_node_type = kwargs.get('new_cluster_node_type', None) + self.new_cluster_spark_conf = kwargs.get('new_cluster_spark_conf', None) + self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'AzureDatabricks' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_databricks_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_databricks_linked_service_py3.py new file mode 100644 index 000000000000..f41395a54f74 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_databricks_linked_service_py3.py @@ -0,0 +1,101 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class AzureDatabricksLinkedService(LinkedService): + """Azure Databricks linked service. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param domain: Required. .azuredatabricks.net, domain name of your + Databricks deployment. Type: string (or Expression with resultType + string). + :type domain: object + :param access_token: Required. Access token for databricks REST API. Refer + to https://docs.azuredatabricks.net/api/latest/authentication.html. Type: + string (or Expression with resultType string). + :type access_token: ~azure.mgmt.datafactory.models.SecretBase + :param existing_cluster_id: The id of an existing cluster that will be + used for all runs of this job. Type: string (or Expression with resultType + string). + :type existing_cluster_id: object + :param new_cluster_version: The Spark version of new cluster. Type: string + (or Expression with resultType string). + :type new_cluster_version: object + :param new_cluster_num_of_worker: Number of worker nodes that new cluster + should have. A string formatted Int32, like '1' means numOfWorker is 1 or + '1:10' means auto-scale from 1 as min and 10 as max. Type: string (or + Expression with resultType string). + :type new_cluster_num_of_worker: object + :param new_cluster_node_type: The node types of new cluster. Type: string + (or Expression with resultType string). + :type new_cluster_node_type: object + :param new_cluster_spark_conf: a set of optional, user-specified Spark + configuration key-value pairs. + :type new_cluster_spark_conf: dict[str, object] + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'domain': {'required': True}, + 'access_token': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'domain': {'key': 'typeProperties.domain', 'type': 'object'}, + 'access_token': {'key': 'typeProperties.accessToken', 'type': 'SecretBase'}, + 'existing_cluster_id': {'key': 'typeProperties.existingClusterId', 'type': 'object'}, + 'new_cluster_version': {'key': 'typeProperties.newClusterVersion', 'type': 'object'}, + 'new_cluster_num_of_worker': {'key': 'typeProperties.newClusterNumOfWorker', 'type': 'object'}, + 'new_cluster_node_type': {'key': 'typeProperties.newClusterNodeType', 'type': 'object'}, + 'new_cluster_spark_conf': {'key': 'typeProperties.newClusterSparkConf', 'type': '{object}'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, *, domain, access_token, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, existing_cluster_id=None, new_cluster_version=None, new_cluster_num_of_worker=None, new_cluster_node_type=None, new_cluster_spark_conf=None, encrypted_credential=None, **kwargs) -> None: + super(AzureDatabricksLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.domain = domain + self.access_token = access_token + self.existing_cluster_id = existing_cluster_id + self.new_cluster_version = new_cluster_version + self.new_cluster_num_of_worker = new_cluster_num_of_worker + self.new_cluster_node_type = new_cluster_node_type + self.new_cluster_spark_conf = new_cluster_spark_conf + self.encrypted_credential = encrypted_credential + self.type = 'AzureDatabricks' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_key_vault_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_key_vault_linked_service.py index c387356020fe..c7ad622591ee 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_key_vault_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_key_vault_linked_service.py @@ -15,6 +15,8 @@ class AzureKeyVaultLinkedService(LinkedService): """Azure Key Vault linked service. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -29,9 +31,9 @@ class AzureKeyVaultLinkedService(LinkedService): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param base_url: The base URL of the Azure Key Vault. e.g. + :param base_url: Required. The base URL of the Azure Key Vault. e.g. https://myakv.vault.azure.net Type: string (or Expression with resultType string). :type base_url: object @@ -52,7 +54,7 @@ class AzureKeyVaultLinkedService(LinkedService): 'base_url': {'key': 'typeProperties.baseUrl', 'type': 'object'}, } - def __init__(self, base_url, additional_properties=None, connect_via=None, description=None, parameters=None, annotations=None): - super(AzureKeyVaultLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations) - self.base_url = base_url + def __init__(self, **kwargs): + super(AzureKeyVaultLinkedService, self).__init__(**kwargs) + self.base_url = kwargs.get('base_url', None) self.type = 'AzureKeyVault' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_key_vault_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_key_vault_linked_service_py3.py new file mode 100644 index 000000000000..0f66d72b1d75 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_key_vault_linked_service_py3.py @@ -0,0 +1,60 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class AzureKeyVaultLinkedService(LinkedService): + """Azure Key Vault linked service. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param base_url: Required. The base URL of the Azure Key Vault. e.g. + https://myakv.vault.azure.net Type: string (or Expression with resultType + string). + :type base_url: object + """ + + _validation = { + 'type': {'required': True}, + 'base_url': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'base_url': {'key': 'typeProperties.baseUrl', 'type': 'object'}, + } + + def __init__(self, *, base_url, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, **kwargs) -> None: + super(AzureKeyVaultLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.base_url = base_url + self.type = 'AzureKeyVault' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_key_vault_secret_reference.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_key_vault_secret_reference.py index 9e5e976fa083..28d3e7d31cee 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_key_vault_secret_reference.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_key_vault_secret_reference.py @@ -15,12 +15,14 @@ class AzureKeyVaultSecretReference(SecretBase): """Azure Key Vault secret reference. - :param type: Constant filled by server. + All required parameters must be populated in order to send to Azure. + + :param type: Required. Constant filled by server. :type type: str - :param store: The Azure Key Vault linked service reference. + :param store: Required. The Azure Key Vault linked service reference. :type store: ~azure.mgmt.datafactory.models.LinkedServiceReference - :param secret_name: The name of the secret in Azure Key Vault. Type: - string (or Expression with resultType string). + :param secret_name: Required. The name of the secret in Azure Key Vault. + Type: string (or Expression with resultType string). :type secret_name: object :param secret_version: The version of the secret in Azure Key Vault. The default value is the latest version of the secret. Type: string (or @@ -41,9 +43,9 @@ class AzureKeyVaultSecretReference(SecretBase): 'secret_version': {'key': 'secretVersion', 'type': 'object'}, } - def __init__(self, store, secret_name, secret_version=None): - super(AzureKeyVaultSecretReference, self).__init__() - self.store = store - self.secret_name = secret_name - self.secret_version = secret_version + def __init__(self, **kwargs): + super(AzureKeyVaultSecretReference, self).__init__(**kwargs) + self.store = kwargs.get('store', None) + self.secret_name = kwargs.get('secret_name', None) + self.secret_version = kwargs.get('secret_version', None) self.type = 'AzureKeyVaultSecret' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_key_vault_secret_reference_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_key_vault_secret_reference_py3.py new file mode 100644 index 000000000000..4ce7e7c32313 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_key_vault_secret_reference_py3.py @@ -0,0 +1,51 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .secret_base import SecretBase + + +class AzureKeyVaultSecretReference(SecretBase): + """Azure Key Vault secret reference. + + All required parameters must be populated in order to send to Azure. + + :param type: Required. Constant filled by server. + :type type: str + :param store: Required. The Azure Key Vault linked service reference. + :type store: ~azure.mgmt.datafactory.models.LinkedServiceReference + :param secret_name: Required. The name of the secret in Azure Key Vault. + Type: string (or Expression with resultType string). + :type secret_name: object + :param secret_version: The version of the secret in Azure Key Vault. The + default value is the latest version of the secret. Type: string (or + Expression with resultType string). + :type secret_version: object + """ + + _validation = { + 'type': {'required': True}, + 'store': {'required': True}, + 'secret_name': {'required': True}, + } + + _attribute_map = { + 'type': {'key': 'type', 'type': 'str'}, + 'store': {'key': 'store', 'type': 'LinkedServiceReference'}, + 'secret_name': {'key': 'secretName', 'type': 'object'}, + 'secret_version': {'key': 'secretVersion', 'type': 'object'}, + } + + def __init__(self, *, store, secret_name, secret_version=None, **kwargs) -> None: + super(AzureKeyVaultSecretReference, self).__init__(, **kwargs) + self.store = store + self.secret_name = secret_name + self.secret_version = secret_version + self.type = 'AzureKeyVaultSecret' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_batch_execution_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_batch_execution_activity.py index 9a9f4da813ff..826846634438 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_batch_execution_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_batch_execution_activity.py @@ -15,16 +15,18 @@ class AzureMLBatchExecutionActivity(ExecutionActivity): """Azure ML Batch Execution activity. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] - :param name: Activity name. + :param name: Required. Activity name. :type name: str :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. :type linked_service_name: @@ -69,9 +71,9 @@ class AzureMLBatchExecutionActivity(ExecutionActivity): 'web_service_inputs': {'key': 'typeProperties.webServiceInputs', 'type': '{AzureMLWebServiceFile}'}, } - def __init__(self, name, additional_properties=None, description=None, depends_on=None, linked_service_name=None, policy=None, global_parameters=None, web_service_outputs=None, web_service_inputs=None): - super(AzureMLBatchExecutionActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, linked_service_name=linked_service_name, policy=policy) - self.global_parameters = global_parameters - self.web_service_outputs = web_service_outputs - self.web_service_inputs = web_service_inputs + def __init__(self, **kwargs): + super(AzureMLBatchExecutionActivity, self).__init__(**kwargs) + self.global_parameters = kwargs.get('global_parameters', None) + self.web_service_outputs = kwargs.get('web_service_outputs', None) + self.web_service_inputs = kwargs.get('web_service_inputs', None) self.type = 'AzureMLBatchExecution' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_batch_execution_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_batch_execution_activity_py3.py new file mode 100644 index 000000000000..7dc7f9d3a154 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_batch_execution_activity_py3.py @@ -0,0 +1,79 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .execution_activity import ExecutionActivity + + +class AzureMLBatchExecutionActivity(ExecutionActivity): + """Azure ML Batch Execution activity. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param name: Required. Activity name. + :type name: str + :param description: Activity description. + :type description: str + :param depends_on: Activity depends on condition. + :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param type: Required. Constant filled by server. + :type type: str + :param linked_service_name: Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param policy: Activity policy. + :type policy: ~azure.mgmt.datafactory.models.ActivityPolicy + :param global_parameters: Key,Value pairs to be passed to the Azure ML + Batch Execution Service endpoint. Keys must match the names of web service + parameters defined in the published Azure ML web service. Values will be + passed in the GlobalParameters property of the Azure ML batch execution + request. + :type global_parameters: dict[str, object] + :param web_service_outputs: Key,Value pairs, mapping the names of Azure ML + endpoint's Web Service Outputs to AzureMLWebServiceFile objects specifying + the output Blob locations. This information will be passed in the + WebServiceOutputs property of the Azure ML batch execution request. + :type web_service_outputs: dict[str, + ~azure.mgmt.datafactory.models.AzureMLWebServiceFile] + :param web_service_inputs: Key,Value pairs, mapping the names of Azure ML + endpoint's Web Service Inputs to AzureMLWebServiceFile objects specifying + the input Blob locations.. This information will be passed in the + WebServiceInputs property of the Azure ML batch execution request. + :type web_service_inputs: dict[str, + ~azure.mgmt.datafactory.models.AzureMLWebServiceFile] + """ + + _validation = { + 'name': {'required': True}, + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'name': {'key': 'name', 'type': 'str'}, + 'description': {'key': 'description', 'type': 'str'}, + 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, + 'global_parameters': {'key': 'typeProperties.globalParameters', 'type': '{object}'}, + 'web_service_outputs': {'key': 'typeProperties.webServiceOutputs', 'type': '{AzureMLWebServiceFile}'}, + 'web_service_inputs': {'key': 'typeProperties.webServiceInputs', 'type': '{AzureMLWebServiceFile}'}, + } + + def __init__(self, *, name: str, additional_properties=None, description: str=None, depends_on=None, linked_service_name=None, policy=None, global_parameters=None, web_service_outputs=None, web_service_inputs=None, **kwargs) -> None: + super(AzureMLBatchExecutionActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, linked_service_name=linked_service_name, policy=policy, **kwargs) + self.global_parameters = global_parameters + self.web_service_outputs = web_service_outputs + self.web_service_inputs = web_service_inputs + self.type = 'AzureMLBatchExecution' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_linked_service.py index ee0fea573933..a6a19be4069b 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_linked_service.py @@ -15,6 +15,8 @@ class AzureMLLinkedService(LinkedService): """Azure ML Web Service linked service. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -29,12 +31,13 @@ class AzureMLLinkedService(LinkedService): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param ml_endpoint: The Batch Execution REST URL for an Azure ML Web - Service endpoint. Type: string (or Expression with resultType string). + :param ml_endpoint: Required. The Batch Execution REST URL for an Azure ML + Web Service endpoint. Type: string (or Expression with resultType string). :type ml_endpoint: object - :param api_key: The API key for accessing the Azure ML model endpoint. + :param api_key: Required. The API key for accessing the Azure ML model + endpoint. :type api_key: ~azure.mgmt.datafactory.models.SecretBase :param update_resource_endpoint: The Update Resource REST URL for an Azure ML Web Service endpoint. Type: string (or Expression with resultType @@ -79,13 +82,13 @@ class AzureMLLinkedService(LinkedService): 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, ml_endpoint, api_key, additional_properties=None, connect_via=None, description=None, parameters=None, annotations=None, update_resource_endpoint=None, service_principal_id=None, service_principal_key=None, tenant=None, encrypted_credential=None): - super(AzureMLLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations) - self.ml_endpoint = ml_endpoint - self.api_key = api_key - self.update_resource_endpoint = update_resource_endpoint - self.service_principal_id = service_principal_id - self.service_principal_key = service_principal_key - self.tenant = tenant - self.encrypted_credential = encrypted_credential + def __init__(self, **kwargs): + super(AzureMLLinkedService, self).__init__(**kwargs) + self.ml_endpoint = kwargs.get('ml_endpoint', None) + self.api_key = kwargs.get('api_key', None) + self.update_resource_endpoint = kwargs.get('update_resource_endpoint', None) + self.service_principal_id = kwargs.get('service_principal_id', None) + self.service_principal_key = kwargs.get('service_principal_key', None) + self.tenant = kwargs.get('tenant', None) + self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'AzureML' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_linked_service_py3.py new file mode 100644 index 000000000000..9bea24015ba7 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_linked_service_py3.py @@ -0,0 +1,94 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class AzureMLLinkedService(LinkedService): + """Azure ML Web Service linked service. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param ml_endpoint: Required. The Batch Execution REST URL for an Azure ML + Web Service endpoint. Type: string (or Expression with resultType string). + :type ml_endpoint: object + :param api_key: Required. The API key for accessing the Azure ML model + endpoint. + :type api_key: ~azure.mgmt.datafactory.models.SecretBase + :param update_resource_endpoint: The Update Resource REST URL for an Azure + ML Web Service endpoint. Type: string (or Expression with resultType + string). + :type update_resource_endpoint: object + :param service_principal_id: The ID of the service principal used to + authenticate against the ARM-based updateResourceEndpoint of an Azure ML + web service. Type: string (or Expression with resultType string). + :type service_principal_id: object + :param service_principal_key: The key of the service principal used to + authenticate against the ARM-based updateResourceEndpoint of an Azure ML + web service. + :type service_principal_key: ~azure.mgmt.datafactory.models.SecretBase + :param tenant: The name or ID of the tenant to which the service principal + belongs. Type: string (or Expression with resultType string). + :type tenant: object + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'ml_endpoint': {'required': True}, + 'api_key': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'ml_endpoint': {'key': 'typeProperties.mlEndpoint', 'type': 'object'}, + 'api_key': {'key': 'typeProperties.apiKey', 'type': 'SecretBase'}, + 'update_resource_endpoint': {'key': 'typeProperties.updateResourceEndpoint', 'type': 'object'}, + 'service_principal_id': {'key': 'typeProperties.servicePrincipalId', 'type': 'object'}, + 'service_principal_key': {'key': 'typeProperties.servicePrincipalKey', 'type': 'SecretBase'}, + 'tenant': {'key': 'typeProperties.tenant', 'type': 'object'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, *, ml_endpoint, api_key, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, update_resource_endpoint=None, service_principal_id=None, service_principal_key=None, tenant=None, encrypted_credential=None, **kwargs) -> None: + super(AzureMLLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.ml_endpoint = ml_endpoint + self.api_key = api_key + self.update_resource_endpoint = update_resource_endpoint + self.service_principal_id = service_principal_id + self.service_principal_key = service_principal_key + self.tenant = tenant + self.encrypted_credential = encrypted_credential + self.type = 'AzureML' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_update_resource_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_update_resource_activity.py index 6800e6cc778c..2a6860f5f41e 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_update_resource_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_update_resource_activity.py @@ -15,32 +15,34 @@ class AzureMLUpdateResourceActivity(ExecutionActivity): """Azure ML Update Resource management activity. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] - :param name: Activity name. + :param name: Required. Activity name. :type name: str :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference :param policy: Activity policy. :type policy: ~azure.mgmt.datafactory.models.ActivityPolicy - :param trained_model_name: Name of the Trained Model module in the Web - Service experiment to be updated. Type: string (or Expression with + :param trained_model_name: Required. Name of the Trained Model module in + the Web Service experiment to be updated. Type: string (or Expression with resultType string). :type trained_model_name: object - :param trained_model_linked_service_name: Name of Azure Storage linked - service holding the .ilearner file that will be uploaded by the update - operation. + :param trained_model_linked_service_name: Required. Name of Azure Storage + linked service holding the .ilearner file that will be uploaded by the + update operation. :type trained_model_linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference - :param trained_model_file_path: The relative file path in + :param trained_model_file_path: Required. The relative file path in trainedModelLinkedService to represent the .ilearner file that will be uploaded by the update operation. Type: string (or Expression with resultType string). @@ -68,9 +70,9 @@ class AzureMLUpdateResourceActivity(ExecutionActivity): 'trained_model_file_path': {'key': 'typeProperties.trainedModelFilePath', 'type': 'object'}, } - def __init__(self, name, trained_model_name, trained_model_linked_service_name, trained_model_file_path, additional_properties=None, description=None, depends_on=None, linked_service_name=None, policy=None): - super(AzureMLUpdateResourceActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, linked_service_name=linked_service_name, policy=policy) - self.trained_model_name = trained_model_name - self.trained_model_linked_service_name = trained_model_linked_service_name - self.trained_model_file_path = trained_model_file_path + def __init__(self, **kwargs): + super(AzureMLUpdateResourceActivity, self).__init__(**kwargs) + self.trained_model_name = kwargs.get('trained_model_name', None) + self.trained_model_linked_service_name = kwargs.get('trained_model_linked_service_name', None) + self.trained_model_file_path = kwargs.get('trained_model_file_path', None) self.type = 'AzureMLUpdateResource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_update_resource_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_update_resource_activity_py3.py new file mode 100644 index 000000000000..d035b2ac4049 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_update_resource_activity_py3.py @@ -0,0 +1,78 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .execution_activity import ExecutionActivity + + +class AzureMLUpdateResourceActivity(ExecutionActivity): + """Azure ML Update Resource management activity. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param name: Required. Activity name. + :type name: str + :param description: Activity description. + :type description: str + :param depends_on: Activity depends on condition. + :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param type: Required. Constant filled by server. + :type type: str + :param linked_service_name: Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param policy: Activity policy. + :type policy: ~azure.mgmt.datafactory.models.ActivityPolicy + :param trained_model_name: Required. Name of the Trained Model module in + the Web Service experiment to be updated. Type: string (or Expression with + resultType string). + :type trained_model_name: object + :param trained_model_linked_service_name: Required. Name of Azure Storage + linked service holding the .ilearner file that will be uploaded by the + update operation. + :type trained_model_linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param trained_model_file_path: Required. The relative file path in + trainedModelLinkedService to represent the .ilearner file that will be + uploaded by the update operation. Type: string (or Expression with + resultType string). + :type trained_model_file_path: object + """ + + _validation = { + 'name': {'required': True}, + 'type': {'required': True}, + 'trained_model_name': {'required': True}, + 'trained_model_linked_service_name': {'required': True}, + 'trained_model_file_path': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'name': {'key': 'name', 'type': 'str'}, + 'description': {'key': 'description', 'type': 'str'}, + 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, + 'trained_model_name': {'key': 'typeProperties.trainedModelName', 'type': 'object'}, + 'trained_model_linked_service_name': {'key': 'typeProperties.trainedModelLinkedServiceName', 'type': 'LinkedServiceReference'}, + 'trained_model_file_path': {'key': 'typeProperties.trainedModelFilePath', 'type': 'object'}, + } + + def __init__(self, *, name: str, trained_model_name, trained_model_linked_service_name, trained_model_file_path, additional_properties=None, description: str=None, depends_on=None, linked_service_name=None, policy=None, **kwargs) -> None: + super(AzureMLUpdateResourceActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, linked_service_name=linked_service_name, policy=policy, **kwargs) + self.trained_model_name = trained_model_name + self.trained_model_linked_service_name = trained_model_linked_service_name + self.trained_model_file_path = trained_model_file_path + self.type = 'AzureMLUpdateResource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_web_service_file.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_web_service_file.py index 381eef2be708..682b24fed830 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_web_service_file.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_web_service_file.py @@ -15,12 +15,14 @@ class AzureMLWebServiceFile(Model): """Azure ML WebService Input/Output file. - :param file_path: The relative file path, including container name, in the - Azure Blob Storage specified by the LinkedService. Type: string (or - Expression with resultType string). + All required parameters must be populated in order to send to Azure. + + :param file_path: Required. The relative file path, including container + name, in the Azure Blob Storage specified by the LinkedService. Type: + string (or Expression with resultType string). :type file_path: object - :param linked_service_name: Reference to an Azure Storage LinkedService, - where Azure ML WebService Input/Output file located. + :param linked_service_name: Required. Reference to an Azure Storage + LinkedService, where Azure ML WebService Input/Output file located. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference """ @@ -35,7 +37,7 @@ class AzureMLWebServiceFile(Model): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, } - def __init__(self, file_path, linked_service_name): - super(AzureMLWebServiceFile, self).__init__() - self.file_path = file_path - self.linked_service_name = linked_service_name + def __init__(self, **kwargs): + super(AzureMLWebServiceFile, self).__init__(**kwargs) + self.file_path = kwargs.get('file_path', None) + self.linked_service_name = kwargs.get('linked_service_name', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_web_service_file_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_web_service_file_py3.py new file mode 100644 index 000000000000..abe75d9d9bf2 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_web_service_file_py3.py @@ -0,0 +1,43 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class AzureMLWebServiceFile(Model): + """Azure ML WebService Input/Output file. + + All required parameters must be populated in order to send to Azure. + + :param file_path: Required. The relative file path, including container + name, in the Azure Blob Storage specified by the LinkedService. Type: + string (or Expression with resultType string). + :type file_path: object + :param linked_service_name: Required. Reference to an Azure Storage + LinkedService, where Azure ML WebService Input/Output file located. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + """ + + _validation = { + 'file_path': {'required': True}, + 'linked_service_name': {'required': True}, + } + + _attribute_map = { + 'file_path': {'key': 'filePath', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + } + + def __init__(self, *, file_path, linked_service_name, **kwargs) -> None: + super(AzureMLWebServiceFile, self).__init__(**kwargs) + self.file_path = file_path + self.linked_service_name = linked_service_name diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_linked_service.py index a74ae09ea0f4..471ecb2e21f0 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_linked_service.py @@ -15,6 +15,8 @@ class AzureMySqlLinkedService(LinkedService): """Azure MySQL database linked service. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -29,9 +31,9 @@ class AzureMySqlLinkedService(LinkedService): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param connection_string: The connection string. + :param connection_string: Required. The connection string. :type connection_string: ~azure.mgmt.datafactory.models.SecretBase :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime @@ -55,8 +57,8 @@ class AzureMySqlLinkedService(LinkedService): 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, connection_string, additional_properties=None, connect_via=None, description=None, parameters=None, annotations=None, encrypted_credential=None): - super(AzureMySqlLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations) - self.connection_string = connection_string - self.encrypted_credential = encrypted_credential + def __init__(self, **kwargs): + super(AzureMySqlLinkedService, self).__init__(**kwargs) + self.connection_string = kwargs.get('connection_string', None) + self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'AzureMySql' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_linked_service_py3.py new file mode 100644 index 000000000000..f6dde454fb27 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_linked_service_py3.py @@ -0,0 +1,64 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class AzureMySqlLinkedService(LinkedService): + """Azure MySQL database linked service. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param connection_string: Required. The connection string. + :type connection_string: ~azure.mgmt.datafactory.models.SecretBase + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'connection_string': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'SecretBase'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, *, connection_string, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, encrypted_credential=None, **kwargs) -> None: + super(AzureMySqlLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.connection_string = connection_string + self.encrypted_credential = encrypted_credential + self.type = 'AzureMySql' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_source.py index fd8490cd0736..7409be73bd09 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_source.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_source.py @@ -15,6 +15,8 @@ class AzureMySqlSource(CopySource): """A copy activity Azure MySQL source. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -25,7 +27,7 @@ class AzureMySqlSource(CopySource): with resultType string), pattern: ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type source_retry_wait: object - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param query: Database query. Type: string (or Expression with resultType string). @@ -44,7 +46,7 @@ class AzureMySqlSource(CopySource): 'query': {'key': 'query', 'type': 'object'}, } - def __init__(self, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None): - super(AzureMySqlSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait) - self.query = query + def __init__(self, **kwargs): + super(AzureMySqlSource, self).__init__(**kwargs) + self.query = kwargs.get('query', None) self.type = 'AzureMySqlSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_source_py3.py new file mode 100644 index 000000000000..aaa6377548ef --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_source_py3.py @@ -0,0 +1,52 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class AzureMySqlSource(CopySource): + """A copy activity Azure MySQL source. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Required. Constant filled by server. + :type type: str + :param query: Database query. Type: string (or Expression with resultType + string). + :type query: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'query': {'key': 'query', 'type': 'object'}, + } + + def __init__(self, *, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None, **kwargs) -> None: + super(AzureMySqlSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, **kwargs) + self.query = query + self.type = 'AzureMySqlSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_table_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_table_dataset.py index 9e5cd9b74643..d342c3528e04 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_table_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_table_dataset.py @@ -15,6 +15,8 @@ class AzureMySqlTableDataset(Dataset): """The Azure MySQL database dataset. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -23,7 +25,7 @@ class AzureMySqlTableDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object - :param linked_service_name: Linked service reference. + :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference :param parameters: Parameters for dataset. @@ -32,7 +34,7 @@ class AzureMySqlTableDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param table_name: The Azure MySQL database table name. Type: string (or Expression with resultType string). @@ -55,7 +57,7 @@ class AzureMySqlTableDataset(Dataset): 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, linked_service_name, additional_properties=None, description=None, structure=None, parameters=None, annotations=None, table_name=None): - super(AzureMySqlTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations) - self.table_name = table_name + def __init__(self, **kwargs): + super(AzureMySqlTableDataset, self).__init__(**kwargs) + self.table_name = kwargs.get('table_name', None) self.type = 'AzureMySqlTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_table_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_table_dataset_py3.py new file mode 100644 index 000000000000..cb8e662e5941 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_table_dataset_py3.py @@ -0,0 +1,63 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class AzureMySqlTableDataset(Dataset): + """The Azure MySQL database dataset. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Required. Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param table_name: The Azure MySQL database table name. Type: string (or + Expression with resultType string). + :type table_name: object + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, + } + + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, table_name=None, **kwargs) -> None: + super(AzureMySqlTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + self.table_name = table_name + self.type = 'AzureMySqlTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_linked_service.py index 971de84ac9c2..62824fc83fcd 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_linked_service.py @@ -15,6 +15,8 @@ class AzurePostgreSqlLinkedService(LinkedService): """Azure PostgreSQL linked service. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -29,7 +31,7 @@ class AzurePostgreSqlLinkedService(LinkedService): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param connection_string: An ODBC connection string. :type connection_string: ~azure.mgmt.datafactory.models.SecretBase @@ -54,8 +56,8 @@ class AzurePostgreSqlLinkedService(LinkedService): 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, additional_properties=None, connect_via=None, description=None, parameters=None, annotations=None, connection_string=None, encrypted_credential=None): - super(AzurePostgreSqlLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations) - self.connection_string = connection_string - self.encrypted_credential = encrypted_credential + def __init__(self, **kwargs): + super(AzurePostgreSqlLinkedService, self).__init__(**kwargs) + self.connection_string = kwargs.get('connection_string', None) + self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'AzurePostgreSql' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_linked_service_py3.py new file mode 100644 index 000000000000..fe496511710a --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_linked_service_py3.py @@ -0,0 +1,63 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class AzurePostgreSqlLinkedService(LinkedService): + """Azure PostgreSQL linked service. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param connection_string: An ODBC connection string. + :type connection_string: ~azure.mgmt.datafactory.models.SecretBase + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'SecretBase'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, *, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, connection_string=None, encrypted_credential=None, **kwargs) -> None: + super(AzurePostgreSqlLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.connection_string = connection_string + self.encrypted_credential = encrypted_credential + self.type = 'AzurePostgreSql' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_source.py index 5183b370bc5d..816e066ecebb 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_source.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_source.py @@ -15,6 +15,8 @@ class AzurePostgreSqlSource(CopySource): """A copy activity Azure PostgreSQL source. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -25,7 +27,7 @@ class AzurePostgreSqlSource(CopySource): with resultType string), pattern: ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type source_retry_wait: object - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). @@ -44,7 +46,7 @@ class AzurePostgreSqlSource(CopySource): 'query': {'key': 'query', 'type': 'object'}, } - def __init__(self, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None): - super(AzurePostgreSqlSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait) - self.query = query + def __init__(self, **kwargs): + super(AzurePostgreSqlSource, self).__init__(**kwargs) + self.query = kwargs.get('query', None) self.type = 'AzurePostgreSqlSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_source_py3.py new file mode 100644 index 000000000000..6dbc32ed38a7 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_source_py3.py @@ -0,0 +1,52 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class AzurePostgreSqlSource(CopySource): + """A copy activity Azure PostgreSQL source. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Required. Constant filled by server. + :type type: str + :param query: A query to retrieve data from source. Type: string (or + Expression with resultType string). + :type query: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'query': {'key': 'query', 'type': 'object'}, + } + + def __init__(self, *, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None, **kwargs) -> None: + super(AzurePostgreSqlSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, **kwargs) + self.query = query + self.type = 'AzurePostgreSqlSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_table_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_table_dataset.py index 17ec62fce24b..0ebd39c8c3ab 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_table_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_table_dataset.py @@ -15,6 +15,8 @@ class AzurePostgreSqlTableDataset(Dataset): """Azure PostgreSQL dataset. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -23,7 +25,7 @@ class AzurePostgreSqlTableDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object - :param linked_service_name: Linked service reference. + :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference :param parameters: Parameters for dataset. @@ -32,7 +34,7 @@ class AzurePostgreSqlTableDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str """ @@ -41,6 +43,16 @@ class AzurePostgreSqlTableDataset(Dataset): 'type': {'required': True}, } - def __init__(self, linked_service_name, additional_properties=None, description=None, structure=None, parameters=None, annotations=None): - super(AzurePostgreSqlTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations) + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, **kwargs): + super(AzurePostgreSqlTableDataset, self).__init__(**kwargs) self.type = 'AzurePostgreSqlTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_table_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_table_dataset_py3.py new file mode 100644 index 000000000000..8aff95ae35f7 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_table_dataset_py3.py @@ -0,0 +1,58 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class AzurePostgreSqlTableDataset(Dataset): + """Azure PostgreSQL dataset. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Required. Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: + super(AzurePostgreSqlTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + self.type = 'AzurePostgreSqlTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_queue_sink.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_queue_sink.py index b7e2f1831161..5ecb911fb94a 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_queue_sink.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_queue_sink.py @@ -15,6 +15,8 @@ class AzureQueueSink(CopySink): """A copy activity Azure Queue sink. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -32,7 +34,7 @@ class AzureQueueSink(CopySink): resultType string), pattern: ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type sink_retry_wait: object - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str """ @@ -40,6 +42,15 @@ class AzureQueueSink(CopySink): 'type': {'required': True}, } - def __init__(self, additional_properties=None, write_batch_size=None, write_batch_timeout=None, sink_retry_count=None, sink_retry_wait=None): - super(AzureQueueSink, self).__init__(additional_properties=additional_properties, write_batch_size=write_batch_size, write_batch_timeout=write_batch_timeout, sink_retry_count=sink_retry_count, sink_retry_wait=sink_retry_wait) + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, + 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, + 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, + 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, **kwargs): + super(AzureQueueSink, self).__init__(**kwargs) self.type = 'AzureQueueSink' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_queue_sink_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_queue_sink_py3.py new file mode 100644 index 000000000000..4357c290185c --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_queue_sink_py3.py @@ -0,0 +1,56 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_sink import CopySink + + +class AzureQueueSink(CopySink): + """A copy activity Azure Queue sink. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param write_batch_size: Write batch size. Type: integer (or Expression + with resultType integer), minimum: 0. + :type write_batch_size: object + :param write_batch_timeout: Write batch timeout. Type: string (or + Expression with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type write_batch_timeout: object + :param sink_retry_count: Sink retry count. Type: integer (or Expression + with resultType integer). + :type sink_retry_count: object + :param sink_retry_wait: Sink retry wait. Type: string (or Expression with + resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type sink_retry_wait: object + :param type: Required. Constant filled by server. + :type type: str + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, + 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, + 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, + 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, *, additional_properties=None, write_batch_size=None, write_batch_timeout=None, sink_retry_count=None, sink_retry_wait=None, **kwargs) -> None: + super(AzureQueueSink, self).__init__(additional_properties=additional_properties, write_batch_size=write_batch_size, write_batch_timeout=write_batch_timeout, sink_retry_count=sink_retry_count, sink_retry_wait=sink_retry_wait, **kwargs) + self.type = 'AzureQueueSink' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_search_index_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_search_index_dataset.py index abc71cc5163d..afe7efe8f66d 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_search_index_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_search_index_dataset.py @@ -15,6 +15,8 @@ class AzureSearchIndexDataset(Dataset): """The Azure Search Index. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -23,7 +25,7 @@ class AzureSearchIndexDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object - :param linked_service_name: Linked service reference. + :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference :param parameters: Parameters for dataset. @@ -32,10 +34,10 @@ class AzureSearchIndexDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param index_name: The name of the Azure Search Index. Type: string (or - Expression with resultType string). + :param index_name: Required. The name of the Azure Search Index. Type: + string (or Expression with resultType string). :type index_name: object """ @@ -56,7 +58,7 @@ class AzureSearchIndexDataset(Dataset): 'index_name': {'key': 'typeProperties.indexName', 'type': 'object'}, } - def __init__(self, linked_service_name, index_name, additional_properties=None, description=None, structure=None, parameters=None, annotations=None): - super(AzureSearchIndexDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations) - self.index_name = index_name + def __init__(self, **kwargs): + super(AzureSearchIndexDataset, self).__init__(**kwargs) + self.index_name = kwargs.get('index_name', None) self.type = 'AzureSearchIndex' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_search_index_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_search_index_dataset_py3.py new file mode 100644 index 000000000000..f97335cba8ee --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_search_index_dataset_py3.py @@ -0,0 +1,64 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class AzureSearchIndexDataset(Dataset): + """The Azure Search Index. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Required. Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param index_name: Required. The name of the Azure Search Index. Type: + string (or Expression with resultType string). + :type index_name: object + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + 'index_name': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'index_name': {'key': 'typeProperties.indexName', 'type': 'object'}, + } + + def __init__(self, *, linked_service_name, index_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: + super(AzureSearchIndexDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + self.index_name = index_name + self.type = 'AzureSearchIndex' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_search_index_sink.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_search_index_sink.py index d98dfaed11b6..c09cd94bfb51 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_search_index_sink.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_search_index_sink.py @@ -15,6 +15,8 @@ class AzureSearchIndexSink(CopySink): """A copy activity Azure Search Index sink. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -32,7 +34,7 @@ class AzureSearchIndexSink(CopySink): resultType string), pattern: ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type sink_retry_wait: object - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param write_behavior: Specify the write behavior when upserting documents into Azure Search Index. Possible values include: 'Merge', 'Upload' @@ -54,7 +56,7 @@ class AzureSearchIndexSink(CopySink): 'write_behavior': {'key': 'writeBehavior', 'type': 'str'}, } - def __init__(self, additional_properties=None, write_batch_size=None, write_batch_timeout=None, sink_retry_count=None, sink_retry_wait=None, write_behavior=None): - super(AzureSearchIndexSink, self).__init__(additional_properties=additional_properties, write_batch_size=write_batch_size, write_batch_timeout=write_batch_timeout, sink_retry_count=sink_retry_count, sink_retry_wait=sink_retry_wait) - self.write_behavior = write_behavior + def __init__(self, **kwargs): + super(AzureSearchIndexSink, self).__init__(**kwargs) + self.write_behavior = kwargs.get('write_behavior', None) self.type = 'AzureSearchIndexSink' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_search_index_sink_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_search_index_sink_py3.py new file mode 100644 index 000000000000..90ed08a6f910 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_search_index_sink_py3.py @@ -0,0 +1,62 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_sink import CopySink + + +class AzureSearchIndexSink(CopySink): + """A copy activity Azure Search Index sink. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param write_batch_size: Write batch size. Type: integer (or Expression + with resultType integer), minimum: 0. + :type write_batch_size: object + :param write_batch_timeout: Write batch timeout. Type: string (or + Expression with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type write_batch_timeout: object + :param sink_retry_count: Sink retry count. Type: integer (or Expression + with resultType integer). + :type sink_retry_count: object + :param sink_retry_wait: Sink retry wait. Type: string (or Expression with + resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type sink_retry_wait: object + :param type: Required. Constant filled by server. + :type type: str + :param write_behavior: Specify the write behavior when upserting documents + into Azure Search Index. Possible values include: 'Merge', 'Upload' + :type write_behavior: str or + ~azure.mgmt.datafactory.models.AzureSearchIndexWriteBehaviorType + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, + 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, + 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, + 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'write_behavior': {'key': 'writeBehavior', 'type': 'str'}, + } + + def __init__(self, *, additional_properties=None, write_batch_size=None, write_batch_timeout=None, sink_retry_count=None, sink_retry_wait=None, write_behavior=None, **kwargs) -> None: + super(AzureSearchIndexSink, self).__init__(additional_properties=additional_properties, write_batch_size=write_batch_size, write_batch_timeout=write_batch_timeout, sink_retry_count=sink_retry_count, sink_retry_wait=sink_retry_wait, **kwargs) + self.write_behavior = write_behavior + self.type = 'AzureSearchIndexSink' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_search_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_search_linked_service.py index 4ed7fc9e1962..18979ed87ca0 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_search_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_search_linked_service.py @@ -15,6 +15,8 @@ class AzureSearchLinkedService(LinkedService): """Linked service for Windows Azure Search Service. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -29,10 +31,10 @@ class AzureSearchLinkedService(LinkedService): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param url: URL for Azure Search service. Type: string (or Expression with - resultType string). + :param url: Required. URL for Azure Search service. Type: string (or + Expression with resultType string). :type url: object :param key: Admin Key for Azure Search service :type key: ~azure.mgmt.datafactory.models.SecretBase @@ -59,9 +61,9 @@ class AzureSearchLinkedService(LinkedService): 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, url, additional_properties=None, connect_via=None, description=None, parameters=None, annotations=None, key=None, encrypted_credential=None): - super(AzureSearchLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations) - self.url = url - self.key = key - self.encrypted_credential = encrypted_credential + def __init__(self, **kwargs): + super(AzureSearchLinkedService, self).__init__(**kwargs) + self.url = kwargs.get('url', None) + self.key = kwargs.get('key', None) + self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'AzureSearch' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_search_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_search_linked_service_py3.py new file mode 100644 index 000000000000..2cd8ef33be9f --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_search_linked_service_py3.py @@ -0,0 +1,69 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class AzureSearchLinkedService(LinkedService): + """Linked service for Windows Azure Search Service. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param url: Required. URL for Azure Search service. Type: string (or + Expression with resultType string). + :type url: object + :param key: Admin Key for Azure Search service + :type key: ~azure.mgmt.datafactory.models.SecretBase + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'url': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'url': {'key': 'typeProperties.url', 'type': 'object'}, + 'key': {'key': 'typeProperties.key', 'type': 'SecretBase'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, *, url, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, key=None, encrypted_credential=None, **kwargs) -> None: + super(AzureSearchLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.url = url + self.key = key + self.encrypted_credential = encrypted_credential + self.type = 'AzureSearch' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_database_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_database_linked_service.py index 6c2d2f45c718..250b004f00be 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_database_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_database_linked_service.py @@ -15,6 +15,8 @@ class AzureSqlDatabaseLinkedService(LinkedService): """Microsoft Azure SQL Database linked service. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -29,9 +31,9 @@ class AzureSqlDatabaseLinkedService(LinkedService): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param connection_string: The connection string. + :param connection_string: Required. The connection string. :type connection_string: ~azure.mgmt.datafactory.models.SecretBase :param service_principal_id: The ID of the service principal used to authenticate against Azure SQL Database. Type: string (or Expression with @@ -68,11 +70,11 @@ class AzureSqlDatabaseLinkedService(LinkedService): 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, connection_string, additional_properties=None, connect_via=None, description=None, parameters=None, annotations=None, service_principal_id=None, service_principal_key=None, tenant=None, encrypted_credential=None): - super(AzureSqlDatabaseLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations) - self.connection_string = connection_string - self.service_principal_id = service_principal_id - self.service_principal_key = service_principal_key - self.tenant = tenant - self.encrypted_credential = encrypted_credential + def __init__(self, **kwargs): + super(AzureSqlDatabaseLinkedService, self).__init__(**kwargs) + self.connection_string = kwargs.get('connection_string', None) + self.service_principal_id = kwargs.get('service_principal_id', None) + self.service_principal_key = kwargs.get('service_principal_key', None) + self.tenant = kwargs.get('tenant', None) + self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'AzureSqlDatabase' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_database_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_database_linked_service_py3.py new file mode 100644 index 000000000000..ffa86ded4b4e --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_database_linked_service_py3.py @@ -0,0 +1,80 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class AzureSqlDatabaseLinkedService(LinkedService): + """Microsoft Azure SQL Database linked service. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param connection_string: Required. The connection string. + :type connection_string: ~azure.mgmt.datafactory.models.SecretBase + :param service_principal_id: The ID of the service principal used to + authenticate against Azure SQL Database. Type: string (or Expression with + resultType string). + :type service_principal_id: object + :param service_principal_key: The key of the service principal used to + authenticate against Azure SQL Database. + :type service_principal_key: ~azure.mgmt.datafactory.models.SecretBase + :param tenant: The name or ID of the tenant to which the service principal + belongs. Type: string (or Expression with resultType string). + :type tenant: object + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'connection_string': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'SecretBase'}, + 'service_principal_id': {'key': 'typeProperties.servicePrincipalId', 'type': 'object'}, + 'service_principal_key': {'key': 'typeProperties.servicePrincipalKey', 'type': 'SecretBase'}, + 'tenant': {'key': 'typeProperties.tenant', 'type': 'object'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, *, connection_string, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, service_principal_id=None, service_principal_key=None, tenant=None, encrypted_credential=None, **kwargs) -> None: + super(AzureSqlDatabaseLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.connection_string = connection_string + self.service_principal_id = service_principal_id + self.service_principal_key = service_principal_key + self.tenant = tenant + self.encrypted_credential = encrypted_credential + self.type = 'AzureSqlDatabase' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_dw_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_dw_linked_service.py index 41273d3090d7..989d8c799955 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_dw_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_dw_linked_service.py @@ -15,6 +15,8 @@ class AzureSqlDWLinkedService(LinkedService): """Azure SQL Data Warehouse linked service. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -29,9 +31,9 @@ class AzureSqlDWLinkedService(LinkedService): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param connection_string: The connection string. + :param connection_string: Required. The connection string. :type connection_string: ~azure.mgmt.datafactory.models.SecretBase :param service_principal_id: The ID of the service principal used to authenticate against Azure SQL Data Warehouse. Type: string (or Expression @@ -68,11 +70,11 @@ class AzureSqlDWLinkedService(LinkedService): 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, connection_string, additional_properties=None, connect_via=None, description=None, parameters=None, annotations=None, service_principal_id=None, service_principal_key=None, tenant=None, encrypted_credential=None): - super(AzureSqlDWLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations) - self.connection_string = connection_string - self.service_principal_id = service_principal_id - self.service_principal_key = service_principal_key - self.tenant = tenant - self.encrypted_credential = encrypted_credential + def __init__(self, **kwargs): + super(AzureSqlDWLinkedService, self).__init__(**kwargs) + self.connection_string = kwargs.get('connection_string', None) + self.service_principal_id = kwargs.get('service_principal_id', None) + self.service_principal_key = kwargs.get('service_principal_key', None) + self.tenant = kwargs.get('tenant', None) + self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'AzureSqlDW' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_dw_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_dw_linked_service_py3.py new file mode 100644 index 000000000000..6d0644a4d37d --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_dw_linked_service_py3.py @@ -0,0 +1,80 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class AzureSqlDWLinkedService(LinkedService): + """Azure SQL Data Warehouse linked service. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param connection_string: Required. The connection string. + :type connection_string: ~azure.mgmt.datafactory.models.SecretBase + :param service_principal_id: The ID of the service principal used to + authenticate against Azure SQL Data Warehouse. Type: string (or Expression + with resultType string). + :type service_principal_id: object + :param service_principal_key: The key of the service principal used to + authenticate against Azure SQL Data Warehouse. + :type service_principal_key: ~azure.mgmt.datafactory.models.SecretBase + :param tenant: The name or ID of the tenant to which the service principal + belongs. Type: string (or Expression with resultType string). + :type tenant: object + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'connection_string': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'SecretBase'}, + 'service_principal_id': {'key': 'typeProperties.servicePrincipalId', 'type': 'object'}, + 'service_principal_key': {'key': 'typeProperties.servicePrincipalKey', 'type': 'SecretBase'}, + 'tenant': {'key': 'typeProperties.tenant', 'type': 'object'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, *, connection_string, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, service_principal_id=None, service_principal_key=None, tenant=None, encrypted_credential=None, **kwargs) -> None: + super(AzureSqlDWLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.connection_string = connection_string + self.service_principal_id = service_principal_id + self.service_principal_key = service_principal_key + self.tenant = tenant + self.encrypted_credential = encrypted_credential + self.type = 'AzureSqlDW' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_dw_table_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_dw_table_dataset.py index bfc814b0fd3f..c66b72f6ef06 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_dw_table_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_dw_table_dataset.py @@ -15,6 +15,8 @@ class AzureSqlDWTableDataset(Dataset): """The Azure SQL Data Warehouse dataset. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -23,7 +25,7 @@ class AzureSqlDWTableDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object - :param linked_service_name: Linked service reference. + :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference :param parameters: Parameters for dataset. @@ -32,10 +34,10 @@ class AzureSqlDWTableDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param table_name: The table name of the Azure SQL Data Warehouse. Type: - string (or Expression with resultType string). + :param table_name: Required. The table name of the Azure SQL Data + Warehouse. Type: string (or Expression with resultType string). :type table_name: object """ @@ -56,7 +58,7 @@ class AzureSqlDWTableDataset(Dataset): 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, linked_service_name, table_name, additional_properties=None, description=None, structure=None, parameters=None, annotations=None): - super(AzureSqlDWTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations) - self.table_name = table_name + def __init__(self, **kwargs): + super(AzureSqlDWTableDataset, self).__init__(**kwargs) + self.table_name = kwargs.get('table_name', None) self.type = 'AzureSqlDWTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_dw_table_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_dw_table_dataset_py3.py new file mode 100644 index 000000000000..2a5040f3cdcc --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_dw_table_dataset_py3.py @@ -0,0 +1,64 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class AzureSqlDWTableDataset(Dataset): + """The Azure SQL Data Warehouse dataset. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Required. Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param table_name: Required. The table name of the Azure SQL Data + Warehouse. Type: string (or Expression with resultType string). + :type table_name: object + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + 'table_name': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, + } + + def __init__(self, *, linked_service_name, table_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: + super(AzureSqlDWTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + self.table_name = table_name + self.type = 'AzureSqlDWTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_table_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_table_dataset.py index c375715078f3..8b6f061514d8 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_table_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_table_dataset.py @@ -15,6 +15,8 @@ class AzureSqlTableDataset(Dataset): """The Azure SQL Server database dataset. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -23,7 +25,7 @@ class AzureSqlTableDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object - :param linked_service_name: Linked service reference. + :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference :param parameters: Parameters for dataset. @@ -32,10 +34,10 @@ class AzureSqlTableDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param table_name: The table name of the Azure SQL database. Type: string - (or Expression with resultType string). + :param table_name: Required. The table name of the Azure SQL database. + Type: string (or Expression with resultType string). :type table_name: object """ @@ -56,7 +58,7 @@ class AzureSqlTableDataset(Dataset): 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, linked_service_name, table_name, additional_properties=None, description=None, structure=None, parameters=None, annotations=None): - super(AzureSqlTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations) - self.table_name = table_name + def __init__(self, **kwargs): + super(AzureSqlTableDataset, self).__init__(**kwargs) + self.table_name = kwargs.get('table_name', None) self.type = 'AzureSqlTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_table_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_table_dataset_py3.py new file mode 100644 index 000000000000..67d236b072fd --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_table_dataset_py3.py @@ -0,0 +1,64 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class AzureSqlTableDataset(Dataset): + """The Azure SQL Server database dataset. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Required. Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param table_name: Required. The table name of the Azure SQL database. + Type: string (or Expression with resultType string). + :type table_name: object + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + 'table_name': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, + } + + def __init__(self, *, linked_service_name, table_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: + super(AzureSqlTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + self.table_name = table_name + self.type = 'AzureSqlTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_storage_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_storage_linked_service.py index f8d4005c9abb..bd78469cfcf5 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_storage_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_storage_linked_service.py @@ -15,6 +15,8 @@ class AzureStorageLinkedService(LinkedService): """The storage account linked service. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -29,7 +31,7 @@ class AzureStorageLinkedService(LinkedService): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param connection_string: The connection string. It is mutually exclusive with sasUri property. @@ -59,9 +61,9 @@ class AzureStorageLinkedService(LinkedService): 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, additional_properties=None, connect_via=None, description=None, parameters=None, annotations=None, connection_string=None, sas_uri=None, encrypted_credential=None): - super(AzureStorageLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations) - self.connection_string = connection_string - self.sas_uri = sas_uri - self.encrypted_credential = encrypted_credential + def __init__(self, **kwargs): + super(AzureStorageLinkedService, self).__init__(**kwargs) + self.connection_string = kwargs.get('connection_string', None) + self.sas_uri = kwargs.get('sas_uri', None) + self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'AzureStorage' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_storage_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_storage_linked_service_py3.py new file mode 100644 index 000000000000..2cbb53b69dcb --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_storage_linked_service_py3.py @@ -0,0 +1,69 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class AzureStorageLinkedService(LinkedService): + """The storage account linked service. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param connection_string: The connection string. It is mutually exclusive + with sasUri property. + :type connection_string: ~azure.mgmt.datafactory.models.SecretBase + :param sas_uri: SAS URI of the Azure Storage resource. It is mutually + exclusive with connectionString property. + :type sas_uri: ~azure.mgmt.datafactory.models.SecretBase + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'SecretBase'}, + 'sas_uri': {'key': 'typeProperties.sasUri', 'type': 'SecretBase'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, *, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, connection_string=None, sas_uri=None, encrypted_credential=None, **kwargs) -> None: + super(AzureStorageLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.connection_string = connection_string + self.sas_uri = sas_uri + self.encrypted_credential = encrypted_credential + self.type = 'AzureStorage' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_table_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_table_dataset.py index 94a4e34d6f52..810d53a4fdca 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_table_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_table_dataset.py @@ -15,6 +15,8 @@ class AzureTableDataset(Dataset): """The Azure Table storage dataset. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -23,7 +25,7 @@ class AzureTableDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object - :param linked_service_name: Linked service reference. + :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference :param parameters: Parameters for dataset. @@ -32,10 +34,10 @@ class AzureTableDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param table_name: The table name of the Azure Table storage. Type: string - (or Expression with resultType string). + :param table_name: Required. The table name of the Azure Table storage. + Type: string (or Expression with resultType string). :type table_name: object """ @@ -56,7 +58,7 @@ class AzureTableDataset(Dataset): 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, linked_service_name, table_name, additional_properties=None, description=None, structure=None, parameters=None, annotations=None): - super(AzureTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations) - self.table_name = table_name + def __init__(self, **kwargs): + super(AzureTableDataset, self).__init__(**kwargs) + self.table_name = kwargs.get('table_name', None) self.type = 'AzureTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_table_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_table_dataset_py3.py new file mode 100644 index 000000000000..98956e68bc72 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_table_dataset_py3.py @@ -0,0 +1,64 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class AzureTableDataset(Dataset): + """The Azure Table storage dataset. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Required. Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param table_name: Required. The table name of the Azure Table storage. + Type: string (or Expression with resultType string). + :type table_name: object + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + 'table_name': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, + } + + def __init__(self, *, linked_service_name, table_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: + super(AzureTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + self.table_name = table_name + self.type = 'AzureTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_table_sink.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_table_sink.py index 9f85f98c8a39..faba497cc734 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_table_sink.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_table_sink.py @@ -15,6 +15,8 @@ class AzureTableSink(CopySink): """A copy activity Azure Table sink. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -32,7 +34,7 @@ class AzureTableSink(CopySink): resultType string), pattern: ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type sink_retry_wait: object - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param azure_table_default_partition_key_value: Azure Table default partition key value. Type: string (or Expression with resultType string). @@ -65,10 +67,10 @@ class AzureTableSink(CopySink): 'azure_table_insert_type': {'key': 'azureTableInsertType', 'type': 'object'}, } - def __init__(self, additional_properties=None, write_batch_size=None, write_batch_timeout=None, sink_retry_count=None, sink_retry_wait=None, azure_table_default_partition_key_value=None, azure_table_partition_key_name=None, azure_table_row_key_name=None, azure_table_insert_type=None): - super(AzureTableSink, self).__init__(additional_properties=additional_properties, write_batch_size=write_batch_size, write_batch_timeout=write_batch_timeout, sink_retry_count=sink_retry_count, sink_retry_wait=sink_retry_wait) - self.azure_table_default_partition_key_value = azure_table_default_partition_key_value - self.azure_table_partition_key_name = azure_table_partition_key_name - self.azure_table_row_key_name = azure_table_row_key_name - self.azure_table_insert_type = azure_table_insert_type + def __init__(self, **kwargs): + super(AzureTableSink, self).__init__(**kwargs) + self.azure_table_default_partition_key_value = kwargs.get('azure_table_default_partition_key_value', None) + self.azure_table_partition_key_name = kwargs.get('azure_table_partition_key_name', None) + self.azure_table_row_key_name = kwargs.get('azure_table_row_key_name', None) + self.azure_table_insert_type = kwargs.get('azure_table_insert_type', None) self.type = 'AzureTableSink' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_table_sink_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_table_sink_py3.py new file mode 100644 index 000000000000..2ba396eead2f --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_table_sink_py3.py @@ -0,0 +1,76 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_sink import CopySink + + +class AzureTableSink(CopySink): + """A copy activity Azure Table sink. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param write_batch_size: Write batch size. Type: integer (or Expression + with resultType integer), minimum: 0. + :type write_batch_size: object + :param write_batch_timeout: Write batch timeout. Type: string (or + Expression with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type write_batch_timeout: object + :param sink_retry_count: Sink retry count. Type: integer (or Expression + with resultType integer). + :type sink_retry_count: object + :param sink_retry_wait: Sink retry wait. Type: string (or Expression with + resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type sink_retry_wait: object + :param type: Required. Constant filled by server. + :type type: str + :param azure_table_default_partition_key_value: Azure Table default + partition key value. Type: string (or Expression with resultType string). + :type azure_table_default_partition_key_value: object + :param azure_table_partition_key_name: Azure Table partition key name. + Type: string (or Expression with resultType string). + :type azure_table_partition_key_name: object + :param azure_table_row_key_name: Azure Table row key name. Type: string + (or Expression with resultType string). + :type azure_table_row_key_name: object + :param azure_table_insert_type: Azure Table insert type. Type: string (or + Expression with resultType string). + :type azure_table_insert_type: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, + 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, + 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, + 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'azure_table_default_partition_key_value': {'key': 'azureTableDefaultPartitionKeyValue', 'type': 'object'}, + 'azure_table_partition_key_name': {'key': 'azureTablePartitionKeyName', 'type': 'object'}, + 'azure_table_row_key_name': {'key': 'azureTableRowKeyName', 'type': 'object'}, + 'azure_table_insert_type': {'key': 'azureTableInsertType', 'type': 'object'}, + } + + def __init__(self, *, additional_properties=None, write_batch_size=None, write_batch_timeout=None, sink_retry_count=None, sink_retry_wait=None, azure_table_default_partition_key_value=None, azure_table_partition_key_name=None, azure_table_row_key_name=None, azure_table_insert_type=None, **kwargs) -> None: + super(AzureTableSink, self).__init__(additional_properties=additional_properties, write_batch_size=write_batch_size, write_batch_timeout=write_batch_timeout, sink_retry_count=sink_retry_count, sink_retry_wait=sink_retry_wait, **kwargs) + self.azure_table_default_partition_key_value = azure_table_default_partition_key_value + self.azure_table_partition_key_name = azure_table_partition_key_name + self.azure_table_row_key_name = azure_table_row_key_name + self.azure_table_insert_type = azure_table_insert_type + self.type = 'AzureTableSink' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_table_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_table_source.py index 4cf1f054b42d..f4046c989f4e 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_table_source.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_table_source.py @@ -15,6 +15,8 @@ class AzureTableSource(CopySource): """A copy activity Azure Table source. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -25,7 +27,7 @@ class AzureTableSource(CopySource): with resultType string), pattern: ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type source_retry_wait: object - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param azure_table_source_query: Azure Table source query. Type: string (or Expression with resultType string). @@ -49,8 +51,8 @@ class AzureTableSource(CopySource): 'azure_table_source_ignore_table_not_found': {'key': 'azureTableSourceIgnoreTableNotFound', 'type': 'object'}, } - def __init__(self, additional_properties=None, source_retry_count=None, source_retry_wait=None, azure_table_source_query=None, azure_table_source_ignore_table_not_found=None): - super(AzureTableSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait) - self.azure_table_source_query = azure_table_source_query - self.azure_table_source_ignore_table_not_found = azure_table_source_ignore_table_not_found + def __init__(self, **kwargs): + super(AzureTableSource, self).__init__(**kwargs) + self.azure_table_source_query = kwargs.get('azure_table_source_query', None) + self.azure_table_source_ignore_table_not_found = kwargs.get('azure_table_source_ignore_table_not_found', None) self.type = 'AzureTableSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_table_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_table_source_py3.py new file mode 100644 index 000000000000..00c0fb93f3dc --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_table_source_py3.py @@ -0,0 +1,58 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class AzureTableSource(CopySource): + """A copy activity Azure Table source. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Required. Constant filled by server. + :type type: str + :param azure_table_source_query: Azure Table source query. Type: string + (or Expression with resultType string). + :type azure_table_source_query: object + :param azure_table_source_ignore_table_not_found: Azure Table source + ignore table not found. Type: boolean (or Expression with resultType + boolean). + :type azure_table_source_ignore_table_not_found: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'azure_table_source_query': {'key': 'azureTableSourceQuery', 'type': 'object'}, + 'azure_table_source_ignore_table_not_found': {'key': 'azureTableSourceIgnoreTableNotFound', 'type': 'object'}, + } + + def __init__(self, *, additional_properties=None, source_retry_count=None, source_retry_wait=None, azure_table_source_query=None, azure_table_source_ignore_table_not_found=None, **kwargs) -> None: + super(AzureTableSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, **kwargs) + self.azure_table_source_query = azure_table_source_query + self.azure_table_source_ignore_table_not_found = azure_table_source_ignore_table_not_found + self.type = 'AzureTableSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/blob_sink.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/blob_sink.py index 53a53c57d4af..fe90f5836faf 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/blob_sink.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/blob_sink.py @@ -15,6 +15,8 @@ class BlobSink(CopySink): """A copy activity Azure Blob sink. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -32,7 +34,7 @@ class BlobSink(CopySink): resultType string), pattern: ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type sink_retry_wait: object - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param blob_writer_overwrite_files: Blob writer overwrite files. Type: boolean (or Expression with resultType boolean). @@ -66,10 +68,10 @@ class BlobSink(CopySink): 'copy_behavior': {'key': 'copyBehavior', 'type': 'str'}, } - def __init__(self, additional_properties=None, write_batch_size=None, write_batch_timeout=None, sink_retry_count=None, sink_retry_wait=None, blob_writer_overwrite_files=None, blob_writer_date_time_format=None, blob_writer_add_header=None, copy_behavior=None): - super(BlobSink, self).__init__(additional_properties=additional_properties, write_batch_size=write_batch_size, write_batch_timeout=write_batch_timeout, sink_retry_count=sink_retry_count, sink_retry_wait=sink_retry_wait) - self.blob_writer_overwrite_files = blob_writer_overwrite_files - self.blob_writer_date_time_format = blob_writer_date_time_format - self.blob_writer_add_header = blob_writer_add_header - self.copy_behavior = copy_behavior + def __init__(self, **kwargs): + super(BlobSink, self).__init__(**kwargs) + self.blob_writer_overwrite_files = kwargs.get('blob_writer_overwrite_files', None) + self.blob_writer_date_time_format = kwargs.get('blob_writer_date_time_format', None) + self.blob_writer_add_header = kwargs.get('blob_writer_add_header', None) + self.copy_behavior = kwargs.get('copy_behavior', None) self.type = 'BlobSink' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/blob_sink_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/blob_sink_py3.py new file mode 100644 index 000000000000..7d8d3dc642fd --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/blob_sink_py3.py @@ -0,0 +1,77 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_sink import CopySink + + +class BlobSink(CopySink): + """A copy activity Azure Blob sink. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param write_batch_size: Write batch size. Type: integer (or Expression + with resultType integer), minimum: 0. + :type write_batch_size: object + :param write_batch_timeout: Write batch timeout. Type: string (or + Expression with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type write_batch_timeout: object + :param sink_retry_count: Sink retry count. Type: integer (or Expression + with resultType integer). + :type sink_retry_count: object + :param sink_retry_wait: Sink retry wait. Type: string (or Expression with + resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type sink_retry_wait: object + :param type: Required. Constant filled by server. + :type type: str + :param blob_writer_overwrite_files: Blob writer overwrite files. Type: + boolean (or Expression with resultType boolean). + :type blob_writer_overwrite_files: object + :param blob_writer_date_time_format: Blob writer date time format. Type: + string (or Expression with resultType string). + :type blob_writer_date_time_format: object + :param blob_writer_add_header: Blob writer add header. Type: boolean (or + Expression with resultType boolean). + :type blob_writer_add_header: object + :param copy_behavior: The type of copy behavior for copy sink. Possible + values include: 'PreserveHierarchy', 'FlattenHierarchy', 'MergeFiles' + :type copy_behavior: str or + ~azure.mgmt.datafactory.models.CopyBehaviorType + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, + 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, + 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, + 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'blob_writer_overwrite_files': {'key': 'blobWriterOverwriteFiles', 'type': 'object'}, + 'blob_writer_date_time_format': {'key': 'blobWriterDateTimeFormat', 'type': 'object'}, + 'blob_writer_add_header': {'key': 'blobWriterAddHeader', 'type': 'object'}, + 'copy_behavior': {'key': 'copyBehavior', 'type': 'str'}, + } + + def __init__(self, *, additional_properties=None, write_batch_size=None, write_batch_timeout=None, sink_retry_count=None, sink_retry_wait=None, blob_writer_overwrite_files=None, blob_writer_date_time_format=None, blob_writer_add_header=None, copy_behavior=None, **kwargs) -> None: + super(BlobSink, self).__init__(additional_properties=additional_properties, write_batch_size=write_batch_size, write_batch_timeout=write_batch_timeout, sink_retry_count=sink_retry_count, sink_retry_wait=sink_retry_wait, **kwargs) + self.blob_writer_overwrite_files = blob_writer_overwrite_files + self.blob_writer_date_time_format = blob_writer_date_time_format + self.blob_writer_add_header = blob_writer_add_header + self.copy_behavior = copy_behavior + self.type = 'BlobSink' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/blob_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/blob_source.py index e8d0c0bb8148..f563d0af1e2d 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/blob_source.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/blob_source.py @@ -15,6 +15,8 @@ class BlobSource(CopySource): """A copy activity Azure Blob source. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -25,7 +27,7 @@ class BlobSource(CopySource): with resultType string), pattern: ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type source_retry_wait: object - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param treat_empty_as_null: Treat empty as null. Type: boolean (or Expression with resultType boolean). @@ -53,9 +55,9 @@ class BlobSource(CopySource): 'recursive': {'key': 'recursive', 'type': 'object'}, } - def __init__(self, additional_properties=None, source_retry_count=None, source_retry_wait=None, treat_empty_as_null=None, skip_header_line_count=None, recursive=None): - super(BlobSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait) - self.treat_empty_as_null = treat_empty_as_null - self.skip_header_line_count = skip_header_line_count - self.recursive = recursive + def __init__(self, **kwargs): + super(BlobSource, self).__init__(**kwargs) + self.treat_empty_as_null = kwargs.get('treat_empty_as_null', None) + self.skip_header_line_count = kwargs.get('skip_header_line_count', None) + self.recursive = kwargs.get('recursive', None) self.type = 'BlobSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/blob_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/blob_source_py3.py new file mode 100644 index 000000000000..8c442e8f408f --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/blob_source_py3.py @@ -0,0 +1,63 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class BlobSource(CopySource): + """A copy activity Azure Blob source. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Required. Constant filled by server. + :type type: str + :param treat_empty_as_null: Treat empty as null. Type: boolean (or + Expression with resultType boolean). + :type treat_empty_as_null: object + :param skip_header_line_count: Number of header lines to skip from each + blob. Type: integer (or Expression with resultType integer). + :type skip_header_line_count: object + :param recursive: If true, files under the folder path will be read + recursively. Default is true. Type: boolean (or Expression with resultType + boolean). + :type recursive: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'treat_empty_as_null': {'key': 'treatEmptyAsNull', 'type': 'object'}, + 'skip_header_line_count': {'key': 'skipHeaderLineCount', 'type': 'object'}, + 'recursive': {'key': 'recursive', 'type': 'object'}, + } + + def __init__(self, *, additional_properties=None, source_retry_count=None, source_retry_wait=None, treat_empty_as_null=None, skip_header_line_count=None, recursive=None, **kwargs) -> None: + super(BlobSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, **kwargs) + self.treat_empty_as_null = treat_empty_as_null + self.skip_header_line_count = skip_header_line_count + self.recursive = recursive + self.type = 'BlobSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/blob_trigger.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/blob_trigger.py index b277506b5597..07504d83c933 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/blob_trigger.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/blob_trigger.py @@ -18,6 +18,8 @@ class BlobTrigger(MultiplePipelineTrigger): Variables are only populated by the server, and will be ignored when sending a request. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -28,18 +30,19 @@ class BlobTrigger(MultiplePipelineTrigger): 'Started', 'Stopped', 'Disabled' :vartype runtime_state: str or ~azure.mgmt.datafactory.models.TriggerRuntimeState - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param pipelines: Pipelines that need to be started. :type pipelines: list[~azure.mgmt.datafactory.models.TriggerPipelineReference] - :param folder_path: The path of the container/folder that will trigger the - pipeline. + :param folder_path: Required. The path of the container/folder that will + trigger the pipeline. :type folder_path: str - :param max_concurrency: The max number of parallel files to handle when it - is triggered. + :param max_concurrency: Required. The max number of parallel files to + handle when it is triggered. :type max_concurrency: int - :param linked_service: The Azure Storage linked service reference. + :param linked_service: Required. The Azure Storage linked service + reference. :type linked_service: ~azure.mgmt.datafactory.models.LinkedServiceReference """ @@ -63,9 +66,9 @@ class BlobTrigger(MultiplePipelineTrigger): 'linked_service': {'key': 'typeProperties.linkedService', 'type': 'LinkedServiceReference'}, } - def __init__(self, folder_path, max_concurrency, linked_service, additional_properties=None, description=None, pipelines=None): - super(BlobTrigger, self).__init__(additional_properties=additional_properties, description=description, pipelines=pipelines) - self.folder_path = folder_path - self.max_concurrency = max_concurrency - self.linked_service = linked_service + def __init__(self, **kwargs): + super(BlobTrigger, self).__init__(**kwargs) + self.folder_path = kwargs.get('folder_path', None) + self.max_concurrency = kwargs.get('max_concurrency', None) + self.linked_service = kwargs.get('linked_service', None) self.type = 'BlobTrigger' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/blob_trigger_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/blob_trigger_py3.py new file mode 100644 index 000000000000..f7123fbcf9bd --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/blob_trigger_py3.py @@ -0,0 +1,74 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .multiple_pipeline_trigger import MultiplePipelineTrigger + + +class BlobTrigger(MultiplePipelineTrigger): + """Trigger that runs everytime the selected Blob container changes. + + Variables are only populated by the server, and will be ignored when + sending a request. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param description: Trigger description. + :type description: str + :ivar runtime_state: Indicates if trigger is running or not. Updated when + Start/Stop APIs are called on the Trigger. Possible values include: + 'Started', 'Stopped', 'Disabled' + :vartype runtime_state: str or + ~azure.mgmt.datafactory.models.TriggerRuntimeState + :param type: Required. Constant filled by server. + :type type: str + :param pipelines: Pipelines that need to be started. + :type pipelines: + list[~azure.mgmt.datafactory.models.TriggerPipelineReference] + :param folder_path: Required. The path of the container/folder that will + trigger the pipeline. + :type folder_path: str + :param max_concurrency: Required. The max number of parallel files to + handle when it is triggered. + :type max_concurrency: int + :param linked_service: Required. The Azure Storage linked service + reference. + :type linked_service: + ~azure.mgmt.datafactory.models.LinkedServiceReference + """ + + _validation = { + 'runtime_state': {'readonly': True}, + 'type': {'required': True}, + 'folder_path': {'required': True}, + 'max_concurrency': {'required': True}, + 'linked_service': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'runtime_state': {'key': 'runtimeState', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'str'}, + 'pipelines': {'key': 'pipelines', 'type': '[TriggerPipelineReference]'}, + 'folder_path': {'key': 'typeProperties.folderPath', 'type': 'str'}, + 'max_concurrency': {'key': 'typeProperties.maxConcurrency', 'type': 'int'}, + 'linked_service': {'key': 'typeProperties.linkedService', 'type': 'LinkedServiceReference'}, + } + + def __init__(self, *, folder_path: str, max_concurrency: int, linked_service, additional_properties=None, description: str=None, pipelines=None, **kwargs) -> None: + super(BlobTrigger, self).__init__(additional_properties=additional_properties, description=description, pipelines=pipelines, **kwargs) + self.folder_path = folder_path + self.max_concurrency = max_concurrency + self.linked_service = linked_service + self.type = 'BlobTrigger' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cassandra_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cassandra_linked_service.py index 33f8489b168d..974ce49a1c62 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cassandra_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cassandra_linked_service.py @@ -15,6 +15,8 @@ class CassandraLinkedService(LinkedService): """Linked service for Cassandra data source. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -29,10 +31,10 @@ class CassandraLinkedService(LinkedService): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param host: Host name for connection. Type: string (or Expression with - resultType string). + :param host: Required. Host name for connection. Type: string (or + Expression with resultType string). :type host: object :param authentication_type: AuthenticationType to be used for connection. Type: string (or Expression with resultType string). @@ -71,12 +73,12 @@ class CassandraLinkedService(LinkedService): 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, host, additional_properties=None, connect_via=None, description=None, parameters=None, annotations=None, authentication_type=None, port=None, username=None, password=None, encrypted_credential=None): - super(CassandraLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations) - self.host = host - self.authentication_type = authentication_type - self.port = port - self.username = username - self.password = password - self.encrypted_credential = encrypted_credential + def __init__(self, **kwargs): + super(CassandraLinkedService, self).__init__(**kwargs) + self.host = kwargs.get('host', None) + self.authentication_type = kwargs.get('authentication_type', None) + self.port = kwargs.get('port', None) + self.username = kwargs.get('username', None) + self.password = kwargs.get('password', None) + self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'Cassandra' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cassandra_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cassandra_linked_service_py3.py new file mode 100644 index 000000000000..43a1e607b58f --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cassandra_linked_service_py3.py @@ -0,0 +1,84 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class CassandraLinkedService(LinkedService): + """Linked service for Cassandra data source. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param host: Required. Host name for connection. Type: string (or + Expression with resultType string). + :type host: object + :param authentication_type: AuthenticationType to be used for connection. + Type: string (or Expression with resultType string). + :type authentication_type: object + :param port: The port for the connection. Type: integer (or Expression + with resultType integer). + :type port: object + :param username: Username for authentication. Type: string (or Expression + with resultType string). + :type username: object + :param password: Password for authentication. + :type password: ~azure.mgmt.datafactory.models.SecretBase + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'host': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'host': {'key': 'typeProperties.host', 'type': 'object'}, + 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'object'}, + 'port': {'key': 'typeProperties.port', 'type': 'object'}, + 'username': {'key': 'typeProperties.username', 'type': 'object'}, + 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, *, host, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, authentication_type=None, port=None, username=None, password=None, encrypted_credential=None, **kwargs) -> None: + super(CassandraLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.host = host + self.authentication_type = authentication_type + self.port = port + self.username = username + self.password = password + self.encrypted_credential = encrypted_credential + self.type = 'Cassandra' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cassandra_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cassandra_source.py index 559e00c8fb9b..fdd0a228d001 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cassandra_source.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cassandra_source.py @@ -15,6 +15,8 @@ class CassandraSource(CopySource): """A copy activity source for a Cassandra database. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -25,7 +27,7 @@ class CassandraSource(CopySource): with resultType string), pattern: ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type source_retry_wait: object - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param query: Database query. Should be a SQL-92 query expression or Cassandra Query Language (CQL) command. Type: string (or Expression with @@ -56,8 +58,8 @@ class CassandraSource(CopySource): 'consistency_level': {'key': 'consistencyLevel', 'type': 'str'}, } - def __init__(self, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None, consistency_level=None): - super(CassandraSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait) - self.query = query - self.consistency_level = consistency_level + def __init__(self, **kwargs): + super(CassandraSource, self).__init__(**kwargs) + self.query = kwargs.get('query', None) + self.consistency_level = kwargs.get('consistency_level', None) self.type = 'CassandraSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cassandra_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cassandra_source_py3.py new file mode 100644 index 000000000000..6c62419c1e29 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cassandra_source_py3.py @@ -0,0 +1,65 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class CassandraSource(CopySource): + """A copy activity source for a Cassandra database. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Required. Constant filled by server. + :type type: str + :param query: Database query. Should be a SQL-92 query expression or + Cassandra Query Language (CQL) command. Type: string (or Expression with + resultType string). + :type query: object + :param consistency_level: The consistency level specifies how many + Cassandra servers must respond to a read request before returning data to + the client application. Cassandra checks the specified number of Cassandra + servers for data to satisfy the read request. Must be one of + cassandraSourceReadConsistencyLevels. The default value is 'ONE'. It is + case-insensitive. Possible values include: 'ALL', 'EACH_QUORUM', 'QUORUM', + 'LOCAL_QUORUM', 'ONE', 'TWO', 'THREE', 'LOCAL_ONE', 'SERIAL', + 'LOCAL_SERIAL' + :type consistency_level: str or + ~azure.mgmt.datafactory.models.CassandraSourceReadConsistencyLevels + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'query': {'key': 'query', 'type': 'object'}, + 'consistency_level': {'key': 'consistencyLevel', 'type': 'str'}, + } + + def __init__(self, *, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None, consistency_level=None, **kwargs) -> None: + super(CassandraSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, **kwargs) + self.query = query + self.consistency_level = consistency_level + self.type = 'CassandraSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cassandra_table_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cassandra_table_dataset.py index 816ca6112fc9..91b5e11aa8f2 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cassandra_table_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cassandra_table_dataset.py @@ -15,6 +15,8 @@ class CassandraTableDataset(Dataset): """The Cassandra database dataset. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -23,7 +25,7 @@ class CassandraTableDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object - :param linked_service_name: Linked service reference. + :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference :param parameters: Parameters for dataset. @@ -32,7 +34,7 @@ class CassandraTableDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param table_name: The table name of the Cassandra database. Type: string (or Expression with resultType string). @@ -59,8 +61,8 @@ class CassandraTableDataset(Dataset): 'keyspace': {'key': 'typeProperties.keyspace', 'type': 'object'}, } - def __init__(self, linked_service_name, additional_properties=None, description=None, structure=None, parameters=None, annotations=None, table_name=None, keyspace=None): - super(CassandraTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations) - self.table_name = table_name - self.keyspace = keyspace + def __init__(self, **kwargs): + super(CassandraTableDataset, self).__init__(**kwargs) + self.table_name = kwargs.get('table_name', None) + self.keyspace = kwargs.get('keyspace', None) self.type = 'CassandraTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cassandra_table_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cassandra_table_dataset_py3.py new file mode 100644 index 000000000000..e8932fe02571 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cassandra_table_dataset_py3.py @@ -0,0 +1,68 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class CassandraTableDataset(Dataset): + """The Cassandra database dataset. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Required. Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param table_name: The table name of the Cassandra database. Type: string + (or Expression with resultType string). + :type table_name: object + :param keyspace: The keyspace of the Cassandra database. Type: string (or + Expression with resultType string). + :type keyspace: object + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, + 'keyspace': {'key': 'typeProperties.keyspace', 'type': 'object'}, + } + + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, table_name=None, keyspace=None, **kwargs) -> None: + super(CassandraTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + self.table_name = table_name + self.keyspace = keyspace + self.type = 'CassandraTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_linked_service.py index 6e5321bb8ca1..f8bc6443d656 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_linked_service.py @@ -15,6 +15,8 @@ class ConcurLinkedService(LinkedService): """Concur Serivce linked service. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -29,11 +31,13 @@ class ConcurLinkedService(LinkedService): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param client_id: Application client_id supplied by Concur App Management. + :param client_id: Required. Application client_id supplied by Concur App + Management. :type client_id: object - :param username: The user name that you use to access Concur Service. + :param username: Required. The user name that you use to access Concur + Service. :type username: object :param password: The password corresponding to the user name that you provided in the username field. @@ -76,13 +80,13 @@ class ConcurLinkedService(LinkedService): 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, client_id, username, additional_properties=None, connect_via=None, description=None, parameters=None, annotations=None, password=None, use_encrypted_endpoints=None, use_host_verification=None, use_peer_verification=None, encrypted_credential=None): - super(ConcurLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations) - self.client_id = client_id - self.username = username - self.password = password - self.use_encrypted_endpoints = use_encrypted_endpoints - self.use_host_verification = use_host_verification - self.use_peer_verification = use_peer_verification - self.encrypted_credential = encrypted_credential + def __init__(self, **kwargs): + super(ConcurLinkedService, self).__init__(**kwargs) + self.client_id = kwargs.get('client_id', None) + self.username = kwargs.get('username', None) + self.password = kwargs.get('password', None) + self.use_encrypted_endpoints = kwargs.get('use_encrypted_endpoints', None) + self.use_host_verification = kwargs.get('use_host_verification', None) + self.use_peer_verification = kwargs.get('use_peer_verification', None) + self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'Concur' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_linked_service_py3.py new file mode 100644 index 000000000000..ba57ab2874c4 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_linked_service_py3.py @@ -0,0 +1,92 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class ConcurLinkedService(LinkedService): + """Concur Serivce linked service. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param client_id: Required. Application client_id supplied by Concur App + Management. + :type client_id: object + :param username: Required. The user name that you use to access Concur + Service. + :type username: object + :param password: The password corresponding to the user name that you + provided in the username field. + :type password: ~azure.mgmt.datafactory.models.SecretBase + :param use_encrypted_endpoints: Specifies whether the data source + endpoints are encrypted using HTTPS. The default value is true. + :type use_encrypted_endpoints: object + :param use_host_verification: Specifies whether to require the host name + in the server's certificate to match the host name of the server when + connecting over SSL. The default value is true. + :type use_host_verification: object + :param use_peer_verification: Specifies whether to verify the identity of + the server when connecting over SSL. The default value is true. + :type use_peer_verification: object + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'client_id': {'required': True}, + 'username': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'client_id': {'key': 'typeProperties.clientId', 'type': 'object'}, + 'username': {'key': 'typeProperties.username', 'type': 'object'}, + 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, + 'use_encrypted_endpoints': {'key': 'typeProperties.useEncryptedEndpoints', 'type': 'object'}, + 'use_host_verification': {'key': 'typeProperties.useHostVerification', 'type': 'object'}, + 'use_peer_verification': {'key': 'typeProperties.usePeerVerification', 'type': 'object'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, *, client_id, username, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, password=None, use_encrypted_endpoints=None, use_host_verification=None, use_peer_verification=None, encrypted_credential=None, **kwargs) -> None: + super(ConcurLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.client_id = client_id + self.username = username + self.password = password + self.use_encrypted_endpoints = use_encrypted_endpoints + self.use_host_verification = use_host_verification + self.use_peer_verification = use_peer_verification + self.encrypted_credential = encrypted_credential + self.type = 'Concur' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_object_dataset.py index 694b6c2c4c6f..00ee6540daa1 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_object_dataset.py @@ -15,6 +15,8 @@ class ConcurObjectDataset(Dataset): """Concur Serivce dataset. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -23,7 +25,7 @@ class ConcurObjectDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object - :param linked_service_name: Linked service reference. + :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference :param parameters: Parameters for dataset. @@ -32,7 +34,7 @@ class ConcurObjectDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str """ @@ -41,6 +43,16 @@ class ConcurObjectDataset(Dataset): 'type': {'required': True}, } - def __init__(self, linked_service_name, additional_properties=None, description=None, structure=None, parameters=None, annotations=None): - super(ConcurObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations) + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, **kwargs): + super(ConcurObjectDataset, self).__init__(**kwargs) self.type = 'ConcurObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_object_dataset_py3.py new file mode 100644 index 000000000000..50b5325f3656 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_object_dataset_py3.py @@ -0,0 +1,58 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class ConcurObjectDataset(Dataset): + """Concur Serivce dataset. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Required. Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: + super(ConcurObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + self.type = 'ConcurObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_source.py index 95dbc403bbf9..7320de6aa008 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_source.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_source.py @@ -15,6 +15,8 @@ class ConcurSource(CopySource): """A copy activity Concur Serivce source. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -25,7 +27,7 @@ class ConcurSource(CopySource): with resultType string), pattern: ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type source_retry_wait: object - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). @@ -44,7 +46,7 @@ class ConcurSource(CopySource): 'query': {'key': 'query', 'type': 'object'}, } - def __init__(self, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None): - super(ConcurSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait) - self.query = query + def __init__(self, **kwargs): + super(ConcurSource, self).__init__(**kwargs) + self.query = kwargs.get('query', None) self.type = 'ConcurSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_source_py3.py new file mode 100644 index 000000000000..d019deb98e5a --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_source_py3.py @@ -0,0 +1,52 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class ConcurSource(CopySource): + """A copy activity Concur Serivce source. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Required. Constant filled by server. + :type type: str + :param query: A query to retrieve data from source. Type: string (or + Expression with resultType string). + :type query: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'query': {'key': 'query', 'type': 'object'}, + } + + def __init__(self, *, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None, **kwargs) -> None: + super(ConcurSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, **kwargs) + self.query = query + self.type = 'ConcurSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/control_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/control_activity.py index 7938a2cf01b7..200740cbe757 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/control_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/control_activity.py @@ -19,16 +19,18 @@ class ControlActivity(Activity): sub-classes are: FilterActivity, UntilActivity, WaitActivity, ForEachActivity, IfConditionActivity, ExecutePipelineActivity + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] - :param name: Activity name. + :param name: Required. Activity name. :type name: str :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str """ @@ -37,10 +39,18 @@ class ControlActivity(Activity): 'type': {'required': True}, } + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'name': {'key': 'name', 'type': 'str'}, + 'description': {'key': 'description', 'type': 'str'}, + 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'type': {'key': 'type', 'type': 'str'}, + } + _subtype_map = { 'type': {'Filter': 'FilterActivity', 'Until': 'UntilActivity', 'Wait': 'WaitActivity', 'ForEach': 'ForEachActivity', 'IfCondition': 'IfConditionActivity', 'ExecutePipeline': 'ExecutePipelineActivity'} } - def __init__(self, name, additional_properties=None, description=None, depends_on=None): - super(ControlActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on) + def __init__(self, **kwargs): + super(ControlActivity, self).__init__(**kwargs) self.type = 'Container' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/control_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/control_activity_py3.py new file mode 100644 index 000000000000..72eb7caec9fb --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/control_activity_py3.py @@ -0,0 +1,56 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .activity import Activity + + +class ControlActivity(Activity): + """Base class for all control activities like IfCondition, ForEach , Until. + + You probably want to use the sub-classes and not this class directly. Known + sub-classes are: FilterActivity, UntilActivity, WaitActivity, + ForEachActivity, IfConditionActivity, ExecutePipelineActivity + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param name: Required. Activity name. + :type name: str + :param description: Activity description. + :type description: str + :param depends_on: Activity depends on condition. + :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param type: Required. Constant filled by server. + :type type: str + """ + + _validation = { + 'name': {'required': True}, + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'name': {'key': 'name', 'type': 'str'}, + 'description': {'key': 'description', 'type': 'str'}, + 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + _subtype_map = { + 'type': {'Filter': 'FilterActivity', 'Until': 'UntilActivity', 'Wait': 'WaitActivity', 'ForEach': 'ForEachActivity', 'IfCondition': 'IfConditionActivity', 'ExecutePipeline': 'ExecutePipelineActivity'} + } + + def __init__(self, *, name: str, additional_properties=None, description: str=None, depends_on=None, **kwargs) -> None: + super(ControlActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, **kwargs) + self.type = 'Container' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_activity.py index d6a5248e65da..e600efe42d06 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_activity.py @@ -15,25 +15,27 @@ class CopyActivity(ExecutionActivity): """Copy activity. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] - :param name: Activity name. + :param name: Required. Activity name. :type name: str :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference :param policy: Activity policy. :type policy: ~azure.mgmt.datafactory.models.ActivityPolicy - :param source: Copy activity source. + :param source: Required. Copy activity source. :type source: ~azure.mgmt.datafactory.models.CopySource - :param sink: Copy activity sink. + :param sink: Required. Copy activity sink. :type sink: ~azure.mgmt.datafactory.models.CopySink :param translator: Copy activity translator. If not specificed, tabular translator is used. @@ -95,17 +97,17 @@ class CopyActivity(ExecutionActivity): 'outputs': {'key': 'outputs', 'type': '[DatasetReference]'}, } - def __init__(self, name, source, sink, additional_properties=None, description=None, depends_on=None, linked_service_name=None, policy=None, translator=None, enable_staging=None, staging_settings=None, parallel_copies=None, cloud_data_movement_units=None, enable_skip_incompatible_row=None, redirect_incompatible_row_settings=None, inputs=None, outputs=None): - super(CopyActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, linked_service_name=linked_service_name, policy=policy) - self.source = source - self.sink = sink - self.translator = translator - self.enable_staging = enable_staging - self.staging_settings = staging_settings - self.parallel_copies = parallel_copies - self.cloud_data_movement_units = cloud_data_movement_units - self.enable_skip_incompatible_row = enable_skip_incompatible_row - self.redirect_incompatible_row_settings = redirect_incompatible_row_settings - self.inputs = inputs - self.outputs = outputs + def __init__(self, **kwargs): + super(CopyActivity, self).__init__(**kwargs) + self.source = kwargs.get('source', None) + self.sink = kwargs.get('sink', None) + self.translator = kwargs.get('translator', None) + self.enable_staging = kwargs.get('enable_staging', None) + self.staging_settings = kwargs.get('staging_settings', None) + self.parallel_copies = kwargs.get('parallel_copies', None) + self.cloud_data_movement_units = kwargs.get('cloud_data_movement_units', None) + self.enable_skip_incompatible_row = kwargs.get('enable_skip_incompatible_row', None) + self.redirect_incompatible_row_settings = kwargs.get('redirect_incompatible_row_settings', None) + self.inputs = kwargs.get('inputs', None) + self.outputs = kwargs.get('outputs', None) self.type = 'Copy' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_activity_py3.py new file mode 100644 index 000000000000..6da5626be374 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_activity_py3.py @@ -0,0 +1,113 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .execution_activity import ExecutionActivity + + +class CopyActivity(ExecutionActivity): + """Copy activity. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param name: Required. Activity name. + :type name: str + :param description: Activity description. + :type description: str + :param depends_on: Activity depends on condition. + :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param type: Required. Constant filled by server. + :type type: str + :param linked_service_name: Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param policy: Activity policy. + :type policy: ~azure.mgmt.datafactory.models.ActivityPolicy + :param source: Required. Copy activity source. + :type source: ~azure.mgmt.datafactory.models.CopySource + :param sink: Required. Copy activity sink. + :type sink: ~azure.mgmt.datafactory.models.CopySink + :param translator: Copy activity translator. If not specificed, tabular + translator is used. + :type translator: ~azure.mgmt.datafactory.models.CopyTranslator + :param enable_staging: Specifies whether to copy data via an interim + staging. Default value is false. Type: boolean (or Expression with + resultType boolean). + :type enable_staging: object + :param staging_settings: Specifies interim staging settings when + EnableStaging is true. + :type staging_settings: ~azure.mgmt.datafactory.models.StagingSettings + :param parallel_copies: Maximum number of concurrent sessions opened on + the source or sink to avoid overloading the data store. Type: integer (or + Expression with resultType integer), minimum: 0. + :type parallel_copies: object + :param cloud_data_movement_units: Maximum number of cloud data movement + units that can be used to perform this data movement. Type: integer (or + Expression with resultType integer), minimum: 0. + :type cloud_data_movement_units: object + :param enable_skip_incompatible_row: Whether to skip incompatible row. + Default value is false. Type: boolean (or Expression with resultType + boolean). + :type enable_skip_incompatible_row: object + :param redirect_incompatible_row_settings: Redirect incompatible row + settings when EnableSkipIncompatibleRow is true. + :type redirect_incompatible_row_settings: + ~azure.mgmt.datafactory.models.RedirectIncompatibleRowSettings + :param inputs: List of inputs for the activity. + :type inputs: list[~azure.mgmt.datafactory.models.DatasetReference] + :param outputs: List of outputs for the activity. + :type outputs: list[~azure.mgmt.datafactory.models.DatasetReference] + """ + + _validation = { + 'name': {'required': True}, + 'type': {'required': True}, + 'source': {'required': True}, + 'sink': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'name': {'key': 'name', 'type': 'str'}, + 'description': {'key': 'description', 'type': 'str'}, + 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, + 'source': {'key': 'typeProperties.source', 'type': 'CopySource'}, + 'sink': {'key': 'typeProperties.sink', 'type': 'CopySink'}, + 'translator': {'key': 'typeProperties.translator', 'type': 'CopyTranslator'}, + 'enable_staging': {'key': 'typeProperties.enableStaging', 'type': 'object'}, + 'staging_settings': {'key': 'typeProperties.stagingSettings', 'type': 'StagingSettings'}, + 'parallel_copies': {'key': 'typeProperties.parallelCopies', 'type': 'object'}, + 'cloud_data_movement_units': {'key': 'typeProperties.cloudDataMovementUnits', 'type': 'object'}, + 'enable_skip_incompatible_row': {'key': 'typeProperties.enableSkipIncompatibleRow', 'type': 'object'}, + 'redirect_incompatible_row_settings': {'key': 'typeProperties.redirectIncompatibleRowSettings', 'type': 'RedirectIncompatibleRowSettings'}, + 'inputs': {'key': 'inputs', 'type': '[DatasetReference]'}, + 'outputs': {'key': 'outputs', 'type': '[DatasetReference]'}, + } + + def __init__(self, *, name: str, source, sink, additional_properties=None, description: str=None, depends_on=None, linked_service_name=None, policy=None, translator=None, enable_staging=None, staging_settings=None, parallel_copies=None, cloud_data_movement_units=None, enable_skip_incompatible_row=None, redirect_incompatible_row_settings=None, inputs=None, outputs=None, **kwargs) -> None: + super(CopyActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, linked_service_name=linked_service_name, policy=policy, **kwargs) + self.source = source + self.sink = sink + self.translator = translator + self.enable_staging = enable_staging + self.staging_settings = staging_settings + self.parallel_copies = parallel_copies + self.cloud_data_movement_units = cloud_data_movement_units + self.enable_skip_incompatible_row = enable_skip_incompatible_row + self.redirect_incompatible_row_settings = redirect_incompatible_row_settings + self.inputs = inputs + self.outputs = outputs + self.type = 'Copy' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_sink.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_sink.py index 9ef5e01ddcb5..58b55bf39bbc 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_sink.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_sink.py @@ -21,6 +21,8 @@ class CopySink(Model): SqlSink, DocumentDbCollectionSink, FileSystemSink, BlobSink, AzureTableSink, AzureQueueSink, SapCloudForCustomerSink + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -38,7 +40,7 @@ class CopySink(Model): resultType string), pattern: ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type sink_retry_wait: object - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str """ @@ -59,11 +61,11 @@ class CopySink(Model): 'type': {'SalesforceSink': 'SalesforceSink', 'DynamicsSink': 'DynamicsSink', 'OdbcSink': 'OdbcSink', 'AzureSearchIndexSink': 'AzureSearchIndexSink', 'AzureDataLakeStoreSink': 'AzureDataLakeStoreSink', 'OracleSink': 'OracleSink', 'SqlDWSink': 'SqlDWSink', 'SqlSink': 'SqlSink', 'DocumentDbCollectionSink': 'DocumentDbCollectionSink', 'FileSystemSink': 'FileSystemSink', 'BlobSink': 'BlobSink', 'AzureTableSink': 'AzureTableSink', 'AzureQueueSink': 'AzureQueueSink', 'SapCloudForCustomerSink': 'SapCloudForCustomerSink'} } - def __init__(self, additional_properties=None, write_batch_size=None, write_batch_timeout=None, sink_retry_count=None, sink_retry_wait=None): - super(CopySink, self).__init__() - self.additional_properties = additional_properties - self.write_batch_size = write_batch_size - self.write_batch_timeout = write_batch_timeout - self.sink_retry_count = sink_retry_count - self.sink_retry_wait = sink_retry_wait + def __init__(self, **kwargs): + super(CopySink, self).__init__(**kwargs) + self.additional_properties = kwargs.get('additional_properties', None) + self.write_batch_size = kwargs.get('write_batch_size', None) + self.write_batch_timeout = kwargs.get('write_batch_timeout', None) + self.sink_retry_count = kwargs.get('sink_retry_count', None) + self.sink_retry_wait = kwargs.get('sink_retry_wait', None) self.type = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_sink_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_sink_py3.py new file mode 100644 index 000000000000..02dfd30c931e --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_sink_py3.py @@ -0,0 +1,71 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class CopySink(Model): + """A copy activity sink. + + You probably want to use the sub-classes and not this class directly. Known + sub-classes are: SalesforceSink, DynamicsSink, OdbcSink, + AzureSearchIndexSink, AzureDataLakeStoreSink, OracleSink, SqlDWSink, + SqlSink, DocumentDbCollectionSink, FileSystemSink, BlobSink, + AzureTableSink, AzureQueueSink, SapCloudForCustomerSink + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param write_batch_size: Write batch size. Type: integer (or Expression + with resultType integer), minimum: 0. + :type write_batch_size: object + :param write_batch_timeout: Write batch timeout. Type: string (or + Expression with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type write_batch_timeout: object + :param sink_retry_count: Sink retry count. Type: integer (or Expression + with resultType integer). + :type sink_retry_count: object + :param sink_retry_wait: Sink retry wait. Type: string (or Expression with + resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type sink_retry_wait: object + :param type: Required. Constant filled by server. + :type type: str + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, + 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, + 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, + 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + _subtype_map = { + 'type': {'SalesforceSink': 'SalesforceSink', 'DynamicsSink': 'DynamicsSink', 'OdbcSink': 'OdbcSink', 'AzureSearchIndexSink': 'AzureSearchIndexSink', 'AzureDataLakeStoreSink': 'AzureDataLakeStoreSink', 'OracleSink': 'OracleSink', 'SqlDWSink': 'SqlDWSink', 'SqlSink': 'SqlSink', 'DocumentDbCollectionSink': 'DocumentDbCollectionSink', 'FileSystemSink': 'FileSystemSink', 'BlobSink': 'BlobSink', 'AzureTableSink': 'AzureTableSink', 'AzureQueueSink': 'AzureQueueSink', 'SapCloudForCustomerSink': 'SapCloudForCustomerSink'} + } + + def __init__(self, *, additional_properties=None, write_batch_size=None, write_batch_timeout=None, sink_retry_count=None, sink_retry_wait=None, **kwargs) -> None: + super(CopySink, self).__init__(**kwargs) + self.additional_properties = additional_properties + self.write_batch_size = write_batch_size + self.write_batch_timeout = write_batch_timeout + self.sink_retry_count = sink_retry_count + self.sink_retry_wait = sink_retry_wait + self.type = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_source.py index a03ac0dccdca..9a11107fc8e8 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_source.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_source.py @@ -16,18 +16,21 @@ class CopySource(Model): """A copy activity source. You probably want to use the sub-classes and not this class directly. Known - sub-classes are: AmazonRedshiftSource, SalesforceMarketingCloudSource, - VerticaSource, NetezzaSource, ZohoSource, XeroSource, SquareSource, - SparkSource, ShopifySource, ServiceNowSource, QuickBooksSource, - PrestoSource, PhoenixSource, PaypalSource, MarketoSource, MariaDBSource, - MagentoSource, JiraSource, ImpalaSource, HubspotSource, HiveSource, - HBaseSource, GreenplumSource, GoogleBigQuerySource, EloquaSource, - DrillSource, CouchbaseSource, ConcurSource, AzurePostgreSqlSource, - AmazonMWSSource, HttpSource, AzureDataLakeStoreSource, MongoDbSource, - CassandraSource, WebSource, OracleSource, AzureMySqlSource, HdfsSource, - FileSystemSource, SqlDWSource, SqlSource, SapEccSource, - SapCloudForCustomerSource, SalesforceSource, RelationalSource, - DynamicsSource, DocumentDbCollectionSource, BlobSource, AzureTableSource + sub-classes are: AmazonRedshiftSource, ResponsysSource, + SalesforceMarketingCloudSource, VerticaSource, NetezzaSource, ZohoSource, + XeroSource, SquareSource, SparkSource, ShopifySource, ServiceNowSource, + QuickBooksSource, PrestoSource, PhoenixSource, PaypalSource, MarketoSource, + MariaDBSource, MagentoSource, JiraSource, ImpalaSource, HubspotSource, + HiveSource, HBaseSource, GreenplumSource, GoogleBigQuerySource, + EloquaSource, DrillSource, CouchbaseSource, ConcurSource, + AzurePostgreSqlSource, AmazonMWSSource, HttpSource, + AzureDataLakeStoreSource, MongoDbSource, CassandraSource, WebSource, + OracleSource, AzureMySqlSource, HdfsSource, FileSystemSource, SqlDWSource, + SqlSource, SapEccSource, SapCloudForCustomerSource, SalesforceSource, + RelationalSource, DynamicsSource, DocumentDbCollectionSource, BlobSource, + AzureTableSource + + All required parameters must be populated in order to send to Azure. :param additional_properties: Unmatched properties from the message are deserialized this collection @@ -39,7 +42,7 @@ class CopySource(Model): with resultType string), pattern: ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type source_retry_wait: object - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str """ @@ -55,12 +58,12 @@ class CopySource(Model): } _subtype_map = { - 'type': {'AmazonRedshiftSource': 'AmazonRedshiftSource', 'SalesforceMarketingCloudSource': 'SalesforceMarketingCloudSource', 'VerticaSource': 'VerticaSource', 'NetezzaSource': 'NetezzaSource', 'ZohoSource': 'ZohoSource', 'XeroSource': 'XeroSource', 'SquareSource': 'SquareSource', 'SparkSource': 'SparkSource', 'ShopifySource': 'ShopifySource', 'ServiceNowSource': 'ServiceNowSource', 'QuickBooksSource': 'QuickBooksSource', 'PrestoSource': 'PrestoSource', 'PhoenixSource': 'PhoenixSource', 'PaypalSource': 'PaypalSource', 'MarketoSource': 'MarketoSource', 'MariaDBSource': 'MariaDBSource', 'MagentoSource': 'MagentoSource', 'JiraSource': 'JiraSource', 'ImpalaSource': 'ImpalaSource', 'HubspotSource': 'HubspotSource', 'HiveSource': 'HiveSource', 'HBaseSource': 'HBaseSource', 'GreenplumSource': 'GreenplumSource', 'GoogleBigQuerySource': 'GoogleBigQuerySource', 'EloquaSource': 'EloquaSource', 'DrillSource': 'DrillSource', 'CouchbaseSource': 'CouchbaseSource', 'ConcurSource': 'ConcurSource', 'AzurePostgreSqlSource': 'AzurePostgreSqlSource', 'AmazonMWSSource': 'AmazonMWSSource', 'HttpSource': 'HttpSource', 'AzureDataLakeStoreSource': 'AzureDataLakeStoreSource', 'MongoDbSource': 'MongoDbSource', 'CassandraSource': 'CassandraSource', 'WebSource': 'WebSource', 'OracleSource': 'OracleSource', 'AzureMySqlSource': 'AzureMySqlSource', 'HdfsSource': 'HdfsSource', 'FileSystemSource': 'FileSystemSource', 'SqlDWSource': 'SqlDWSource', 'SqlSource': 'SqlSource', 'SapEccSource': 'SapEccSource', 'SapCloudForCustomerSource': 'SapCloudForCustomerSource', 'SalesforceSource': 'SalesforceSource', 'RelationalSource': 'RelationalSource', 'DynamicsSource': 'DynamicsSource', 'DocumentDbCollectionSource': 'DocumentDbCollectionSource', 'BlobSource': 'BlobSource', 'AzureTableSource': 'AzureTableSource'} + 'type': {'AmazonRedshiftSource': 'AmazonRedshiftSource', 'ResponsysSource': 'ResponsysSource', 'SalesforceMarketingCloudSource': 'SalesforceMarketingCloudSource', 'VerticaSource': 'VerticaSource', 'NetezzaSource': 'NetezzaSource', 'ZohoSource': 'ZohoSource', 'XeroSource': 'XeroSource', 'SquareSource': 'SquareSource', 'SparkSource': 'SparkSource', 'ShopifySource': 'ShopifySource', 'ServiceNowSource': 'ServiceNowSource', 'QuickBooksSource': 'QuickBooksSource', 'PrestoSource': 'PrestoSource', 'PhoenixSource': 'PhoenixSource', 'PaypalSource': 'PaypalSource', 'MarketoSource': 'MarketoSource', 'MariaDBSource': 'MariaDBSource', 'MagentoSource': 'MagentoSource', 'JiraSource': 'JiraSource', 'ImpalaSource': 'ImpalaSource', 'HubspotSource': 'HubspotSource', 'HiveSource': 'HiveSource', 'HBaseSource': 'HBaseSource', 'GreenplumSource': 'GreenplumSource', 'GoogleBigQuerySource': 'GoogleBigQuerySource', 'EloquaSource': 'EloquaSource', 'DrillSource': 'DrillSource', 'CouchbaseSource': 'CouchbaseSource', 'ConcurSource': 'ConcurSource', 'AzurePostgreSqlSource': 'AzurePostgreSqlSource', 'AmazonMWSSource': 'AmazonMWSSource', 'HttpSource': 'HttpSource', 'AzureDataLakeStoreSource': 'AzureDataLakeStoreSource', 'MongoDbSource': 'MongoDbSource', 'CassandraSource': 'CassandraSource', 'WebSource': 'WebSource', 'OracleSource': 'OracleSource', 'AzureMySqlSource': 'AzureMySqlSource', 'HdfsSource': 'HdfsSource', 'FileSystemSource': 'FileSystemSource', 'SqlDWSource': 'SqlDWSource', 'SqlSource': 'SqlSource', 'SapEccSource': 'SapEccSource', 'SapCloudForCustomerSource': 'SapCloudForCustomerSource', 'SalesforceSource': 'SalesforceSource', 'RelationalSource': 'RelationalSource', 'DynamicsSource': 'DynamicsSource', 'DocumentDbCollectionSource': 'DocumentDbCollectionSource', 'BlobSource': 'BlobSource', 'AzureTableSource': 'AzureTableSource'} } - def __init__(self, additional_properties=None, source_retry_count=None, source_retry_wait=None): - super(CopySource, self).__init__() - self.additional_properties = additional_properties - self.source_retry_count = source_retry_count - self.source_retry_wait = source_retry_wait + def __init__(self, **kwargs): + super(CopySource, self).__init__(**kwargs) + self.additional_properties = kwargs.get('additional_properties', None) + self.source_retry_count = kwargs.get('source_retry_count', None) + self.source_retry_wait = kwargs.get('source_retry_wait', None) self.type = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_source_py3.py new file mode 100644 index 000000000000..7c1a96b2897a --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_source_py3.py @@ -0,0 +1,69 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class CopySource(Model): + """A copy activity source. + + You probably want to use the sub-classes and not this class directly. Known + sub-classes are: AmazonRedshiftSource, ResponsysSource, + SalesforceMarketingCloudSource, VerticaSource, NetezzaSource, ZohoSource, + XeroSource, SquareSource, SparkSource, ShopifySource, ServiceNowSource, + QuickBooksSource, PrestoSource, PhoenixSource, PaypalSource, MarketoSource, + MariaDBSource, MagentoSource, JiraSource, ImpalaSource, HubspotSource, + HiveSource, HBaseSource, GreenplumSource, GoogleBigQuerySource, + EloquaSource, DrillSource, CouchbaseSource, ConcurSource, + AzurePostgreSqlSource, AmazonMWSSource, HttpSource, + AzureDataLakeStoreSource, MongoDbSource, CassandraSource, WebSource, + OracleSource, AzureMySqlSource, HdfsSource, FileSystemSource, SqlDWSource, + SqlSource, SapEccSource, SapCloudForCustomerSource, SalesforceSource, + RelationalSource, DynamicsSource, DocumentDbCollectionSource, BlobSource, + AzureTableSource + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Required. Constant filled by server. + :type type: str + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + _subtype_map = { + 'type': {'AmazonRedshiftSource': 'AmazonRedshiftSource', 'ResponsysSource': 'ResponsysSource', 'SalesforceMarketingCloudSource': 'SalesforceMarketingCloudSource', 'VerticaSource': 'VerticaSource', 'NetezzaSource': 'NetezzaSource', 'ZohoSource': 'ZohoSource', 'XeroSource': 'XeroSource', 'SquareSource': 'SquareSource', 'SparkSource': 'SparkSource', 'ShopifySource': 'ShopifySource', 'ServiceNowSource': 'ServiceNowSource', 'QuickBooksSource': 'QuickBooksSource', 'PrestoSource': 'PrestoSource', 'PhoenixSource': 'PhoenixSource', 'PaypalSource': 'PaypalSource', 'MarketoSource': 'MarketoSource', 'MariaDBSource': 'MariaDBSource', 'MagentoSource': 'MagentoSource', 'JiraSource': 'JiraSource', 'ImpalaSource': 'ImpalaSource', 'HubspotSource': 'HubspotSource', 'HiveSource': 'HiveSource', 'HBaseSource': 'HBaseSource', 'GreenplumSource': 'GreenplumSource', 'GoogleBigQuerySource': 'GoogleBigQuerySource', 'EloquaSource': 'EloquaSource', 'DrillSource': 'DrillSource', 'CouchbaseSource': 'CouchbaseSource', 'ConcurSource': 'ConcurSource', 'AzurePostgreSqlSource': 'AzurePostgreSqlSource', 'AmazonMWSSource': 'AmazonMWSSource', 'HttpSource': 'HttpSource', 'AzureDataLakeStoreSource': 'AzureDataLakeStoreSource', 'MongoDbSource': 'MongoDbSource', 'CassandraSource': 'CassandraSource', 'WebSource': 'WebSource', 'OracleSource': 'OracleSource', 'AzureMySqlSource': 'AzureMySqlSource', 'HdfsSource': 'HdfsSource', 'FileSystemSource': 'FileSystemSource', 'SqlDWSource': 'SqlDWSource', 'SqlSource': 'SqlSource', 'SapEccSource': 'SapEccSource', 'SapCloudForCustomerSource': 'SapCloudForCustomerSource', 'SalesforceSource': 'SalesforceSource', 'RelationalSource': 'RelationalSource', 'DynamicsSource': 'DynamicsSource', 'DocumentDbCollectionSource': 'DocumentDbCollectionSource', 'BlobSource': 'BlobSource', 'AzureTableSource': 'AzureTableSource'} + } + + def __init__(self, *, additional_properties=None, source_retry_count=None, source_retry_wait=None, **kwargs) -> None: + super(CopySource, self).__init__(**kwargs) + self.additional_properties = additional_properties + self.source_retry_count = source_retry_count + self.source_retry_wait = source_retry_wait + self.type = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_translator.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_translator.py index 9a5452f7c9fa..2b0242ef997c 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_translator.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_translator.py @@ -18,10 +18,12 @@ class CopyTranslator(Model): You probably want to use the sub-classes and not this class directly. Known sub-classes are: TabularTranslator + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str """ @@ -38,7 +40,7 @@ class CopyTranslator(Model): 'type': {'TabularTranslator': 'TabularTranslator'} } - def __init__(self, additional_properties=None): - super(CopyTranslator, self).__init__() - self.additional_properties = additional_properties + def __init__(self, **kwargs): + super(CopyTranslator, self).__init__(**kwargs) + self.additional_properties = kwargs.get('additional_properties', None) self.type = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_translator_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_translator_py3.py new file mode 100644 index 000000000000..3fef58394fd0 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_translator_py3.py @@ -0,0 +1,46 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class CopyTranslator(Model): + """A copy activity translator. + + You probably want to use the sub-classes and not this class directly. Known + sub-classes are: TabularTranslator + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param type: Required. Constant filled by server. + :type type: str + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + _subtype_map = { + 'type': {'TabularTranslator': 'TabularTranslator'} + } + + def __init__(self, *, additional_properties=None, **kwargs) -> None: + super(CopyTranslator, self).__init__(**kwargs) + self.additional_properties = additional_properties + self.type = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cosmos_db_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cosmos_db_linked_service.py index 0461d58e7f0d..1ca42b56e86d 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cosmos_db_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cosmos_db_linked_service.py @@ -15,6 +15,8 @@ class CosmosDbLinkedService(LinkedService): """Microsoft Azure Cosmos Database (CosmosDB) linked service. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -29,9 +31,9 @@ class CosmosDbLinkedService(LinkedService): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param connection_string: The connection string. + :param connection_string: Required. The connection string. :type connection_string: ~azure.mgmt.datafactory.models.SecretBase :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime @@ -55,8 +57,8 @@ class CosmosDbLinkedService(LinkedService): 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, connection_string, additional_properties=None, connect_via=None, description=None, parameters=None, annotations=None, encrypted_credential=None): - super(CosmosDbLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations) - self.connection_string = connection_string - self.encrypted_credential = encrypted_credential + def __init__(self, **kwargs): + super(CosmosDbLinkedService, self).__init__(**kwargs) + self.connection_string = kwargs.get('connection_string', None) + self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'CosmosDb' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cosmos_db_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cosmos_db_linked_service_py3.py new file mode 100644 index 000000000000..8bf8ce4bf74a --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cosmos_db_linked_service_py3.py @@ -0,0 +1,64 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class CosmosDbLinkedService(LinkedService): + """Microsoft Azure Cosmos Database (CosmosDB) linked service. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param connection_string: Required. The connection string. + :type connection_string: ~azure.mgmt.datafactory.models.SecretBase + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'connection_string': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'SecretBase'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, *, connection_string, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, encrypted_credential=None, **kwargs) -> None: + super(CosmosDbLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.connection_string = connection_string + self.encrypted_credential = encrypted_credential + self.type = 'CosmosDb' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_linked_service.py index 562648752f21..63787690a7ab 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_linked_service.py @@ -15,6 +15,8 @@ class CouchbaseLinkedService(LinkedService): """Couchbase server linked service. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -29,7 +31,7 @@ class CouchbaseLinkedService(LinkedService): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param connection_string: An ODBC connection string. :type connection_string: ~azure.mgmt.datafactory.models.SecretBase @@ -54,8 +56,8 @@ class CouchbaseLinkedService(LinkedService): 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, additional_properties=None, connect_via=None, description=None, parameters=None, annotations=None, connection_string=None, encrypted_credential=None): - super(CouchbaseLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations) - self.connection_string = connection_string - self.encrypted_credential = encrypted_credential + def __init__(self, **kwargs): + super(CouchbaseLinkedService, self).__init__(**kwargs) + self.connection_string = kwargs.get('connection_string', None) + self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'Couchbase' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_linked_service_py3.py new file mode 100644 index 000000000000..3c53fb08cd4e --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_linked_service_py3.py @@ -0,0 +1,63 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class CouchbaseLinkedService(LinkedService): + """Couchbase server linked service. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param connection_string: An ODBC connection string. + :type connection_string: ~azure.mgmt.datafactory.models.SecretBase + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'SecretBase'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, *, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, connection_string=None, encrypted_credential=None, **kwargs) -> None: + super(CouchbaseLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.connection_string = connection_string + self.encrypted_credential = encrypted_credential + self.type = 'Couchbase' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_source.py index c570a4582ec6..bfab638594a3 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_source.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_source.py @@ -15,6 +15,8 @@ class CouchbaseSource(CopySource): """A copy activity Couchbase server source. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -25,7 +27,7 @@ class CouchbaseSource(CopySource): with resultType string), pattern: ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type source_retry_wait: object - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). @@ -44,7 +46,7 @@ class CouchbaseSource(CopySource): 'query': {'key': 'query', 'type': 'object'}, } - def __init__(self, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None): - super(CouchbaseSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait) - self.query = query + def __init__(self, **kwargs): + super(CouchbaseSource, self).__init__(**kwargs) + self.query = kwargs.get('query', None) self.type = 'CouchbaseSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_source_py3.py new file mode 100644 index 000000000000..21535b3f1987 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_source_py3.py @@ -0,0 +1,52 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class CouchbaseSource(CopySource): + """A copy activity Couchbase server source. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Required. Constant filled by server. + :type type: str + :param query: A query to retrieve data from source. Type: string (or + Expression with resultType string). + :type query: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'query': {'key': 'query', 'type': 'object'}, + } + + def __init__(self, *, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None, **kwargs) -> None: + super(CouchbaseSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, **kwargs) + self.query = query + self.type = 'CouchbaseSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_table_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_table_dataset.py index 425eefb0594d..21e4a5d39363 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_table_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_table_dataset.py @@ -15,6 +15,8 @@ class CouchbaseTableDataset(Dataset): """Couchbase server dataset. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -23,7 +25,7 @@ class CouchbaseTableDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object - :param linked_service_name: Linked service reference. + :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference :param parameters: Parameters for dataset. @@ -32,7 +34,7 @@ class CouchbaseTableDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str """ @@ -41,6 +43,16 @@ class CouchbaseTableDataset(Dataset): 'type': {'required': True}, } - def __init__(self, linked_service_name, additional_properties=None, description=None, structure=None, parameters=None, annotations=None): - super(CouchbaseTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations) + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, **kwargs): + super(CouchbaseTableDataset, self).__init__(**kwargs) self.type = 'CouchbaseTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_table_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_table_dataset_py3.py new file mode 100644 index 000000000000..99f9d2e8d218 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_table_dataset_py3.py @@ -0,0 +1,58 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class CouchbaseTableDataset(Dataset): + """Couchbase server dataset. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Required. Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: + super(CouchbaseTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + self.type = 'CouchbaseTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/create_run_response.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/create_run_response.py index 696b3fcea4c4..18ec9f963e65 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/create_run_response.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/create_run_response.py @@ -15,7 +15,9 @@ class CreateRunResponse(Model): """Response body with a run identifier. - :param run_id: Identifier of a run. + All required parameters must be populated in order to send to Azure. + + :param run_id: Required. Identifier of a run. :type run_id: str """ @@ -27,6 +29,6 @@ class CreateRunResponse(Model): 'run_id': {'key': 'runId', 'type': 'str'}, } - def __init__(self, run_id): - super(CreateRunResponse, self).__init__() - self.run_id = run_id + def __init__(self, **kwargs): + super(CreateRunResponse, self).__init__(**kwargs) + self.run_id = kwargs.get('run_id', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/create_run_response_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/create_run_response_py3.py new file mode 100644 index 000000000000..bb280441ae90 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/create_run_response_py3.py @@ -0,0 +1,34 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class CreateRunResponse(Model): + """Response body with a run identifier. + + All required parameters must be populated in order to send to Azure. + + :param run_id: Required. Identifier of a run. + :type run_id: str + """ + + _validation = { + 'run_id': {'required': True}, + } + + _attribute_map = { + 'run_id': {'key': 'runId', 'type': 'str'}, + } + + def __init__(self, *, run_id: str, **kwargs) -> None: + super(CreateRunResponse, self).__init__(**kwargs) + self.run_id = run_id diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_activity.py index 83b6f4e28214..2ebe9c2bc401 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_activity.py @@ -15,24 +15,26 @@ class CustomActivity(ExecutionActivity): """Custom activity type. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] - :param name: Activity name. + :param name: Required. Activity name. :type name: str :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference :param policy: Activity policy. :type policy: ~azure.mgmt.datafactory.models.ActivityPolicy - :param command: Command for custom activity Type: string (or Expression - with resultType string). + :param command: Required. Command for custom activity Type: string (or + Expression with resultType string). :type command: object :param resource_linked_service: Resource linked service reference. :type resource_linked_service: @@ -71,11 +73,11 @@ class CustomActivity(ExecutionActivity): 'extended_properties': {'key': 'typeProperties.extendedProperties', 'type': '{object}'}, } - def __init__(self, name, command, additional_properties=None, description=None, depends_on=None, linked_service_name=None, policy=None, resource_linked_service=None, folder_path=None, reference_objects=None, extended_properties=None): - super(CustomActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, linked_service_name=linked_service_name, policy=policy) - self.command = command - self.resource_linked_service = resource_linked_service - self.folder_path = folder_path - self.reference_objects = reference_objects - self.extended_properties = extended_properties + def __init__(self, **kwargs): + super(CustomActivity, self).__init__(**kwargs) + self.command = kwargs.get('command', None) + self.resource_linked_service = kwargs.get('resource_linked_service', None) + self.folder_path = kwargs.get('folder_path', None) + self.reference_objects = kwargs.get('reference_objects', None) + self.extended_properties = kwargs.get('extended_properties', None) self.type = 'Custom' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_activity_py3.py new file mode 100644 index 000000000000..65520468a96c --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_activity_py3.py @@ -0,0 +1,83 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .execution_activity import ExecutionActivity + + +class CustomActivity(ExecutionActivity): + """Custom activity type. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param name: Required. Activity name. + :type name: str + :param description: Activity description. + :type description: str + :param depends_on: Activity depends on condition. + :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param type: Required. Constant filled by server. + :type type: str + :param linked_service_name: Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param policy: Activity policy. + :type policy: ~azure.mgmt.datafactory.models.ActivityPolicy + :param command: Required. Command for custom activity Type: string (or + Expression with resultType string). + :type command: object + :param resource_linked_service: Resource linked service reference. + :type resource_linked_service: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param folder_path: Folder path for resource files Type: string (or + Expression with resultType string). + :type folder_path: object + :param reference_objects: Reference objects + :type reference_objects: + ~azure.mgmt.datafactory.models.CustomActivityReferenceObject + :param extended_properties: User defined property bag. There is no + restriction on the keys or values that can be used. The user specified + custom activity has the full responsibility to consume and interpret the + content defined. + :type extended_properties: dict[str, object] + """ + + _validation = { + 'name': {'required': True}, + 'type': {'required': True}, + 'command': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'name': {'key': 'name', 'type': 'str'}, + 'description': {'key': 'description', 'type': 'str'}, + 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, + 'command': {'key': 'typeProperties.command', 'type': 'object'}, + 'resource_linked_service': {'key': 'typeProperties.resourceLinkedService', 'type': 'LinkedServiceReference'}, + 'folder_path': {'key': 'typeProperties.folderPath', 'type': 'object'}, + 'reference_objects': {'key': 'typeProperties.referenceObjects', 'type': 'CustomActivityReferenceObject'}, + 'extended_properties': {'key': 'typeProperties.extendedProperties', 'type': '{object}'}, + } + + def __init__(self, *, name: str, command, additional_properties=None, description: str=None, depends_on=None, linked_service_name=None, policy=None, resource_linked_service=None, folder_path=None, reference_objects=None, extended_properties=None, **kwargs) -> None: + super(CustomActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, linked_service_name=linked_service_name, policy=policy, **kwargs) + self.command = command + self.resource_linked_service = resource_linked_service + self.folder_path = folder_path + self.reference_objects = reference_objects + self.extended_properties = extended_properties + self.type = 'Custom' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_activity_reference_object.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_activity_reference_object.py index bcf61066590b..5f95a54612dd 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_activity_reference_object.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_activity_reference_object.py @@ -27,7 +27,7 @@ class CustomActivityReferenceObject(Model): 'datasets': {'key': 'datasets', 'type': '[DatasetReference]'}, } - def __init__(self, linked_services=None, datasets=None): - super(CustomActivityReferenceObject, self).__init__() - self.linked_services = linked_services - self.datasets = datasets + def __init__(self, **kwargs): + super(CustomActivityReferenceObject, self).__init__(**kwargs) + self.linked_services = kwargs.get('linked_services', None) + self.datasets = kwargs.get('datasets', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_activity_reference_object_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_activity_reference_object_py3.py new file mode 100644 index 000000000000..f860f0141bd0 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_activity_reference_object_py3.py @@ -0,0 +1,33 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class CustomActivityReferenceObject(Model): + """Reference objects for custom activity. + + :param linked_services: Linked service references. + :type linked_services: + list[~azure.mgmt.datafactory.models.LinkedServiceReference] + :param datasets: Dataset references. + :type datasets: list[~azure.mgmt.datafactory.models.DatasetReference] + """ + + _attribute_map = { + 'linked_services': {'key': 'linkedServices', 'type': '[LinkedServiceReference]'}, + 'datasets': {'key': 'datasets', 'type': '[DatasetReference]'}, + } + + def __init__(self, *, linked_services=None, datasets=None, **kwargs) -> None: + super(CustomActivityReferenceObject, self).__init__(**kwargs) + self.linked_services = linked_services + self.datasets = datasets diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_data_source_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_data_source_linked_service.py index 4a5b4a2ed37a..4bc3a2863fc3 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_data_source_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_data_source_linked_service.py @@ -15,6 +15,8 @@ class CustomDataSourceLinkedService(LinkedService): """Custom linked service. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -29,9 +31,9 @@ class CustomDataSourceLinkedService(LinkedService): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param type_properties: Custom linked service properties. + :param type_properties: Required. Custom linked service properties. :type type_properties: object """ @@ -50,7 +52,7 @@ class CustomDataSourceLinkedService(LinkedService): 'type_properties': {'key': 'typeProperties', 'type': 'object'}, } - def __init__(self, type_properties, additional_properties=None, connect_via=None, description=None, parameters=None, annotations=None): - super(CustomDataSourceLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations) - self.type_properties = type_properties + def __init__(self, **kwargs): + super(CustomDataSourceLinkedService, self).__init__(**kwargs) + self.type_properties = kwargs.get('type_properties', None) self.type = 'CustomDataSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_data_source_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_data_source_linked_service_py3.py new file mode 100644 index 000000000000..d11e5f8f8076 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_data_source_linked_service_py3.py @@ -0,0 +1,58 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class CustomDataSourceLinkedService(LinkedService): + """Custom linked service. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param type_properties: Required. Custom linked service properties. + :type type_properties: object + """ + + _validation = { + 'type': {'required': True}, + 'type_properties': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'type_properties': {'key': 'typeProperties', 'type': 'object'}, + } + + def __init__(self, *, type_properties, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, **kwargs) -> None: + super(CustomDataSourceLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.type_properties = type_properties + self.type = 'CustomDataSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_dataset.py index 9447713b24f0..4dc34f7442a9 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_dataset.py @@ -15,6 +15,8 @@ class CustomDataset(Dataset): """The custom dataset. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -23,7 +25,7 @@ class CustomDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object - :param linked_service_name: Linked service reference. + :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference :param parameters: Parameters for dataset. @@ -32,9 +34,9 @@ class CustomDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param type_properties: Custom dataset properties. + :param type_properties: Required. Custom dataset properties. :type type_properties: object """ @@ -55,7 +57,7 @@ class CustomDataset(Dataset): 'type_properties': {'key': 'typeProperties', 'type': 'object'}, } - def __init__(self, linked_service_name, type_properties, additional_properties=None, description=None, structure=None, parameters=None, annotations=None): - super(CustomDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations) - self.type_properties = type_properties + def __init__(self, **kwargs): + super(CustomDataset, self).__init__(**kwargs) + self.type_properties = kwargs.get('type_properties', None) self.type = 'CustomDataset' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_dataset_py3.py new file mode 100644 index 000000000000..e8c97da85a2d --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_dataset_py3.py @@ -0,0 +1,63 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class CustomDataset(Dataset): + """The custom dataset. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Required. Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param type_properties: Required. Custom dataset properties. + :type type_properties: object + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + 'type_properties': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'type_properties': {'key': 'typeProperties', 'type': 'object'}, + } + + def __init__(self, *, linked_service_name, type_properties, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: + super(CustomDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + self.type_properties = type_properties + self.type = 'CustomDataset' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_factory_management_client_enums.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_factory_management_client_enums.py index c51c0f9e368b..cb828e14276b 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_factory_management_client_enums.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_factory_management_client_enums.py @@ -12,7 +12,7 @@ from enum import Enum -class IntegrationRuntimeState(Enum): +class IntegrationRuntimeState(str, Enum): initial = "Initial" stopped = "Stopped" @@ -25,13 +25,13 @@ class IntegrationRuntimeState(Enum): offline = "Offline" -class IntegrationRuntimeAutoUpdate(Enum): +class IntegrationRuntimeAutoUpdate(str, Enum): on = "On" off = "Off" -class ParameterType(Enum): +class ParameterType(str, Enum): object_enum = "Object" string = "String" @@ -42,7 +42,7 @@ class ParameterType(Enum): secure_string = "SecureString" -class DependencyCondition(Enum): +class DependencyCondition(str, Enum): succeeded = "Succeeded" failed = "Failed" @@ -50,14 +50,14 @@ class DependencyCondition(Enum): completed = "Completed" -class TriggerRuntimeState(Enum): +class TriggerRuntimeState(str, Enum): started = "Started" stopped = "Stopped" disabled = "Disabled" -class PipelineRunQueryFilterOperand(Enum): +class PipelineRunQueryFilterOperand(str, Enum): pipeline_name = "PipelineName" status = "Status" @@ -65,7 +65,7 @@ class PipelineRunQueryFilterOperand(Enum): run_end = "RunEnd" -class PipelineRunQueryFilterOperator(Enum): +class PipelineRunQueryFilterOperator(str, Enum): equals = "Equals" not_equals = "NotEquals" @@ -73,40 +73,40 @@ class PipelineRunQueryFilterOperator(Enum): not_in = "NotIn" -class PipelineRunQueryOrderByField(Enum): +class PipelineRunQueryOrderByField(str, Enum): run_start = "RunStart" run_end = "RunEnd" -class PipelineRunQueryOrder(Enum): +class PipelineRunQueryOrder(str, Enum): asc = "ASC" desc = "DESC" -class TriggerRunStatus(Enum): +class TriggerRunStatus(str, Enum): succeeded = "Succeeded" failed = "Failed" inprogress = "Inprogress" -class SparkServerType(Enum): +class SparkServerType(str, Enum): shark_server = "SharkServer" shark_server2 = "SharkServer2" spark_thrift_server = "SparkThriftServer" -class SparkThriftTransportProtocol(Enum): +class SparkThriftTransportProtocol(str, Enum): binary = "Binary" sasl = "SASL" http = "HTTP " -class SparkAuthenticationType(Enum): +class SparkAuthenticationType(str, Enum): anonymous = "Anonymous" username = "Username" @@ -114,47 +114,47 @@ class SparkAuthenticationType(Enum): windows_azure_hd_insight_service = "WindowsAzureHDInsightService" -class ServiceNowAuthenticationType(Enum): +class ServiceNowAuthenticationType(str, Enum): basic = "Basic" oauth2 = "OAuth2" -class PrestoAuthenticationType(Enum): +class PrestoAuthenticationType(str, Enum): anonymous = "Anonymous" ldap = "LDAP" -class PhoenixAuthenticationType(Enum): +class PhoenixAuthenticationType(str, Enum): anonymous = "Anonymous" username_and_password = "UsernameAndPassword" windows_azure_hd_insight_service = "WindowsAzureHDInsightService" -class ImpalaAuthenticationType(Enum): +class ImpalaAuthenticationType(str, Enum): anonymous = "Anonymous" sasl_username = "SASLUsername" username_and_password = "UsernameAndPassword" -class HiveServerType(Enum): +class HiveServerType(str, Enum): hive_server1 = "HiveServer1" hive_server2 = "HiveServer2" hive_thrift_server = "HiveThriftServer" -class HiveThriftTransportProtocol(Enum): +class HiveThriftTransportProtocol(str, Enum): binary = "Binary" sasl = "SASL" http = "HTTP " -class HiveAuthenticationType(Enum): +class HiveAuthenticationType(str, Enum): anonymous = "Anonymous" username = "Username" @@ -162,37 +162,37 @@ class HiveAuthenticationType(Enum): windows_azure_hd_insight_service = "WindowsAzureHDInsightService" -class HBaseAuthenticationType(Enum): +class HBaseAuthenticationType(str, Enum): anonymous = "Anonymous" basic = "Basic" -class GoogleBigQueryAuthenticationType(Enum): +class GoogleBigQueryAuthenticationType(str, Enum): service_authentication = "ServiceAuthentication" user_authentication = "UserAuthentication" -class SapHanaAuthenticationType(Enum): +class SapHanaAuthenticationType(str, Enum): basic = "Basic" windows = "Windows" -class SftpAuthenticationType(Enum): +class SftpAuthenticationType(str, Enum): basic = "Basic" ssh_public_key = "SshPublicKey" -class FtpAuthenticationType(Enum): +class FtpAuthenticationType(str, Enum): basic = "Basic" anonymous = "Anonymous" -class HttpAuthenticationType(Enum): +class HttpAuthenticationType(str, Enum): basic = "Basic" anonymous = "Anonymous" @@ -201,54 +201,54 @@ class HttpAuthenticationType(Enum): client_certificate = "ClientCertificate" -class MongoDbAuthenticationType(Enum): +class MongoDbAuthenticationType(str, Enum): basic = "Basic" anonymous = "Anonymous" -class ODataAuthenticationType(Enum): +class ODataAuthenticationType(str, Enum): basic = "Basic" anonymous = "Anonymous" -class TeradataAuthenticationType(Enum): +class TeradataAuthenticationType(str, Enum): basic = "Basic" windows = "Windows" -class Db2AuthenticationType(Enum): +class Db2AuthenticationType(str, Enum): basic = "Basic" -class SybaseAuthenticationType(Enum): +class SybaseAuthenticationType(str, Enum): basic = "Basic" windows = "Windows" -class DatasetCompressionLevel(Enum): +class DatasetCompressionLevel(str, Enum): optimal = "Optimal" fastest = "Fastest" -class JsonFormatFilePattern(Enum): +class JsonFormatFilePattern(str, Enum): set_of_objects = "setOfObjects" array_of_objects = "arrayOfObjects" -class TumblingWindowFrequency(Enum): +class TumblingWindowFrequency(str, Enum): minute = "Minute" hour = "Hour" -class DayOfWeek(Enum): +class DayOfWeek(str, Enum): sunday = "Sunday" monday = "Monday" @@ -259,7 +259,7 @@ class DayOfWeek(Enum): saturday = "Saturday" -class DaysOfWeek(Enum): +class DaysOfWeek(str, Enum): sunday = "Sunday" monday = "Monday" @@ -270,7 +270,7 @@ class DaysOfWeek(Enum): saturday = "Saturday" -class RecurrenceFrequency(Enum): +class RecurrenceFrequency(str, Enum): not_specified = "NotSpecified" minute = "Minute" @@ -281,7 +281,7 @@ class RecurrenceFrequency(Enum): year = "Year" -class WebActivityMethod(Enum): +class WebActivityMethod(str, Enum): get = "GET" post = "POST" @@ -289,7 +289,7 @@ class WebActivityMethod(Enum): delete = "DELETE" -class CassandraSourceReadConsistencyLevels(Enum): +class CassandraSourceReadConsistencyLevels(str, Enum): all = "ALL" each_quorum = "EACH_QUORUM" @@ -303,7 +303,7 @@ class CassandraSourceReadConsistencyLevels(Enum): local_serial = "LOCAL_SERIAL" -class StoredProcedureParameterType(Enum): +class StoredProcedureParameterType(str, Enum): string = "String" int_enum = "Int" @@ -313,63 +313,63 @@ class StoredProcedureParameterType(Enum): date_enum = "Date" -class SalesforceSourceReadBehavior(Enum): +class SalesforceSourceReadBehavior(str, Enum): query = "Query" query_all = "QueryAll" -class SSISExecutionRuntime(Enum): +class SSISExecutionRuntime(str, Enum): x64 = "x64" x86 = "x86" -class HDInsightActivityDebugInfoOption(Enum): +class HDInsightActivityDebugInfoOption(str, Enum): none = "None" always = "Always" failure = "Failure" -class SalesforceSinkWriteBehavior(Enum): +class SalesforceSinkWriteBehavior(str, Enum): insert = "Insert" upsert = "Upsert" -class AzureSearchIndexWriteBehaviorType(Enum): +class AzureSearchIndexWriteBehaviorType(str, Enum): merge = "Merge" upload = "Upload" -class CopyBehaviorType(Enum): +class CopyBehaviorType(str, Enum): preserve_hierarchy = "PreserveHierarchy" flatten_hierarchy = "FlattenHierarchy" merge_files = "MergeFiles" -class PolybaseSettingsRejectType(Enum): +class PolybaseSettingsRejectType(str, Enum): value = "value" percentage = "percentage" -class SapCloudForCustomerSinkWriteBehavior(Enum): +class SapCloudForCustomerSinkWriteBehavior(str, Enum): insert = "Insert" update = "Update" -class IntegrationRuntimeType(Enum): +class IntegrationRuntimeType(str, Enum): managed = "Managed" self_hosted = "SelfHosted" -class SelfHostedIntegrationRuntimeNodeStatus(Enum): +class SelfHostedIntegrationRuntimeNodeStatus(str, Enum): need_registration = "NeedRegistration" online = "Online" @@ -380,20 +380,20 @@ class SelfHostedIntegrationRuntimeNodeStatus(Enum): initialize_failed = "InitializeFailed" -class IntegrationRuntimeUpdateResult(Enum): +class IntegrationRuntimeUpdateResult(str, Enum): succeed = "Succeed" fail = "Fail" -class IntegrationRuntimeInternalChannelEncryptionMode(Enum): +class IntegrationRuntimeInternalChannelEncryptionMode(str, Enum): not_set = "NotSet" ssl_encrypted = "SslEncrypted" not_encrypted = "NotEncrypted" -class ManagedIntegrationRuntimeNodeStatus(Enum): +class ManagedIntegrationRuntimeNodeStatus(str, Enum): starting = "Starting" available = "Available" @@ -401,7 +401,7 @@ class ManagedIntegrationRuntimeNodeStatus(Enum): unavailable = "Unavailable" -class IntegrationRuntimeSsisCatalogPricingTier(Enum): +class IntegrationRuntimeSsisCatalogPricingTier(str, Enum): basic = "Basic" standard = "Standard" @@ -409,19 +409,19 @@ class IntegrationRuntimeSsisCatalogPricingTier(Enum): premium_rs = "PremiumRS" -class IntegrationRuntimeLicenseType(Enum): +class IntegrationRuntimeLicenseType(str, Enum): base_price = "BasePrice" license_included = "LicenseIncluded" -class IntegrationRuntimeEdition(Enum): +class IntegrationRuntimeEdition(str, Enum): standard = "Standard" enterprise = "Enterprise" -class IntegrationRuntimeAuthKeyName(Enum): +class IntegrationRuntimeAuthKeyName(str, Enum): auth_key1 = "authKey1" auth_key2 = "authKey2" diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_lake_analytics_usql_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_lake_analytics_usql_activity.py index dca87f252397..a4b8dcdc70ab 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_lake_analytics_usql_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_lake_analytics_usql_activity.py @@ -15,26 +15,28 @@ class DataLakeAnalyticsUSQLActivity(ExecutionActivity): """Data Lake Analytics U-SQL activity. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] - :param name: Activity name. + :param name: Required. Activity name. :type name: str :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference :param policy: Activity policy. :type policy: ~azure.mgmt.datafactory.models.ActivityPolicy - :param script_path: Case-sensitive path to folder that contains the U-SQL - script. Type: string (or Expression with resultType string). + :param script_path: Required. Case-sensitive path to folder that contains + the U-SQL script. Type: string (or Expression with resultType string). :type script_path: object - :param script_linked_service: Script linked service reference. + :param script_linked_service: Required. Script linked service reference. :type script_linked_service: ~azure.mgmt.datafactory.models.LinkedServiceReference :param degree_of_parallelism: The maximum number of nodes simultaneously @@ -81,13 +83,13 @@ class DataLakeAnalyticsUSQLActivity(ExecutionActivity): 'compilation_mode': {'key': 'typeProperties.compilationMode', 'type': 'object'}, } - def __init__(self, name, script_path, script_linked_service, additional_properties=None, description=None, depends_on=None, linked_service_name=None, policy=None, degree_of_parallelism=None, priority=None, parameters=None, runtime_version=None, compilation_mode=None): - super(DataLakeAnalyticsUSQLActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, linked_service_name=linked_service_name, policy=policy) - self.script_path = script_path - self.script_linked_service = script_linked_service - self.degree_of_parallelism = degree_of_parallelism - self.priority = priority - self.parameters = parameters - self.runtime_version = runtime_version - self.compilation_mode = compilation_mode + def __init__(self, **kwargs): + super(DataLakeAnalyticsUSQLActivity, self).__init__(**kwargs) + self.script_path = kwargs.get('script_path', None) + self.script_linked_service = kwargs.get('script_linked_service', None) + self.degree_of_parallelism = kwargs.get('degree_of_parallelism', None) + self.priority = kwargs.get('priority', None) + self.parameters = kwargs.get('parameters', None) + self.runtime_version = kwargs.get('runtime_version', None) + self.compilation_mode = kwargs.get('compilation_mode', None) self.type = 'DataLakeAnalyticsU-SQL' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_lake_analytics_usql_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_lake_analytics_usql_activity_py3.py new file mode 100644 index 000000000000..022ec770081d --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_lake_analytics_usql_activity_py3.py @@ -0,0 +1,95 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .execution_activity import ExecutionActivity + + +class DataLakeAnalyticsUSQLActivity(ExecutionActivity): + """Data Lake Analytics U-SQL activity. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param name: Required. Activity name. + :type name: str + :param description: Activity description. + :type description: str + :param depends_on: Activity depends on condition. + :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param type: Required. Constant filled by server. + :type type: str + :param linked_service_name: Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param policy: Activity policy. + :type policy: ~azure.mgmt.datafactory.models.ActivityPolicy + :param script_path: Required. Case-sensitive path to folder that contains + the U-SQL script. Type: string (or Expression with resultType string). + :type script_path: object + :param script_linked_service: Required. Script linked service reference. + :type script_linked_service: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param degree_of_parallelism: The maximum number of nodes simultaneously + used to run the job. Default value is 1. Type: integer (or Expression with + resultType integer), minimum: 1. + :type degree_of_parallelism: object + :param priority: Determines which jobs out of all that are queued should + be selected to run first. The lower the number, the higher the priority. + Default value is 1000. Type: integer (or Expression with resultType + integer), minimum: 1. + :type priority: object + :param parameters: Parameters for U-SQL job request. + :type parameters: dict[str, object] + :param runtime_version: Runtime version of the U-SQL engine to use. Type: + string (or Expression with resultType string). + :type runtime_version: object + :param compilation_mode: Compilation mode of U-SQL. Must be one of these + values : Semantic, Full and SingleBox. Type: string (or Expression with + resultType string). + :type compilation_mode: object + """ + + _validation = { + 'name': {'required': True}, + 'type': {'required': True}, + 'script_path': {'required': True}, + 'script_linked_service': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'name': {'key': 'name', 'type': 'str'}, + 'description': {'key': 'description', 'type': 'str'}, + 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, + 'script_path': {'key': 'typeProperties.scriptPath', 'type': 'object'}, + 'script_linked_service': {'key': 'typeProperties.scriptLinkedService', 'type': 'LinkedServiceReference'}, + 'degree_of_parallelism': {'key': 'typeProperties.degreeOfParallelism', 'type': 'object'}, + 'priority': {'key': 'typeProperties.priority', 'type': 'object'}, + 'parameters': {'key': 'typeProperties.parameters', 'type': '{object}'}, + 'runtime_version': {'key': 'typeProperties.runtimeVersion', 'type': 'object'}, + 'compilation_mode': {'key': 'typeProperties.compilationMode', 'type': 'object'}, + } + + def __init__(self, *, name: str, script_path, script_linked_service, additional_properties=None, description: str=None, depends_on=None, linked_service_name=None, policy=None, degree_of_parallelism=None, priority=None, parameters=None, runtime_version=None, compilation_mode=None, **kwargs) -> None: + super(DataLakeAnalyticsUSQLActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, linked_service_name=linked_service_name, policy=policy, **kwargs) + self.script_path = script_path + self.script_linked_service = script_linked_service + self.degree_of_parallelism = degree_of_parallelism + self.priority = priority + self.parameters = parameters + self.runtime_version = runtime_version + self.compilation_mode = compilation_mode + self.type = 'DataLakeAnalyticsU-SQL' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_notebook_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_notebook_activity.py index b23a8fc21092..136ee7debaca 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_notebook_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_notebook_activity.py @@ -15,25 +15,27 @@ class DatabricksNotebookActivity(ExecutionActivity): """DatabricksNotebook activity. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] - :param name: Activity name. + :param name: Required. Activity name. :type name: str :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference :param policy: Activity policy. :type policy: ~azure.mgmt.datafactory.models.ActivityPolicy - :param notebook_path: The absolute path of the notebook to be run in the - Databricks Workspace. This path must begin with a slash. Type: string (or - Expression with resultType string). + :param notebook_path: Required. The absolute path of the notebook to be + run in the Databricks Workspace. This path must begin with a slash. Type: + string (or Expression with resultType string). :type notebook_path: object :param base_parameters: Base parameters to be used for each run of this job.If the notebook takes a parameter that is not specified, the default @@ -59,8 +61,8 @@ class DatabricksNotebookActivity(ExecutionActivity): 'base_parameters': {'key': 'typeProperties.baseParameters', 'type': '{object}'}, } - def __init__(self, name, notebook_path, additional_properties=None, description=None, depends_on=None, linked_service_name=None, policy=None, base_parameters=None): - super(DatabricksNotebookActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, linked_service_name=linked_service_name, policy=policy) - self.notebook_path = notebook_path - self.base_parameters = base_parameters + def __init__(self, **kwargs): + super(DatabricksNotebookActivity, self).__init__(**kwargs) + self.notebook_path = kwargs.get('notebook_path', None) + self.base_parameters = kwargs.get('base_parameters', None) self.type = 'DatabricksNotebook' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_notebook_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_notebook_activity_py3.py new file mode 100644 index 000000000000..a2beeb2fe9dd --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_notebook_activity_py3.py @@ -0,0 +1,68 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .execution_activity import ExecutionActivity + + +class DatabricksNotebookActivity(ExecutionActivity): + """DatabricksNotebook activity. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param name: Required. Activity name. + :type name: str + :param description: Activity description. + :type description: str + :param depends_on: Activity depends on condition. + :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param type: Required. Constant filled by server. + :type type: str + :param linked_service_name: Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param policy: Activity policy. + :type policy: ~azure.mgmt.datafactory.models.ActivityPolicy + :param notebook_path: Required. The absolute path of the notebook to be + run in the Databricks Workspace. This path must begin with a slash. Type: + string (or Expression with resultType string). + :type notebook_path: object + :param base_parameters: Base parameters to be used for each run of this + job.If the notebook takes a parameter that is not specified, the default + value from the notebook will be used. + :type base_parameters: dict[str, object] + """ + + _validation = { + 'name': {'required': True}, + 'type': {'required': True}, + 'notebook_path': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'name': {'key': 'name', 'type': 'str'}, + 'description': {'key': 'description', 'type': 'str'}, + 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, + 'notebook_path': {'key': 'typeProperties.notebookPath', 'type': 'object'}, + 'base_parameters': {'key': 'typeProperties.baseParameters', 'type': '{object}'}, + } + + def __init__(self, *, name: str, notebook_path, additional_properties=None, description: str=None, depends_on=None, linked_service_name=None, policy=None, base_parameters=None, **kwargs) -> None: + super(DatabricksNotebookActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, linked_service_name=linked_service_name, policy=policy, **kwargs) + self.notebook_path = notebook_path + self.base_parameters = base_parameters + self.type = 'DatabricksNotebook' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset.py index e3f4fabdd82d..4ebebafdda73 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset.py @@ -17,18 +17,19 @@ class Dataset(Model): data stores, such as tables, files, folders, and documents. You probably want to use the sub-classes and not this class directly. Known - sub-classes are: SalesforceMarketingCloudObjectDataset, - VerticaTableDataset, NetezzaTableDataset, ZohoObjectDataset, - XeroObjectDataset, SquareObjectDataset, SparkObjectDataset, - ShopifyObjectDataset, ServiceNowObjectDataset, QuickBooksObjectDataset, - PrestoObjectDataset, PhoenixObjectDataset, PaypalObjectDataset, - MarketoObjectDataset, MariaDBTableDataset, MagentoObjectDataset, - JiraObjectDataset, ImpalaObjectDataset, HubspotObjectDataset, - HiveObjectDataset, HBaseObjectDataset, GreenplumTableDataset, - GoogleBigQueryObjectDataset, EloquaObjectDataset, DrillTableDataset, - CouchbaseTableDataset, ConcurObjectDataset, AzurePostgreSqlTableDataset, - AmazonMWSObjectDataset, HttpDataset, AzureSearchIndexDataset, - WebTableDataset, SqlServerTableDataset, SapEccResourceDataset, + sub-classes are: ResponsysObjectDataset, + SalesforceMarketingCloudObjectDataset, VerticaTableDataset, + NetezzaTableDataset, ZohoObjectDataset, XeroObjectDataset, + SquareObjectDataset, SparkObjectDataset, ShopifyObjectDataset, + ServiceNowObjectDataset, QuickBooksObjectDataset, PrestoObjectDataset, + PhoenixObjectDataset, PaypalObjectDataset, MarketoObjectDataset, + MariaDBTableDataset, MagentoObjectDataset, JiraObjectDataset, + ImpalaObjectDataset, HubspotObjectDataset, HiveObjectDataset, + HBaseObjectDataset, GreenplumTableDataset, GoogleBigQueryObjectDataset, + EloquaObjectDataset, DrillTableDataset, CouchbaseTableDataset, + ConcurObjectDataset, AzurePostgreSqlTableDataset, AmazonMWSObjectDataset, + HttpDataset, AzureSearchIndexDataset, WebTableDataset, + SqlServerTableDataset, SapEccResourceDataset, SapCloudForCustomerResourceDataset, SalesforceObjectDataset, RelationalTableDataset, AzureMySqlTableDataset, OracleTableDataset, ODataResourceDataset, MongoDbCollectionDataset, FileShareDataset, @@ -37,6 +38,8 @@ class Dataset(Model): AzureSqlDWTableDataset, AzureSqlTableDataset, AzureTableDataset, AzureBlobDataset, AmazonS3Dataset + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -45,7 +48,7 @@ class Dataset(Model): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object - :param linked_service_name: Linked service reference. + :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference :param parameters: Parameters for dataset. @@ -54,7 +57,7 @@ class Dataset(Model): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str """ @@ -74,15 +77,15 @@ class Dataset(Model): } _subtype_map = { - 'type': {'SalesforceMarketingCloudObject': 'SalesforceMarketingCloudObjectDataset', 'VerticaTable': 'VerticaTableDataset', 'NetezzaTable': 'NetezzaTableDataset', 'ZohoObject': 'ZohoObjectDataset', 'XeroObject': 'XeroObjectDataset', 'SquareObject': 'SquareObjectDataset', 'SparkObject': 'SparkObjectDataset', 'ShopifyObject': 'ShopifyObjectDataset', 'ServiceNowObject': 'ServiceNowObjectDataset', 'QuickBooksObject': 'QuickBooksObjectDataset', 'PrestoObject': 'PrestoObjectDataset', 'PhoenixObject': 'PhoenixObjectDataset', 'PaypalObject': 'PaypalObjectDataset', 'MarketoObject': 'MarketoObjectDataset', 'MariaDBTable': 'MariaDBTableDataset', 'MagentoObject': 'MagentoObjectDataset', 'JiraObject': 'JiraObjectDataset', 'ImpalaObject': 'ImpalaObjectDataset', 'HubspotObject': 'HubspotObjectDataset', 'HiveObject': 'HiveObjectDataset', 'HBaseObject': 'HBaseObjectDataset', 'GreenplumTable': 'GreenplumTableDataset', 'GoogleBigQueryObject': 'GoogleBigQueryObjectDataset', 'EloquaObject': 'EloquaObjectDataset', 'DrillTable': 'DrillTableDataset', 'CouchbaseTable': 'CouchbaseTableDataset', 'ConcurObject': 'ConcurObjectDataset', 'AzurePostgreSqlTable': 'AzurePostgreSqlTableDataset', 'AmazonMWSObject': 'AmazonMWSObjectDataset', 'HttpFile': 'HttpDataset', 'AzureSearchIndex': 'AzureSearchIndexDataset', 'WebTable': 'WebTableDataset', 'SqlServerTable': 'SqlServerTableDataset', 'SapEccResource': 'SapEccResourceDataset', 'SapCloudForCustomerResource': 'SapCloudForCustomerResourceDataset', 'SalesforceObject': 'SalesforceObjectDataset', 'RelationalTable': 'RelationalTableDataset', 'AzureMySqlTable': 'AzureMySqlTableDataset', 'OracleTable': 'OracleTableDataset', 'ODataResource': 'ODataResourceDataset', 'MongoDbCollection': 'MongoDbCollectionDataset', 'FileShare': 'FileShareDataset', 'AzureDataLakeStoreFile': 'AzureDataLakeStoreDataset', 'DynamicsEntity': 'DynamicsEntityDataset', 'DocumentDbCollection': 'DocumentDbCollectionDataset', 'CustomDataset': 'CustomDataset', 'CassandraTable': 'CassandraTableDataset', 'AzureSqlDWTable': 'AzureSqlDWTableDataset', 'AzureSqlTable': 'AzureSqlTableDataset', 'AzureTable': 'AzureTableDataset', 'AzureBlob': 'AzureBlobDataset', 'AmazonS3Object': 'AmazonS3Dataset'} + 'type': {'ResponsysObject': 'ResponsysObjectDataset', 'SalesforceMarketingCloudObject': 'SalesforceMarketingCloudObjectDataset', 'VerticaTable': 'VerticaTableDataset', 'NetezzaTable': 'NetezzaTableDataset', 'ZohoObject': 'ZohoObjectDataset', 'XeroObject': 'XeroObjectDataset', 'SquareObject': 'SquareObjectDataset', 'SparkObject': 'SparkObjectDataset', 'ShopifyObject': 'ShopifyObjectDataset', 'ServiceNowObject': 'ServiceNowObjectDataset', 'QuickBooksObject': 'QuickBooksObjectDataset', 'PrestoObject': 'PrestoObjectDataset', 'PhoenixObject': 'PhoenixObjectDataset', 'PaypalObject': 'PaypalObjectDataset', 'MarketoObject': 'MarketoObjectDataset', 'MariaDBTable': 'MariaDBTableDataset', 'MagentoObject': 'MagentoObjectDataset', 'JiraObject': 'JiraObjectDataset', 'ImpalaObject': 'ImpalaObjectDataset', 'HubspotObject': 'HubspotObjectDataset', 'HiveObject': 'HiveObjectDataset', 'HBaseObject': 'HBaseObjectDataset', 'GreenplumTable': 'GreenplumTableDataset', 'GoogleBigQueryObject': 'GoogleBigQueryObjectDataset', 'EloquaObject': 'EloquaObjectDataset', 'DrillTable': 'DrillTableDataset', 'CouchbaseTable': 'CouchbaseTableDataset', 'ConcurObject': 'ConcurObjectDataset', 'AzurePostgreSqlTable': 'AzurePostgreSqlTableDataset', 'AmazonMWSObject': 'AmazonMWSObjectDataset', 'HttpFile': 'HttpDataset', 'AzureSearchIndex': 'AzureSearchIndexDataset', 'WebTable': 'WebTableDataset', 'SqlServerTable': 'SqlServerTableDataset', 'SapEccResource': 'SapEccResourceDataset', 'SapCloudForCustomerResource': 'SapCloudForCustomerResourceDataset', 'SalesforceObject': 'SalesforceObjectDataset', 'RelationalTable': 'RelationalTableDataset', 'AzureMySqlTable': 'AzureMySqlTableDataset', 'OracleTable': 'OracleTableDataset', 'ODataResource': 'ODataResourceDataset', 'MongoDbCollection': 'MongoDbCollectionDataset', 'FileShare': 'FileShareDataset', 'AzureDataLakeStoreFile': 'AzureDataLakeStoreDataset', 'DynamicsEntity': 'DynamicsEntityDataset', 'DocumentDbCollection': 'DocumentDbCollectionDataset', 'CustomDataset': 'CustomDataset', 'CassandraTable': 'CassandraTableDataset', 'AzureSqlDWTable': 'AzureSqlDWTableDataset', 'AzureSqlTable': 'AzureSqlTableDataset', 'AzureTable': 'AzureTableDataset', 'AzureBlob': 'AzureBlobDataset', 'AmazonS3Object': 'AmazonS3Dataset'} } - def __init__(self, linked_service_name, additional_properties=None, description=None, structure=None, parameters=None, annotations=None): - super(Dataset, self).__init__() - self.additional_properties = additional_properties - self.description = description - self.structure = structure - self.linked_service_name = linked_service_name - self.parameters = parameters - self.annotations = annotations + def __init__(self, **kwargs): + super(Dataset, self).__init__(**kwargs) + self.additional_properties = kwargs.get('additional_properties', None) + self.description = kwargs.get('description', None) + self.structure = kwargs.get('structure', None) + self.linked_service_name = kwargs.get('linked_service_name', None) + self.parameters = kwargs.get('parameters', None) + self.annotations = kwargs.get('annotations', None) self.type = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_bzip2_compression.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_bzip2_compression.py index d31981df108e..71b041c5eb5b 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_bzip2_compression.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_bzip2_compression.py @@ -15,10 +15,12 @@ class DatasetBZip2Compression(DatasetCompression): """The BZip2 compression method used on a dataset. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str """ @@ -26,6 +28,11 @@ class DatasetBZip2Compression(DatasetCompression): 'type': {'required': True}, } - def __init__(self, additional_properties=None): - super(DatasetBZip2Compression, self).__init__(additional_properties=additional_properties) + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, **kwargs): + super(DatasetBZip2Compression, self).__init__(**kwargs) self.type = 'BZip2' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_bzip2_compression_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_bzip2_compression_py3.py new file mode 100644 index 000000000000..fbf34be77561 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_bzip2_compression_py3.py @@ -0,0 +1,38 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset_compression import DatasetCompression + + +class DatasetBZip2Compression(DatasetCompression): + """The BZip2 compression method used on a dataset. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param type: Required. Constant filled by server. + :type type: str + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, *, additional_properties=None, **kwargs) -> None: + super(DatasetBZip2Compression, self).__init__(additional_properties=additional_properties, **kwargs) + self.type = 'BZip2' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_compression.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_compression.py index 78066e1cb4d8..c0c4e3d52624 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_compression.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_compression.py @@ -19,10 +19,12 @@ class DatasetCompression(Model): sub-classes are: DatasetZipDeflateCompression, DatasetDeflateCompression, DatasetGZipCompression, DatasetBZip2Compression + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str """ @@ -39,7 +41,7 @@ class DatasetCompression(Model): 'type': {'ZipDeflate': 'DatasetZipDeflateCompression', 'Deflate': 'DatasetDeflateCompression', 'GZip': 'DatasetGZipCompression', 'BZip2': 'DatasetBZip2Compression'} } - def __init__(self, additional_properties=None): - super(DatasetCompression, self).__init__() - self.additional_properties = additional_properties + def __init__(self, **kwargs): + super(DatasetCompression, self).__init__(**kwargs) + self.additional_properties = kwargs.get('additional_properties', None) self.type = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_compression_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_compression_py3.py new file mode 100644 index 000000000000..3b10abc69abf --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_compression_py3.py @@ -0,0 +1,47 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class DatasetCompression(Model): + """The compression method used on a dataset. + + You probably want to use the sub-classes and not this class directly. Known + sub-classes are: DatasetZipDeflateCompression, DatasetDeflateCompression, + DatasetGZipCompression, DatasetBZip2Compression + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param type: Required. Constant filled by server. + :type type: str + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + _subtype_map = { + 'type': {'ZipDeflate': 'DatasetZipDeflateCompression', 'Deflate': 'DatasetDeflateCompression', 'GZip': 'DatasetGZipCompression', 'BZip2': 'DatasetBZip2Compression'} + } + + def __init__(self, *, additional_properties=None, **kwargs) -> None: + super(DatasetCompression, self).__init__(**kwargs) + self.additional_properties = additional_properties + self.type = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_deflate_compression.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_deflate_compression.py index 7dcf469279c3..c16c0611b364 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_deflate_compression.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_deflate_compression.py @@ -15,10 +15,12 @@ class DatasetDeflateCompression(DatasetCompression): """The Deflate compression method used on a dataset. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param level: The Deflate compression level. Possible values include: 'Optimal', 'Fastest' @@ -35,7 +37,7 @@ class DatasetDeflateCompression(DatasetCompression): 'level': {'key': 'level', 'type': 'str'}, } - def __init__(self, additional_properties=None, level=None): - super(DatasetDeflateCompression, self).__init__(additional_properties=additional_properties) - self.level = level + def __init__(self, **kwargs): + super(DatasetDeflateCompression, self).__init__(**kwargs) + self.level = kwargs.get('level', None) self.type = 'Deflate' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_deflate_compression_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_deflate_compression_py3.py new file mode 100644 index 000000000000..14a7e8be45c9 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_deflate_compression_py3.py @@ -0,0 +1,43 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset_compression import DatasetCompression + + +class DatasetDeflateCompression(DatasetCompression): + """The Deflate compression method used on a dataset. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param type: Required. Constant filled by server. + :type type: str + :param level: The Deflate compression level. Possible values include: + 'Optimal', 'Fastest' + :type level: str or ~azure.mgmt.datafactory.models.DatasetCompressionLevel + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'type': {'key': 'type', 'type': 'str'}, + 'level': {'key': 'level', 'type': 'str'}, + } + + def __init__(self, *, additional_properties=None, level=None, **kwargs) -> None: + super(DatasetDeflateCompression, self).__init__(additional_properties=additional_properties, **kwargs) + self.level = level + self.type = 'Deflate' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_gzip_compression.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_gzip_compression.py index 31cad4106af7..48317d06f34e 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_gzip_compression.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_gzip_compression.py @@ -15,10 +15,12 @@ class DatasetGZipCompression(DatasetCompression): """The GZip compression method used on a dataset. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param level: The GZip compression level. Possible values include: 'Optimal', 'Fastest' @@ -35,7 +37,7 @@ class DatasetGZipCompression(DatasetCompression): 'level': {'key': 'level', 'type': 'str'}, } - def __init__(self, additional_properties=None, level=None): - super(DatasetGZipCompression, self).__init__(additional_properties=additional_properties) - self.level = level + def __init__(self, **kwargs): + super(DatasetGZipCompression, self).__init__(**kwargs) + self.level = kwargs.get('level', None) self.type = 'GZip' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_gzip_compression_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_gzip_compression_py3.py new file mode 100644 index 000000000000..7e8f11cf8e87 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_gzip_compression_py3.py @@ -0,0 +1,43 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset_compression import DatasetCompression + + +class DatasetGZipCompression(DatasetCompression): + """The GZip compression method used on a dataset. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param type: Required. Constant filled by server. + :type type: str + :param level: The GZip compression level. Possible values include: + 'Optimal', 'Fastest' + :type level: str or ~azure.mgmt.datafactory.models.DatasetCompressionLevel + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'type': {'key': 'type', 'type': 'str'}, + 'level': {'key': 'level', 'type': 'str'}, + } + + def __init__(self, *, additional_properties=None, level=None, **kwargs) -> None: + super(DatasetGZipCompression, self).__init__(additional_properties=additional_properties, **kwargs) + self.level = level + self.type = 'GZip' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_py3.py new file mode 100644 index 000000000000..65f43811d7a7 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_py3.py @@ -0,0 +1,91 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class Dataset(Model): + """The Azure Data Factory nested object which identifies data within different + data stores, such as tables, files, folders, and documents. + + You probably want to use the sub-classes and not this class directly. Known + sub-classes are: ResponsysObjectDataset, + SalesforceMarketingCloudObjectDataset, VerticaTableDataset, + NetezzaTableDataset, ZohoObjectDataset, XeroObjectDataset, + SquareObjectDataset, SparkObjectDataset, ShopifyObjectDataset, + ServiceNowObjectDataset, QuickBooksObjectDataset, PrestoObjectDataset, + PhoenixObjectDataset, PaypalObjectDataset, MarketoObjectDataset, + MariaDBTableDataset, MagentoObjectDataset, JiraObjectDataset, + ImpalaObjectDataset, HubspotObjectDataset, HiveObjectDataset, + HBaseObjectDataset, GreenplumTableDataset, GoogleBigQueryObjectDataset, + EloquaObjectDataset, DrillTableDataset, CouchbaseTableDataset, + ConcurObjectDataset, AzurePostgreSqlTableDataset, AmazonMWSObjectDataset, + HttpDataset, AzureSearchIndexDataset, WebTableDataset, + SqlServerTableDataset, SapEccResourceDataset, + SapCloudForCustomerResourceDataset, SalesforceObjectDataset, + RelationalTableDataset, AzureMySqlTableDataset, OracleTableDataset, + ODataResourceDataset, MongoDbCollectionDataset, FileShareDataset, + AzureDataLakeStoreDataset, DynamicsEntityDataset, + DocumentDbCollectionDataset, CustomDataset, CassandraTableDataset, + AzureSqlDWTableDataset, AzureSqlTableDataset, AzureTableDataset, + AzureBlobDataset, AmazonS3Dataset + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Required. Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + _subtype_map = { + 'type': {'ResponsysObject': 'ResponsysObjectDataset', 'SalesforceMarketingCloudObject': 'SalesforceMarketingCloudObjectDataset', 'VerticaTable': 'VerticaTableDataset', 'NetezzaTable': 'NetezzaTableDataset', 'ZohoObject': 'ZohoObjectDataset', 'XeroObject': 'XeroObjectDataset', 'SquareObject': 'SquareObjectDataset', 'SparkObject': 'SparkObjectDataset', 'ShopifyObject': 'ShopifyObjectDataset', 'ServiceNowObject': 'ServiceNowObjectDataset', 'QuickBooksObject': 'QuickBooksObjectDataset', 'PrestoObject': 'PrestoObjectDataset', 'PhoenixObject': 'PhoenixObjectDataset', 'PaypalObject': 'PaypalObjectDataset', 'MarketoObject': 'MarketoObjectDataset', 'MariaDBTable': 'MariaDBTableDataset', 'MagentoObject': 'MagentoObjectDataset', 'JiraObject': 'JiraObjectDataset', 'ImpalaObject': 'ImpalaObjectDataset', 'HubspotObject': 'HubspotObjectDataset', 'HiveObject': 'HiveObjectDataset', 'HBaseObject': 'HBaseObjectDataset', 'GreenplumTable': 'GreenplumTableDataset', 'GoogleBigQueryObject': 'GoogleBigQueryObjectDataset', 'EloquaObject': 'EloquaObjectDataset', 'DrillTable': 'DrillTableDataset', 'CouchbaseTable': 'CouchbaseTableDataset', 'ConcurObject': 'ConcurObjectDataset', 'AzurePostgreSqlTable': 'AzurePostgreSqlTableDataset', 'AmazonMWSObject': 'AmazonMWSObjectDataset', 'HttpFile': 'HttpDataset', 'AzureSearchIndex': 'AzureSearchIndexDataset', 'WebTable': 'WebTableDataset', 'SqlServerTable': 'SqlServerTableDataset', 'SapEccResource': 'SapEccResourceDataset', 'SapCloudForCustomerResource': 'SapCloudForCustomerResourceDataset', 'SalesforceObject': 'SalesforceObjectDataset', 'RelationalTable': 'RelationalTableDataset', 'AzureMySqlTable': 'AzureMySqlTableDataset', 'OracleTable': 'OracleTableDataset', 'ODataResource': 'ODataResourceDataset', 'MongoDbCollection': 'MongoDbCollectionDataset', 'FileShare': 'FileShareDataset', 'AzureDataLakeStoreFile': 'AzureDataLakeStoreDataset', 'DynamicsEntity': 'DynamicsEntityDataset', 'DocumentDbCollection': 'DocumentDbCollectionDataset', 'CustomDataset': 'CustomDataset', 'CassandraTable': 'CassandraTableDataset', 'AzureSqlDWTable': 'AzureSqlDWTableDataset', 'AzureSqlTable': 'AzureSqlTableDataset', 'AzureTable': 'AzureTableDataset', 'AzureBlob': 'AzureBlobDataset', 'AmazonS3Object': 'AmazonS3Dataset'} + } + + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: + super(Dataset, self).__init__(**kwargs) + self.additional_properties = additional_properties + self.description = description + self.structure = structure + self.linked_service_name = linked_service_name + self.parameters = parameters + self.annotations = annotations + self.type = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_reference.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_reference.py index 006074933fe7..ca3d385f31ce 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_reference.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_reference.py @@ -18,9 +18,12 @@ class DatasetReference(Model): Variables are only populated by the server, and will be ignored when sending a request. - :ivar type: Dataset reference type. Default value: "DatasetReference" . + All required parameters must be populated in order to send to Azure. + + :ivar type: Required. Dataset reference type. Default value: + "DatasetReference" . :vartype type: str - :param reference_name: Reference dataset name. + :param reference_name: Required. Reference dataset name. :type reference_name: str :param parameters: Arguments for dataset. :type parameters: dict[str, object] @@ -39,7 +42,7 @@ class DatasetReference(Model): type = "DatasetReference" - def __init__(self, reference_name, parameters=None): - super(DatasetReference, self).__init__() - self.reference_name = reference_name - self.parameters = parameters + def __init__(self, **kwargs): + super(DatasetReference, self).__init__(**kwargs) + self.reference_name = kwargs.get('reference_name', None) + self.parameters = kwargs.get('parameters', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_reference_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_reference_py3.py new file mode 100644 index 000000000000..80162fd77da1 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_reference_py3.py @@ -0,0 +1,48 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class DatasetReference(Model): + """Dataset reference type. + + Variables are only populated by the server, and will be ignored when + sending a request. + + All required parameters must be populated in order to send to Azure. + + :ivar type: Required. Dataset reference type. Default value: + "DatasetReference" . + :vartype type: str + :param reference_name: Required. Reference dataset name. + :type reference_name: str + :param parameters: Arguments for dataset. + :type parameters: dict[str, object] + """ + + _validation = { + 'type': {'required': True, 'constant': True}, + 'reference_name': {'required': True}, + } + + _attribute_map = { + 'type': {'key': 'type', 'type': 'str'}, + 'reference_name': {'key': 'referenceName', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{object}'}, + } + + type = "DatasetReference" + + def __init__(self, *, reference_name: str, parameters=None, **kwargs) -> None: + super(DatasetReference, self).__init__(**kwargs) + self.reference_name = reference_name + self.parameters = parameters diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_resource.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_resource.py index ec6a78c21f79..a68fb563e425 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_resource.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_resource.py @@ -18,6 +18,8 @@ class DatasetResource(SubResource): Variables are only populated by the server, and will be ignored when sending a request. + All required parameters must be populated in order to send to Azure. + :ivar id: The resource identifier. :vartype id: str :ivar name: The resource name. @@ -26,7 +28,7 @@ class DatasetResource(SubResource): :vartype type: str :ivar etag: Etag identifies change in the resource. :vartype etag: str - :param properties: Dataset properties. + :param properties: Required. Dataset properties. :type properties: ~azure.mgmt.datafactory.models.Dataset """ @@ -46,6 +48,6 @@ class DatasetResource(SubResource): 'properties': {'key': 'properties', 'type': 'Dataset'}, } - def __init__(self, properties): - super(DatasetResource, self).__init__() - self.properties = properties + def __init__(self, **kwargs): + super(DatasetResource, self).__init__(**kwargs) + self.properties = kwargs.get('properties', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_resource_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_resource_py3.py new file mode 100644 index 000000000000..1e82ef161d47 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_resource_py3.py @@ -0,0 +1,53 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .sub_resource import SubResource + + +class DatasetResource(SubResource): + """Dataset resource type. + + Variables are only populated by the server, and will be ignored when + sending a request. + + All required parameters must be populated in order to send to Azure. + + :ivar id: The resource identifier. + :vartype id: str + :ivar name: The resource name. + :vartype name: str + :ivar type: The resource type. + :vartype type: str + :ivar etag: Etag identifies change in the resource. + :vartype etag: str + :param properties: Required. Dataset properties. + :type properties: ~azure.mgmt.datafactory.models.Dataset + """ + + _validation = { + 'id': {'readonly': True}, + 'name': {'readonly': True}, + 'type': {'readonly': True}, + 'etag': {'readonly': True}, + 'properties': {'required': True}, + } + + _attribute_map = { + 'id': {'key': 'id', 'type': 'str'}, + 'name': {'key': 'name', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'str'}, + 'etag': {'key': 'etag', 'type': 'str'}, + 'properties': {'key': 'properties', 'type': 'Dataset'}, + } + + def __init__(self, *, properties, **kwargs) -> None: + super(DatasetResource, self).__init__(, **kwargs) + self.properties = properties diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_storage_format.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_storage_format.py index 2ae233fac19f..b3160565230d 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_storage_format.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_storage_format.py @@ -19,6 +19,8 @@ class DatasetStorageFormat(Model): sub-classes are: ParquetFormat, OrcFormat, AvroFormat, JsonFormat, TextFormat + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -28,7 +30,7 @@ class DatasetStorageFormat(Model): :param deserializer: Deserializer. Type: string (or Expression with resultType string). :type deserializer: object - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str """ @@ -47,9 +49,9 @@ class DatasetStorageFormat(Model): 'type': {'ParquetFormat': 'ParquetFormat', 'OrcFormat': 'OrcFormat', 'AvroFormat': 'AvroFormat', 'JsonFormat': 'JsonFormat', 'TextFormat': 'TextFormat'} } - def __init__(self, additional_properties=None, serializer=None, deserializer=None): - super(DatasetStorageFormat, self).__init__() - self.additional_properties = additional_properties - self.serializer = serializer - self.deserializer = deserializer + def __init__(self, **kwargs): + super(DatasetStorageFormat, self).__init__(**kwargs) + self.additional_properties = kwargs.get('additional_properties', None) + self.serializer = kwargs.get('serializer', None) + self.deserializer = kwargs.get('deserializer', None) self.type = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_storage_format_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_storage_format_py3.py new file mode 100644 index 000000000000..faf746642d9e --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_storage_format_py3.py @@ -0,0 +1,57 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class DatasetStorageFormat(Model): + """The format definition of a storage. + + You probably want to use the sub-classes and not this class directly. Known + sub-classes are: ParquetFormat, OrcFormat, AvroFormat, JsonFormat, + TextFormat + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param serializer: Serializer. Type: string (or Expression with resultType + string). + :type serializer: object + :param deserializer: Deserializer. Type: string (or Expression with + resultType string). + :type deserializer: object + :param type: Required. Constant filled by server. + :type type: str + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'serializer': {'key': 'serializer', 'type': 'object'}, + 'deserializer': {'key': 'deserializer', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + _subtype_map = { + 'type': {'ParquetFormat': 'ParquetFormat', 'OrcFormat': 'OrcFormat', 'AvroFormat': 'AvroFormat', 'JsonFormat': 'JsonFormat', 'TextFormat': 'TextFormat'} + } + + def __init__(self, *, additional_properties=None, serializer=None, deserializer=None, **kwargs) -> None: + super(DatasetStorageFormat, self).__init__(**kwargs) + self.additional_properties = additional_properties + self.serializer = serializer + self.deserializer = deserializer + self.type = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_zip_deflate_compression.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_zip_deflate_compression.py index ca8fbf3c768c..9312098be5a3 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_zip_deflate_compression.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_zip_deflate_compression.py @@ -15,10 +15,12 @@ class DatasetZipDeflateCompression(DatasetCompression): """The ZipDeflate compression method used on a dataset. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param level: The ZipDeflate compression level. Possible values include: 'Optimal', 'Fastest' @@ -35,7 +37,7 @@ class DatasetZipDeflateCompression(DatasetCompression): 'level': {'key': 'level', 'type': 'str'}, } - def __init__(self, additional_properties=None, level=None): - super(DatasetZipDeflateCompression, self).__init__(additional_properties=additional_properties) - self.level = level + def __init__(self, **kwargs): + super(DatasetZipDeflateCompression, self).__init__(**kwargs) + self.level = kwargs.get('level', None) self.type = 'ZipDeflate' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_zip_deflate_compression_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_zip_deflate_compression_py3.py new file mode 100644 index 000000000000..c6b0bae79f0c --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_zip_deflate_compression_py3.py @@ -0,0 +1,43 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset_compression import DatasetCompression + + +class DatasetZipDeflateCompression(DatasetCompression): + """The ZipDeflate compression method used on a dataset. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param type: Required. Constant filled by server. + :type type: str + :param level: The ZipDeflate compression level. Possible values include: + 'Optimal', 'Fastest' + :type level: str or ~azure.mgmt.datafactory.models.DatasetCompressionLevel + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'type': {'key': 'type', 'type': 'str'}, + 'level': {'key': 'level', 'type': 'str'}, + } + + def __init__(self, *, additional_properties=None, level=None, **kwargs) -> None: + super(DatasetZipDeflateCompression, self).__init__(additional_properties=additional_properties, **kwargs) + self.level = level + self.type = 'ZipDeflate' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/db2_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/db2_linked_service.py index 031025b25175..a3349c5b7ab2 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/db2_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/db2_linked_service.py @@ -15,6 +15,8 @@ class Db2LinkedService(LinkedService): """Linked service for DB2 data source. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -29,13 +31,13 @@ class Db2LinkedService(LinkedService): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param server: Server name for connection. Type: string (or Expression - with resultType string). + :param server: Required. Server name for connection. Type: string (or + Expression with resultType string). :type server: object - :param database: Database name for connection. Type: string (or Expression - with resultType string). + :param database: Required. Database name for connection. Type: string (or + Expression with resultType string). :type database: object :param schema: Schema name for connection. Type: string (or Expression with resultType string). @@ -77,13 +79,13 @@ class Db2LinkedService(LinkedService): 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, server, database, additional_properties=None, connect_via=None, description=None, parameters=None, annotations=None, schema=None, authentication_type=None, username=None, password=None, encrypted_credential=None): - super(Db2LinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations) - self.server = server - self.database = database - self.schema = schema - self.authentication_type = authentication_type - self.username = username - self.password = password - self.encrypted_credential = encrypted_credential + def __init__(self, **kwargs): + super(Db2LinkedService, self).__init__(**kwargs) + self.server = kwargs.get('server', None) + self.database = kwargs.get('database', None) + self.schema = kwargs.get('schema', None) + self.authentication_type = kwargs.get('authentication_type', None) + self.username = kwargs.get('username', None) + self.password = kwargs.get('password', None) + self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'Db2' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/db2_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/db2_linked_service_py3.py new file mode 100644 index 000000000000..ff86655a9386 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/db2_linked_service_py3.py @@ -0,0 +1,91 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class Db2LinkedService(LinkedService): + """Linked service for DB2 data source. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param server: Required. Server name for connection. Type: string (or + Expression with resultType string). + :type server: object + :param database: Required. Database name for connection. Type: string (or + Expression with resultType string). + :type database: object + :param schema: Schema name for connection. Type: string (or Expression + with resultType string). + :type schema: object + :param authentication_type: AuthenticationType to be used for connection. + Possible values include: 'Basic' + :type authentication_type: str or + ~azure.mgmt.datafactory.models.Db2AuthenticationType + :param username: Username for authentication. Type: string (or Expression + with resultType string). + :type username: object + :param password: Password for authentication. + :type password: ~azure.mgmt.datafactory.models.SecretBase + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'server': {'required': True}, + 'database': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'server': {'key': 'typeProperties.server', 'type': 'object'}, + 'database': {'key': 'typeProperties.database', 'type': 'object'}, + 'schema': {'key': 'typeProperties.schema', 'type': 'object'}, + 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, + 'username': {'key': 'typeProperties.username', 'type': 'object'}, + 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, *, server, database, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, schema=None, authentication_type=None, username=None, password=None, encrypted_credential=None, **kwargs) -> None: + super(Db2LinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.server = server + self.database = database + self.schema = schema + self.authentication_type = authentication_type + self.username = username + self.password = password + self.encrypted_credential = encrypted_credential + self.type = 'Db2' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/distcp_settings.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/distcp_settings.py index fde14d4a8c35..a8065ec3cc06 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/distcp_settings.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/distcp_settings.py @@ -15,13 +15,16 @@ class DistcpSettings(Model): """Distcp settings. - :param resource_manager_endpoint: Specifies the Yarn ResourceManager - endpoint. Type: string (or Expression with resultType string). + All required parameters must be populated in order to send to Azure. + + :param resource_manager_endpoint: Required. Specifies the Yarn + ResourceManager endpoint. Type: string (or Expression with resultType + string). :type resource_manager_endpoint: object - :param temp_script_path: Specifies an existing folder path which will be - used to store temp Distcp command script. The script file is generated by - ADF and will be removed after Copy job finished. Type: string (or - Expression with resultType string). + :param temp_script_path: Required. Specifies an existing folder path which + will be used to store temp Distcp command script. The script file is + generated by ADF and will be removed after Copy job finished. Type: string + (or Expression with resultType string). :type temp_script_path: object :param distcp_options: Specifies the Distcp options. Type: string (or Expression with resultType string). @@ -39,8 +42,8 @@ class DistcpSettings(Model): 'distcp_options': {'key': 'distcpOptions', 'type': 'object'}, } - def __init__(self, resource_manager_endpoint, temp_script_path, distcp_options=None): - super(DistcpSettings, self).__init__() - self.resource_manager_endpoint = resource_manager_endpoint - self.temp_script_path = temp_script_path - self.distcp_options = distcp_options + def __init__(self, **kwargs): + super(DistcpSettings, self).__init__(**kwargs) + self.resource_manager_endpoint = kwargs.get('resource_manager_endpoint', None) + self.temp_script_path = kwargs.get('temp_script_path', None) + self.distcp_options = kwargs.get('distcp_options', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/distcp_settings_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/distcp_settings_py3.py new file mode 100644 index 000000000000..628e2d207f8e --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/distcp_settings_py3.py @@ -0,0 +1,49 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class DistcpSettings(Model): + """Distcp settings. + + All required parameters must be populated in order to send to Azure. + + :param resource_manager_endpoint: Required. Specifies the Yarn + ResourceManager endpoint. Type: string (or Expression with resultType + string). + :type resource_manager_endpoint: object + :param temp_script_path: Required. Specifies an existing folder path which + will be used to store temp Distcp command script. The script file is + generated by ADF and will be removed after Copy job finished. Type: string + (or Expression with resultType string). + :type temp_script_path: object + :param distcp_options: Specifies the Distcp options. Type: string (or + Expression with resultType string). + :type distcp_options: object + """ + + _validation = { + 'resource_manager_endpoint': {'required': True}, + 'temp_script_path': {'required': True}, + } + + _attribute_map = { + 'resource_manager_endpoint': {'key': 'resourceManagerEndpoint', 'type': 'object'}, + 'temp_script_path': {'key': 'tempScriptPath', 'type': 'object'}, + 'distcp_options': {'key': 'distcpOptions', 'type': 'object'}, + } + + def __init__(self, *, resource_manager_endpoint, temp_script_path, distcp_options=None, **kwargs) -> None: + super(DistcpSettings, self).__init__(**kwargs) + self.resource_manager_endpoint = resource_manager_endpoint + self.temp_script_path = temp_script_path + self.distcp_options = distcp_options diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/document_db_collection_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/document_db_collection_dataset.py index 60c3df76af60..58f055d3693e 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/document_db_collection_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/document_db_collection_dataset.py @@ -15,6 +15,8 @@ class DocumentDbCollectionDataset(Dataset): """Microsoft Azure Document Database Collection dataset. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -23,7 +25,7 @@ class DocumentDbCollectionDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object - :param linked_service_name: Linked service reference. + :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference :param parameters: Parameters for dataset. @@ -32,10 +34,10 @@ class DocumentDbCollectionDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param collection_name: Document Database collection name. Type: string - (or Expression with resultType string). + :param collection_name: Required. Document Database collection name. Type: + string (or Expression with resultType string). :type collection_name: object """ @@ -56,7 +58,7 @@ class DocumentDbCollectionDataset(Dataset): 'collection_name': {'key': 'typeProperties.collectionName', 'type': 'object'}, } - def __init__(self, linked_service_name, collection_name, additional_properties=None, description=None, structure=None, parameters=None, annotations=None): - super(DocumentDbCollectionDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations) - self.collection_name = collection_name + def __init__(self, **kwargs): + super(DocumentDbCollectionDataset, self).__init__(**kwargs) + self.collection_name = kwargs.get('collection_name', None) self.type = 'DocumentDbCollection' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/document_db_collection_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/document_db_collection_dataset_py3.py new file mode 100644 index 000000000000..e05756941c1c --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/document_db_collection_dataset_py3.py @@ -0,0 +1,64 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class DocumentDbCollectionDataset(Dataset): + """Microsoft Azure Document Database Collection dataset. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Required. Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param collection_name: Required. Document Database collection name. Type: + string (or Expression with resultType string). + :type collection_name: object + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + 'collection_name': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'collection_name': {'key': 'typeProperties.collectionName', 'type': 'object'}, + } + + def __init__(self, *, linked_service_name, collection_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: + super(DocumentDbCollectionDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + self.collection_name = collection_name + self.type = 'DocumentDbCollection' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/document_db_collection_sink.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/document_db_collection_sink.py index 02e63c1f17ea..43253aff51d0 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/document_db_collection_sink.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/document_db_collection_sink.py @@ -15,6 +15,8 @@ class DocumentDbCollectionSink(CopySink): """A copy activity Document Database Collection sink. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -32,7 +34,7 @@ class DocumentDbCollectionSink(CopySink): resultType string), pattern: ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type sink_retry_wait: object - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param nesting_separator: Nested properties separator. Default is . (dot). Type: string (or Expression with resultType string). @@ -53,7 +55,7 @@ class DocumentDbCollectionSink(CopySink): 'nesting_separator': {'key': 'nestingSeparator', 'type': 'object'}, } - def __init__(self, additional_properties=None, write_batch_size=None, write_batch_timeout=None, sink_retry_count=None, sink_retry_wait=None, nesting_separator=None): - super(DocumentDbCollectionSink, self).__init__(additional_properties=additional_properties, write_batch_size=write_batch_size, write_batch_timeout=write_batch_timeout, sink_retry_count=sink_retry_count, sink_retry_wait=sink_retry_wait) - self.nesting_separator = nesting_separator + def __init__(self, **kwargs): + super(DocumentDbCollectionSink, self).__init__(**kwargs) + self.nesting_separator = kwargs.get('nesting_separator', None) self.type = 'DocumentDbCollectionSink' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/document_db_collection_sink_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/document_db_collection_sink_py3.py new file mode 100644 index 000000000000..be10d5704bb4 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/document_db_collection_sink_py3.py @@ -0,0 +1,61 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_sink import CopySink + + +class DocumentDbCollectionSink(CopySink): + """A copy activity Document Database Collection sink. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param write_batch_size: Write batch size. Type: integer (or Expression + with resultType integer), minimum: 0. + :type write_batch_size: object + :param write_batch_timeout: Write batch timeout. Type: string (or + Expression with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type write_batch_timeout: object + :param sink_retry_count: Sink retry count. Type: integer (or Expression + with resultType integer). + :type sink_retry_count: object + :param sink_retry_wait: Sink retry wait. Type: string (or Expression with + resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type sink_retry_wait: object + :param type: Required. Constant filled by server. + :type type: str + :param nesting_separator: Nested properties separator. Default is . (dot). + Type: string (or Expression with resultType string). + :type nesting_separator: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, + 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, + 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, + 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'nesting_separator': {'key': 'nestingSeparator', 'type': 'object'}, + } + + def __init__(self, *, additional_properties=None, write_batch_size=None, write_batch_timeout=None, sink_retry_count=None, sink_retry_wait=None, nesting_separator=None, **kwargs) -> None: + super(DocumentDbCollectionSink, self).__init__(additional_properties=additional_properties, write_batch_size=write_batch_size, write_batch_timeout=write_batch_timeout, sink_retry_count=sink_retry_count, sink_retry_wait=sink_retry_wait, **kwargs) + self.nesting_separator = nesting_separator + self.type = 'DocumentDbCollectionSink' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/document_db_collection_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/document_db_collection_source.py index d483eb87c3cb..ac6bd77955c8 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/document_db_collection_source.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/document_db_collection_source.py @@ -15,6 +15,8 @@ class DocumentDbCollectionSource(CopySource): """A copy activity Document Database Collection source. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -25,7 +27,7 @@ class DocumentDbCollectionSource(CopySource): with resultType string), pattern: ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type source_retry_wait: object - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param query: Documents query. Type: string (or Expression with resultType string). @@ -48,8 +50,8 @@ class DocumentDbCollectionSource(CopySource): 'nesting_separator': {'key': 'nestingSeparator', 'type': 'object'}, } - def __init__(self, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None, nesting_separator=None): - super(DocumentDbCollectionSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait) - self.query = query - self.nesting_separator = nesting_separator + def __init__(self, **kwargs): + super(DocumentDbCollectionSource, self).__init__(**kwargs) + self.query = kwargs.get('query', None) + self.nesting_separator = kwargs.get('nesting_separator', None) self.type = 'DocumentDbCollectionSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/document_db_collection_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/document_db_collection_source_py3.py new file mode 100644 index 000000000000..7e72aa6f9800 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/document_db_collection_source_py3.py @@ -0,0 +1,57 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class DocumentDbCollectionSource(CopySource): + """A copy activity Document Database Collection source. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Required. Constant filled by server. + :type type: str + :param query: Documents query. Type: string (or Expression with resultType + string). + :type query: object + :param nesting_separator: Nested properties separator. Type: string (or + Expression with resultType string). + :type nesting_separator: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'query': {'key': 'query', 'type': 'object'}, + 'nesting_separator': {'key': 'nestingSeparator', 'type': 'object'}, + } + + def __init__(self, *, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None, nesting_separator=None, **kwargs) -> None: + super(DocumentDbCollectionSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, **kwargs) + self.query = query + self.nesting_separator = nesting_separator + self.type = 'DocumentDbCollectionSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_linked_service.py index ab34fd2c56a4..2cd19dd6fdd8 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_linked_service.py @@ -15,6 +15,8 @@ class DrillLinkedService(LinkedService): """Drill server linked service. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -29,7 +31,7 @@ class DrillLinkedService(LinkedService): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param connection_string: An ODBC connection string. :type connection_string: ~azure.mgmt.datafactory.models.SecretBase @@ -54,8 +56,8 @@ class DrillLinkedService(LinkedService): 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, additional_properties=None, connect_via=None, description=None, parameters=None, annotations=None, connection_string=None, encrypted_credential=None): - super(DrillLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations) - self.connection_string = connection_string - self.encrypted_credential = encrypted_credential + def __init__(self, **kwargs): + super(DrillLinkedService, self).__init__(**kwargs) + self.connection_string = kwargs.get('connection_string', None) + self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'Drill' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_linked_service_py3.py new file mode 100644 index 000000000000..7deba2b1e73d --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_linked_service_py3.py @@ -0,0 +1,63 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class DrillLinkedService(LinkedService): + """Drill server linked service. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param connection_string: An ODBC connection string. + :type connection_string: ~azure.mgmt.datafactory.models.SecretBase + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'SecretBase'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, *, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, connection_string=None, encrypted_credential=None, **kwargs) -> None: + super(DrillLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.connection_string = connection_string + self.encrypted_credential = encrypted_credential + self.type = 'Drill' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_source.py index e265ae6450ae..c2e390308b81 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_source.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_source.py @@ -15,6 +15,8 @@ class DrillSource(CopySource): """A copy activity Drill server source. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -25,7 +27,7 @@ class DrillSource(CopySource): with resultType string), pattern: ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type source_retry_wait: object - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). @@ -44,7 +46,7 @@ class DrillSource(CopySource): 'query': {'key': 'query', 'type': 'object'}, } - def __init__(self, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None): - super(DrillSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait) - self.query = query + def __init__(self, **kwargs): + super(DrillSource, self).__init__(**kwargs) + self.query = kwargs.get('query', None) self.type = 'DrillSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_source_py3.py new file mode 100644 index 000000000000..a68ea79a5238 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_source_py3.py @@ -0,0 +1,52 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class DrillSource(CopySource): + """A copy activity Drill server source. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Required. Constant filled by server. + :type type: str + :param query: A query to retrieve data from source. Type: string (or + Expression with resultType string). + :type query: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'query': {'key': 'query', 'type': 'object'}, + } + + def __init__(self, *, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None, **kwargs) -> None: + super(DrillSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, **kwargs) + self.query = query + self.type = 'DrillSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_table_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_table_dataset.py index 2731f23a6f10..21e7e392ec7b 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_table_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_table_dataset.py @@ -15,6 +15,8 @@ class DrillTableDataset(Dataset): """Drill server dataset. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -23,7 +25,7 @@ class DrillTableDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object - :param linked_service_name: Linked service reference. + :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference :param parameters: Parameters for dataset. @@ -32,7 +34,7 @@ class DrillTableDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str """ @@ -41,6 +43,16 @@ class DrillTableDataset(Dataset): 'type': {'required': True}, } - def __init__(self, linked_service_name, additional_properties=None, description=None, structure=None, parameters=None, annotations=None): - super(DrillTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations) + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, **kwargs): + super(DrillTableDataset, self).__init__(**kwargs) self.type = 'DrillTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_table_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_table_dataset_py3.py new file mode 100644 index 000000000000..325085161353 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_table_dataset_py3.py @@ -0,0 +1,58 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class DrillTableDataset(Dataset): + """Drill server dataset. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Required. Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: + super(DrillTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + self.type = 'DrillTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_entity_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_entity_dataset.py index 212dc23594c4..23eb34e5c6b2 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_entity_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_entity_dataset.py @@ -15,6 +15,8 @@ class DynamicsEntityDataset(Dataset): """The Dynamics entity dataset. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -23,7 +25,7 @@ class DynamicsEntityDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object - :param linked_service_name: Linked service reference. + :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference :param parameters: Parameters for dataset. @@ -32,7 +34,7 @@ class DynamicsEntityDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param entity_name: The logical name of the entity. Type: string (or Expression with resultType string). @@ -55,7 +57,7 @@ class DynamicsEntityDataset(Dataset): 'entity_name': {'key': 'typeProperties.entityName', 'type': 'object'}, } - def __init__(self, linked_service_name, additional_properties=None, description=None, structure=None, parameters=None, annotations=None, entity_name=None): - super(DynamicsEntityDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations) - self.entity_name = entity_name + def __init__(self, **kwargs): + super(DynamicsEntityDataset, self).__init__(**kwargs) + self.entity_name = kwargs.get('entity_name', None) self.type = 'DynamicsEntity' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_entity_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_entity_dataset_py3.py new file mode 100644 index 000000000000..50c02ae54167 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_entity_dataset_py3.py @@ -0,0 +1,63 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class DynamicsEntityDataset(Dataset): + """The Dynamics entity dataset. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Required. Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param entity_name: The logical name of the entity. Type: string (or + Expression with resultType string). + :type entity_name: object + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'entity_name': {'key': 'typeProperties.entityName', 'type': 'object'}, + } + + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, entity_name=None, **kwargs) -> None: + super(DynamicsEntityDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + self.entity_name = entity_name + self.type = 'DynamicsEntity' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_linked_service.py index 8161c702bd69..50ec75f79523 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_linked_service.py @@ -15,6 +15,8 @@ class DynamicsLinkedService(LinkedService): """Dynamics linked service. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -29,11 +31,12 @@ class DynamicsLinkedService(LinkedService): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param deployment_type: The deployment type of the Dynamics instance. - 'Online' for Dynamics Online and 'OnPremisesWithIfd' for Dynamics - on-premises with Ifd. Type: string (or Expression with resultType string). + :param deployment_type: Required. The deployment type of the Dynamics + instance. 'Online' for Dynamics Online and 'OnPremisesWithIfd' for + Dynamics on-premises with Ifd. Type: string (or Expression with resultType + string). :type deployment_type: object :param host_name: The host name of the on-premises Dynamics server. The property is required for on-prem and not allowed for online. Type: string @@ -52,12 +55,12 @@ class DynamicsLinkedService(LinkedService): are more than one Dynamics instances associated with the user. Type: string (or Expression with resultType string). :type organization_name: object - :param authentication_type: The authentication type to connect to Dynamics - server. 'Office365' for online scenario, 'Ifd' for on-premises with Ifd - scenario. Type: string (or Expression with resultType string). + :param authentication_type: Required. The authentication type to connect + to Dynamics server. 'Office365' for online scenario, 'Ifd' for on-premises + with Ifd scenario. Type: string (or Expression with resultType string). :type authentication_type: object - :param username: User name to access the Dynamics instance. Type: string - (or Expression with resultType string). + :param username: Required. User name to access the Dynamics instance. + Type: string (or Expression with resultType string). :type username: object :param password: Password to access the Dynamics instance. :type password: ~azure.mgmt.datafactory.models.SecretBase @@ -92,15 +95,15 @@ class DynamicsLinkedService(LinkedService): 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, deployment_type, authentication_type, username, additional_properties=None, connect_via=None, description=None, parameters=None, annotations=None, host_name=None, port=None, service_uri=None, organization_name=None, password=None, encrypted_credential=None): - super(DynamicsLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations) - self.deployment_type = deployment_type - self.host_name = host_name - self.port = port - self.service_uri = service_uri - self.organization_name = organization_name - self.authentication_type = authentication_type - self.username = username - self.password = password - self.encrypted_credential = encrypted_credential + def __init__(self, **kwargs): + super(DynamicsLinkedService, self).__init__(**kwargs) + self.deployment_type = kwargs.get('deployment_type', None) + self.host_name = kwargs.get('host_name', None) + self.port = kwargs.get('port', None) + self.service_uri = kwargs.get('service_uri', None) + self.organization_name = kwargs.get('organization_name', None) + self.authentication_type = kwargs.get('authentication_type', None) + self.username = kwargs.get('username', None) + self.password = kwargs.get('password', None) + self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'Dynamics' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_linked_service_py3.py new file mode 100644 index 000000000000..f5349731c817 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_linked_service_py3.py @@ -0,0 +1,109 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class DynamicsLinkedService(LinkedService): + """Dynamics linked service. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param deployment_type: Required. The deployment type of the Dynamics + instance. 'Online' for Dynamics Online and 'OnPremisesWithIfd' for + Dynamics on-premises with Ifd. Type: string (or Expression with resultType + string). + :type deployment_type: object + :param host_name: The host name of the on-premises Dynamics server. The + property is required for on-prem and not allowed for online. Type: string + (or Expression with resultType string). + :type host_name: object + :param port: The port of on-premises Dynamics server. The property is + required for on-prem and not allowed for online. Default is 443. Type: + integer (or Expression with resultType integer), minimum: 0. + :type port: object + :param service_uri: The URL to the Microsoft Dynamics server. The property + is required for on-line and not allowed for on-prem. Type: string (or + Expression with resultType string). + :type service_uri: object + :param organization_name: The organization name of the Dynamics instance. + The property is required for on-prem and required for online when there + are more than one Dynamics instances associated with the user. Type: + string (or Expression with resultType string). + :type organization_name: object + :param authentication_type: Required. The authentication type to connect + to Dynamics server. 'Office365' for online scenario, 'Ifd' for on-premises + with Ifd scenario. Type: string (or Expression with resultType string). + :type authentication_type: object + :param username: Required. User name to access the Dynamics instance. + Type: string (or Expression with resultType string). + :type username: object + :param password: Password to access the Dynamics instance. + :type password: ~azure.mgmt.datafactory.models.SecretBase + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'deployment_type': {'required': True}, + 'authentication_type': {'required': True}, + 'username': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'deployment_type': {'key': 'typeProperties.deploymentType', 'type': 'object'}, + 'host_name': {'key': 'typeProperties.hostName', 'type': 'object'}, + 'port': {'key': 'typeProperties.port', 'type': 'object'}, + 'service_uri': {'key': 'typeProperties.serviceUri', 'type': 'object'}, + 'organization_name': {'key': 'typeProperties.organizationName', 'type': 'object'}, + 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'object'}, + 'username': {'key': 'typeProperties.username', 'type': 'object'}, + 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, *, deployment_type, authentication_type, username, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, host_name=None, port=None, service_uri=None, organization_name=None, password=None, encrypted_credential=None, **kwargs) -> None: + super(DynamicsLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.deployment_type = deployment_type + self.host_name = host_name + self.port = port + self.service_uri = service_uri + self.organization_name = organization_name + self.authentication_type = authentication_type + self.username = username + self.password = password + self.encrypted_credential = encrypted_credential + self.type = 'Dynamics' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_sink.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_sink.py index 756cbc8879fd..7eb8be963583 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_sink.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_sink.py @@ -18,6 +18,8 @@ class DynamicsSink(CopySink): Variables are only populated by the server, and will be ignored when sending a request. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -35,10 +37,10 @@ class DynamicsSink(CopySink): resultType string), pattern: ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type sink_retry_wait: object - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :ivar write_behavior: The write behavior for the operation. Default value: - "Upsert" . + :ivar write_behavior: Required. The write behavior for the operation. + Default value: "Upsert" . :vartype write_behavior: str :param ignore_null_values: The flag indicating whether ignore null values from input dataset (except key fields) during write operation. Default is @@ -64,7 +66,7 @@ class DynamicsSink(CopySink): write_behavior = "Upsert" - def __init__(self, additional_properties=None, write_batch_size=None, write_batch_timeout=None, sink_retry_count=None, sink_retry_wait=None, ignore_null_values=None): - super(DynamicsSink, self).__init__(additional_properties=additional_properties, write_batch_size=write_batch_size, write_batch_timeout=write_batch_timeout, sink_retry_count=sink_retry_count, sink_retry_wait=sink_retry_wait) - self.ignore_null_values = ignore_null_values + def __init__(self, **kwargs): + super(DynamicsSink, self).__init__(**kwargs) + self.ignore_null_values = kwargs.get('ignore_null_values', None) self.type = 'DynamicsSink' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_sink_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_sink_py3.py new file mode 100644 index 000000000000..6b301a3f75d7 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_sink_py3.py @@ -0,0 +1,72 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_sink import CopySink + + +class DynamicsSink(CopySink): + """A copy activity Dynamics sink. + + Variables are only populated by the server, and will be ignored when + sending a request. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param write_batch_size: Write batch size. Type: integer (or Expression + with resultType integer), minimum: 0. + :type write_batch_size: object + :param write_batch_timeout: Write batch timeout. Type: string (or + Expression with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type write_batch_timeout: object + :param sink_retry_count: Sink retry count. Type: integer (or Expression + with resultType integer). + :type sink_retry_count: object + :param sink_retry_wait: Sink retry wait. Type: string (or Expression with + resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type sink_retry_wait: object + :param type: Required. Constant filled by server. + :type type: str + :ivar write_behavior: Required. The write behavior for the operation. + Default value: "Upsert" . + :vartype write_behavior: str + :param ignore_null_values: The flag indicating whether ignore null values + from input dataset (except key fields) during write operation. Default is + false. Type: boolean (or Expression with resultType boolean). + :type ignore_null_values: object + """ + + _validation = { + 'type': {'required': True}, + 'write_behavior': {'required': True, 'constant': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, + 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, + 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, + 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'write_behavior': {'key': 'writeBehavior', 'type': 'str'}, + 'ignore_null_values': {'key': 'ignoreNullValues', 'type': 'object'}, + } + + write_behavior = "Upsert" + + def __init__(self, *, additional_properties=None, write_batch_size=None, write_batch_timeout=None, sink_retry_count=None, sink_retry_wait=None, ignore_null_values=None, **kwargs) -> None: + super(DynamicsSink, self).__init__(additional_properties=additional_properties, write_batch_size=write_batch_size, write_batch_timeout=write_batch_timeout, sink_retry_count=sink_retry_count, sink_retry_wait=sink_retry_wait, **kwargs) + self.ignore_null_values = ignore_null_values + self.type = 'DynamicsSink' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_source.py index 0ca0cfd7dd0f..09c04a8d09a6 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_source.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_source.py @@ -15,6 +15,8 @@ class DynamicsSource(CopySource): """A copy activity Dynamics source. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -25,7 +27,7 @@ class DynamicsSource(CopySource): with resultType string), pattern: ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type source_retry_wait: object - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param query: FetchXML is a proprietary query language that is used in Microsoft Dynamics (online & on-premises). Type: string (or Expression @@ -45,7 +47,7 @@ class DynamicsSource(CopySource): 'query': {'key': 'query', 'type': 'object'}, } - def __init__(self, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None): - super(DynamicsSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait) - self.query = query + def __init__(self, **kwargs): + super(DynamicsSource, self).__init__(**kwargs) + self.query = kwargs.get('query', None) self.type = 'DynamicsSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_source_py3.py new file mode 100644 index 000000000000..a01986da60db --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_source_py3.py @@ -0,0 +1,53 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class DynamicsSource(CopySource): + """A copy activity Dynamics source. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Required. Constant filled by server. + :type type: str + :param query: FetchXML is a proprietary query language that is used in + Microsoft Dynamics (online & on-premises). Type: string (or Expression + with resultType string). + :type query: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'query': {'key': 'query', 'type': 'object'}, + } + + def __init__(self, *, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None, **kwargs) -> None: + super(DynamicsSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, **kwargs) + self.query = query + self.type = 'DynamicsSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/eloqua_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/eloqua_linked_service.py index d7914f0db7ce..ae14064ae523 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/eloqua_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/eloqua_linked_service.py @@ -15,6 +15,8 @@ class EloquaLinkedService(LinkedService): """Eloqua server linked service. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -29,13 +31,13 @@ class EloquaLinkedService(LinkedService): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param endpoint: The endpoint of the Eloqua server. (i.e. + :param endpoint: Required. The endpoint of the Eloqua server. (i.e. eloqua.example.com) :type endpoint: object - :param username: The site name and user name of your Eloqua account in the - form: sitename/username. (i.e. Eloqua/Alice) + :param username: Required. The site name and user name of your Eloqua + account in the form: sitename/username. (i.e. Eloqua/Alice) :type username: object :param password: The password corresponding to the user name. :type password: ~azure.mgmt.datafactory.models.SecretBase @@ -77,13 +79,13 @@ class EloquaLinkedService(LinkedService): 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, endpoint, username, additional_properties=None, connect_via=None, description=None, parameters=None, annotations=None, password=None, use_encrypted_endpoints=None, use_host_verification=None, use_peer_verification=None, encrypted_credential=None): - super(EloquaLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations) - self.endpoint = endpoint - self.username = username - self.password = password - self.use_encrypted_endpoints = use_encrypted_endpoints - self.use_host_verification = use_host_verification - self.use_peer_verification = use_peer_verification - self.encrypted_credential = encrypted_credential + def __init__(self, **kwargs): + super(EloquaLinkedService, self).__init__(**kwargs) + self.endpoint = kwargs.get('endpoint', None) + self.username = kwargs.get('username', None) + self.password = kwargs.get('password', None) + self.use_encrypted_endpoints = kwargs.get('use_encrypted_endpoints', None) + self.use_host_verification = kwargs.get('use_host_verification', None) + self.use_peer_verification = kwargs.get('use_peer_verification', None) + self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'Eloqua' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/eloqua_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/eloqua_linked_service_py3.py new file mode 100644 index 000000000000..ec0af56a9dd1 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/eloqua_linked_service_py3.py @@ -0,0 +1,91 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class EloquaLinkedService(LinkedService): + """Eloqua server linked service. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param endpoint: Required. The endpoint of the Eloqua server. (i.e. + eloqua.example.com) + :type endpoint: object + :param username: Required. The site name and user name of your Eloqua + account in the form: sitename/username. (i.e. Eloqua/Alice) + :type username: object + :param password: The password corresponding to the user name. + :type password: ~azure.mgmt.datafactory.models.SecretBase + :param use_encrypted_endpoints: Specifies whether the data source + endpoints are encrypted using HTTPS. The default value is true. + :type use_encrypted_endpoints: object + :param use_host_verification: Specifies whether to require the host name + in the server's certificate to match the host name of the server when + connecting over SSL. The default value is true. + :type use_host_verification: object + :param use_peer_verification: Specifies whether to verify the identity of + the server when connecting over SSL. The default value is true. + :type use_peer_verification: object + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'endpoint': {'required': True}, + 'username': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'endpoint': {'key': 'typeProperties.endpoint', 'type': 'object'}, + 'username': {'key': 'typeProperties.username', 'type': 'object'}, + 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, + 'use_encrypted_endpoints': {'key': 'typeProperties.useEncryptedEndpoints', 'type': 'object'}, + 'use_host_verification': {'key': 'typeProperties.useHostVerification', 'type': 'object'}, + 'use_peer_verification': {'key': 'typeProperties.usePeerVerification', 'type': 'object'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, *, endpoint, username, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, password=None, use_encrypted_endpoints=None, use_host_verification=None, use_peer_verification=None, encrypted_credential=None, **kwargs) -> None: + super(EloquaLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.endpoint = endpoint + self.username = username + self.password = password + self.use_encrypted_endpoints = use_encrypted_endpoints + self.use_host_verification = use_host_verification + self.use_peer_verification = use_peer_verification + self.encrypted_credential = encrypted_credential + self.type = 'Eloqua' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/eloqua_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/eloqua_object_dataset.py index d8a0b7d91d75..95d41cef4d5e 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/eloqua_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/eloqua_object_dataset.py @@ -15,6 +15,8 @@ class EloquaObjectDataset(Dataset): """Eloqua server dataset. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -23,7 +25,7 @@ class EloquaObjectDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object - :param linked_service_name: Linked service reference. + :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference :param parameters: Parameters for dataset. @@ -32,7 +34,7 @@ class EloquaObjectDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str """ @@ -41,6 +43,16 @@ class EloquaObjectDataset(Dataset): 'type': {'required': True}, } - def __init__(self, linked_service_name, additional_properties=None, description=None, structure=None, parameters=None, annotations=None): - super(EloquaObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations) + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, **kwargs): + super(EloquaObjectDataset, self).__init__(**kwargs) self.type = 'EloquaObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/eloqua_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/eloqua_object_dataset_py3.py new file mode 100644 index 000000000000..4150adf6e7ad --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/eloqua_object_dataset_py3.py @@ -0,0 +1,58 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class EloquaObjectDataset(Dataset): + """Eloqua server dataset. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Required. Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: + super(EloquaObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + self.type = 'EloquaObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/eloqua_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/eloqua_source.py index 18725852d52b..694282ebcd8a 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/eloqua_source.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/eloqua_source.py @@ -15,6 +15,8 @@ class EloquaSource(CopySource): """A copy activity Eloqua server source. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -25,7 +27,7 @@ class EloquaSource(CopySource): with resultType string), pattern: ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type source_retry_wait: object - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). @@ -44,7 +46,7 @@ class EloquaSource(CopySource): 'query': {'key': 'query', 'type': 'object'}, } - def __init__(self, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None): - super(EloquaSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait) - self.query = query + def __init__(self, **kwargs): + super(EloquaSource, self).__init__(**kwargs) + self.query = kwargs.get('query', None) self.type = 'EloquaSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/eloqua_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/eloqua_source_py3.py new file mode 100644 index 000000000000..f613b9a0e905 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/eloqua_source_py3.py @@ -0,0 +1,52 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class EloquaSource(CopySource): + """A copy activity Eloqua server source. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Required. Constant filled by server. + :type type: str + :param query: A query to retrieve data from source. Type: string (or + Expression with resultType string). + :type query: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'query': {'key': 'query', 'type': 'object'}, + } + + def __init__(self, *, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None, **kwargs) -> None: + super(EloquaSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, **kwargs) + self.query = query + self.type = 'EloquaSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/error_response.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/error_response.py index 5ff54055741a..2a48122fa578 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/error_response.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/error_response.py @@ -16,9 +16,11 @@ class ErrorResponse(Model): """The object that defines the structure of an Azure Data Factory response. - :param code: Error code. + All required parameters must be populated in order to send to Azure. + + :param code: Required. Error code. :type code: str - :param message: Error message. + :param message: Required. Error message. :type message: str :param target: Property name/path in request associated with error. :type target: str @@ -38,12 +40,12 @@ class ErrorResponse(Model): 'details': {'key': 'details', 'type': '[ErrorResponse]'}, } - def __init__(self, code, message, target=None, details=None): - super(ErrorResponse, self).__init__() - self.code = code - self.message = message - self.target = target - self.details = details + def __init__(self, **kwargs): + super(ErrorResponse, self).__init__(**kwargs) + self.code = kwargs.get('code', None) + self.message = kwargs.get('message', None) + self.target = kwargs.get('target', None) + self.details = kwargs.get('details', None) class ErrorResponseException(HttpOperationError): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/error_response_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/error_response_py3.py new file mode 100644 index 000000000000..6b36d51c90df --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/error_response_py3.py @@ -0,0 +1,60 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model +from msrest.exceptions import HttpOperationError + + +class ErrorResponse(Model): + """The object that defines the structure of an Azure Data Factory response. + + All required parameters must be populated in order to send to Azure. + + :param code: Required. Error code. + :type code: str + :param message: Required. Error message. + :type message: str + :param target: Property name/path in request associated with error. + :type target: str + :param details: Array with additional error details. + :type details: list[~azure.mgmt.datafactory.models.ErrorResponse] + """ + + _validation = { + 'code': {'required': True}, + 'message': {'required': True}, + } + + _attribute_map = { + 'code': {'key': 'code', 'type': 'str'}, + 'message': {'key': 'message', 'type': 'str'}, + 'target': {'key': 'target', 'type': 'str'}, + 'details': {'key': 'details', 'type': '[ErrorResponse]'}, + } + + def __init__(self, *, code: str, message: str, target: str=None, details=None, **kwargs) -> None: + super(ErrorResponse, self).__init__(**kwargs) + self.code = code + self.message = message + self.target = target + self.details = details + + +class ErrorResponseException(HttpOperationError): + """Server responsed with exception of type: 'ErrorResponse'. + + :param deserialize: A deserializer + :param response: Server response to be deserialized. + """ + + def __init__(self, deserialize, response, *args): + + super(ErrorResponseException, self).__init__(deserialize, response, 'ErrorResponse', *args) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_pipeline_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_pipeline_activity.py index a558b2f78d82..4e3b48e75bdf 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_pipeline_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_pipeline_activity.py @@ -15,18 +15,20 @@ class ExecutePipelineActivity(ControlActivity): """Execute pipeline activity. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] - :param name: Activity name. + :param name: Required. Activity name. :type name: str :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param pipeline: Pipeline reference. + :param pipeline: Required. Pipeline reference. :type pipeline: ~azure.mgmt.datafactory.models.PipelineReference :param parameters: Pipeline parameters. :type parameters: dict[str, object] @@ -52,9 +54,9 @@ class ExecutePipelineActivity(ControlActivity): 'wait_on_completion': {'key': 'typeProperties.waitOnCompletion', 'type': 'bool'}, } - def __init__(self, name, pipeline, additional_properties=None, description=None, depends_on=None, parameters=None, wait_on_completion=None): - super(ExecutePipelineActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on) - self.pipeline = pipeline - self.parameters = parameters - self.wait_on_completion = wait_on_completion + def __init__(self, **kwargs): + super(ExecutePipelineActivity, self).__init__(**kwargs) + self.pipeline = kwargs.get('pipeline', None) + self.parameters = kwargs.get('parameters', None) + self.wait_on_completion = kwargs.get('wait_on_completion', None) self.type = 'ExecutePipeline' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_pipeline_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_pipeline_activity_py3.py new file mode 100644 index 000000000000..72fd5fa3015e --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_pipeline_activity_py3.py @@ -0,0 +1,62 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .control_activity import ControlActivity + + +class ExecutePipelineActivity(ControlActivity): + """Execute pipeline activity. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param name: Required. Activity name. + :type name: str + :param description: Activity description. + :type description: str + :param depends_on: Activity depends on condition. + :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param type: Required. Constant filled by server. + :type type: str + :param pipeline: Required. Pipeline reference. + :type pipeline: ~azure.mgmt.datafactory.models.PipelineReference + :param parameters: Pipeline parameters. + :type parameters: dict[str, object] + :param wait_on_completion: Defines whether activity execution will wait + for the dependent pipeline execution to finish. Default is false. + :type wait_on_completion: bool + """ + + _validation = { + 'name': {'required': True}, + 'type': {'required': True}, + 'pipeline': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'name': {'key': 'name', 'type': 'str'}, + 'description': {'key': 'description', 'type': 'str'}, + 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'pipeline': {'key': 'typeProperties.pipeline', 'type': 'PipelineReference'}, + 'parameters': {'key': 'typeProperties.parameters', 'type': '{object}'}, + 'wait_on_completion': {'key': 'typeProperties.waitOnCompletion', 'type': 'bool'}, + } + + def __init__(self, *, name: str, pipeline, additional_properties=None, description: str=None, depends_on=None, parameters=None, wait_on_completion: bool=None, **kwargs) -> None: + super(ExecutePipelineActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, **kwargs) + self.pipeline = pipeline + self.parameters = parameters + self.wait_on_completion = wait_on_completion + self.type = 'ExecutePipeline' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_ssis_package_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_ssis_package_activity.py index 70aa00448f34..a72ca4bbafdb 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_ssis_package_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_ssis_package_activity.py @@ -15,23 +15,25 @@ class ExecuteSSISPackageActivity(ExecutionActivity): """Execute SSIS package activity. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] - :param name: Activity name. + :param name: Required. Activity name. :type name: str :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference :param policy: Activity policy. :type policy: ~azure.mgmt.datafactory.models.ActivityPolicy - :param package_location: SSIS package location. + :param package_location: Required. SSIS package location. :type package_location: ~azure.mgmt.datafactory.models.SSISPackageLocation :param runtime: Specifies the runtime to execute SSIS package. Possible values include: 'x64', 'x86' @@ -41,7 +43,7 @@ class ExecuteSSISPackageActivity(ExecutionActivity): :param environment_path: The environment path to execution the SSIS package. :type environment_path: str - :param connect_via: The integration runtime reference. + :param connect_via: Required. The integration runtime reference. :type connect_via: ~azure.mgmt.datafactory.models.IntegrationRuntimeReference """ @@ -68,11 +70,11 @@ class ExecuteSSISPackageActivity(ExecutionActivity): 'connect_via': {'key': 'typeProperties.connectVia', 'type': 'IntegrationRuntimeReference'}, } - def __init__(self, name, package_location, connect_via, additional_properties=None, description=None, depends_on=None, linked_service_name=None, policy=None, runtime=None, logging_level=None, environment_path=None): - super(ExecuteSSISPackageActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, linked_service_name=linked_service_name, policy=policy) - self.package_location = package_location - self.runtime = runtime - self.logging_level = logging_level - self.environment_path = environment_path - self.connect_via = connect_via + def __init__(self, **kwargs): + super(ExecuteSSISPackageActivity, self).__init__(**kwargs) + self.package_location = kwargs.get('package_location', None) + self.runtime = kwargs.get('runtime', None) + self.logging_level = kwargs.get('logging_level', None) + self.environment_path = kwargs.get('environment_path', None) + self.connect_via = kwargs.get('connect_via', None) self.type = 'ExecuteSSISPackage' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_ssis_package_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_ssis_package_activity_py3.py new file mode 100644 index 000000000000..df053457dd16 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_ssis_package_activity_py3.py @@ -0,0 +1,80 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .execution_activity import ExecutionActivity + + +class ExecuteSSISPackageActivity(ExecutionActivity): + """Execute SSIS package activity. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param name: Required. Activity name. + :type name: str + :param description: Activity description. + :type description: str + :param depends_on: Activity depends on condition. + :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param type: Required. Constant filled by server. + :type type: str + :param linked_service_name: Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param policy: Activity policy. + :type policy: ~azure.mgmt.datafactory.models.ActivityPolicy + :param package_location: Required. SSIS package location. + :type package_location: ~azure.mgmt.datafactory.models.SSISPackageLocation + :param runtime: Specifies the runtime to execute SSIS package. Possible + values include: 'x64', 'x86' + :type runtime: str or ~azure.mgmt.datafactory.models.SSISExecutionRuntime + :param logging_level: The logging level of SSIS package execution. + :type logging_level: str + :param environment_path: The environment path to execution the SSIS + package. + :type environment_path: str + :param connect_via: Required. The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + """ + + _validation = { + 'name': {'required': True}, + 'type': {'required': True}, + 'package_location': {'required': True}, + 'connect_via': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'name': {'key': 'name', 'type': 'str'}, + 'description': {'key': 'description', 'type': 'str'}, + 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, + 'package_location': {'key': 'typeProperties.packageLocation', 'type': 'SSISPackageLocation'}, + 'runtime': {'key': 'typeProperties.runtime', 'type': 'str'}, + 'logging_level': {'key': 'typeProperties.loggingLevel', 'type': 'str'}, + 'environment_path': {'key': 'typeProperties.environmentPath', 'type': 'str'}, + 'connect_via': {'key': 'typeProperties.connectVia', 'type': 'IntegrationRuntimeReference'}, + } + + def __init__(self, *, name: str, package_location, connect_via, additional_properties=None, description: str=None, depends_on=None, linked_service_name=None, policy=None, runtime=None, logging_level: str=None, environment_path: str=None, **kwargs) -> None: + super(ExecuteSSISPackageActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, linked_service_name=linked_service_name, policy=policy, **kwargs) + self.package_location = package_location + self.runtime = runtime + self.logging_level = logging_level + self.environment_path = environment_path + self.connect_via = connect_via + self.type = 'ExecuteSSISPackage' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execution_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execution_activity.py index c65430db8229..736ece86c29c 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execution_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execution_activity.py @@ -24,16 +24,18 @@ class ExecutionActivity(Activity): HDInsightStreamingActivity, HDInsightMapReduceActivity, HDInsightPigActivity, HDInsightHiveActivity, CopyActivity + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] - :param name: Activity name. + :param name: Required. Activity name. :type name: str :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. :type linked_service_name: @@ -61,8 +63,8 @@ class ExecutionActivity(Activity): 'type': {'DatabricksNotebook': 'DatabricksNotebookActivity', 'DataLakeAnalyticsU-SQL': 'DataLakeAnalyticsUSQLActivity', 'AzureMLUpdateResource': 'AzureMLUpdateResourceActivity', 'AzureMLBatchExecution': 'AzureMLBatchExecutionActivity', 'GetMetadata': 'GetMetadataActivity', 'WebActivity': 'WebActivity', 'Lookup': 'LookupActivity', 'SqlServerStoredProcedure': 'SqlServerStoredProcedureActivity', 'Custom': 'CustomActivity', 'ExecuteSSISPackage': 'ExecuteSSISPackageActivity', 'HDInsightSpark': 'HDInsightSparkActivity', 'HDInsightStreaming': 'HDInsightStreamingActivity', 'HDInsightMapReduce': 'HDInsightMapReduceActivity', 'HDInsightPig': 'HDInsightPigActivity', 'HDInsightHive': 'HDInsightHiveActivity', 'Copy': 'CopyActivity'} } - def __init__(self, name, additional_properties=None, description=None, depends_on=None, linked_service_name=None, policy=None): - super(ExecutionActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on) - self.linked_service_name = linked_service_name - self.policy = policy + def __init__(self, **kwargs): + super(ExecutionActivity, self).__init__(**kwargs) + self.linked_service_name = kwargs.get('linked_service_name', None) + self.policy = kwargs.get('policy', None) self.type = 'Execution' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execution_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execution_activity_py3.py new file mode 100644 index 000000000000..2d9e3324ffd0 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execution_activity_py3.py @@ -0,0 +1,70 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .activity import Activity + + +class ExecutionActivity(Activity): + """Base class for all execution activities. + + You probably want to use the sub-classes and not this class directly. Known + sub-classes are: DatabricksNotebookActivity, DataLakeAnalyticsUSQLActivity, + AzureMLUpdateResourceActivity, AzureMLBatchExecutionActivity, + GetMetadataActivity, WebActivity, LookupActivity, + SqlServerStoredProcedureActivity, CustomActivity, + ExecuteSSISPackageActivity, HDInsightSparkActivity, + HDInsightStreamingActivity, HDInsightMapReduceActivity, + HDInsightPigActivity, HDInsightHiveActivity, CopyActivity + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param name: Required. Activity name. + :type name: str + :param description: Activity description. + :type description: str + :param depends_on: Activity depends on condition. + :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param type: Required. Constant filled by server. + :type type: str + :param linked_service_name: Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param policy: Activity policy. + :type policy: ~azure.mgmt.datafactory.models.ActivityPolicy + """ + + _validation = { + 'name': {'required': True}, + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'name': {'key': 'name', 'type': 'str'}, + 'description': {'key': 'description', 'type': 'str'}, + 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, + } + + _subtype_map = { + 'type': {'DatabricksNotebook': 'DatabricksNotebookActivity', 'DataLakeAnalyticsU-SQL': 'DataLakeAnalyticsUSQLActivity', 'AzureMLUpdateResource': 'AzureMLUpdateResourceActivity', 'AzureMLBatchExecution': 'AzureMLBatchExecutionActivity', 'GetMetadata': 'GetMetadataActivity', 'WebActivity': 'WebActivity', 'Lookup': 'LookupActivity', 'SqlServerStoredProcedure': 'SqlServerStoredProcedureActivity', 'Custom': 'CustomActivity', 'ExecuteSSISPackage': 'ExecuteSSISPackageActivity', 'HDInsightSpark': 'HDInsightSparkActivity', 'HDInsightStreaming': 'HDInsightStreamingActivity', 'HDInsightMapReduce': 'HDInsightMapReduceActivity', 'HDInsightPig': 'HDInsightPigActivity', 'HDInsightHive': 'HDInsightHiveActivity', 'Copy': 'CopyActivity'} + } + + def __init__(self, *, name: str, additional_properties=None, description: str=None, depends_on=None, linked_service_name=None, policy=None, **kwargs) -> None: + super(ExecutionActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, **kwargs) + self.linked_service_name = linked_service_name + self.policy = policy + self.type = 'Execution' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/expression.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/expression.py index 1dcebd0c48de..4b16ceca2794 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/expression.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/expression.py @@ -18,9 +18,11 @@ class Expression(Model): Variables are only populated by the server, and will be ignored when sending a request. - :ivar type: Expression type. Default value: "Expression" . + All required parameters must be populated in order to send to Azure. + + :ivar type: Required. Expression type. Default value: "Expression" . :vartype type: str - :param value: Expression value. + :param value: Required. Expression value. :type value: str """ @@ -36,6 +38,6 @@ class Expression(Model): type = "Expression" - def __init__(self, value): - super(Expression, self).__init__() - self.value = value + def __init__(self, **kwargs): + super(Expression, self).__init__(**kwargs) + self.value = kwargs.get('value', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/expression_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/expression_py3.py new file mode 100644 index 000000000000..c6ad023a57ed --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/expression_py3.py @@ -0,0 +1,43 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class Expression(Model): + """Azure Data Factory expression definition. + + Variables are only populated by the server, and will be ignored when + sending a request. + + All required parameters must be populated in order to send to Azure. + + :ivar type: Required. Expression type. Default value: "Expression" . + :vartype type: str + :param value: Required. Expression value. + :type value: str + """ + + _validation = { + 'type': {'required': True, 'constant': True}, + 'value': {'required': True}, + } + + _attribute_map = { + 'type': {'key': 'type', 'type': 'str'}, + 'value': {'key': 'value', 'type': 'str'}, + } + + type = "Expression" + + def __init__(self, *, value: str, **kwargs) -> None: + super(Expression, self).__init__(**kwargs) + self.value = value diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory.py index 0f4b48f0d8d0..3c1fdbada034 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory.py @@ -63,10 +63,10 @@ class Factory(Resource): 'version': {'key': 'properties.version', 'type': 'str'}, } - def __init__(self, location=None, tags=None, additional_properties=None, identity=None): - super(Factory, self).__init__(location=location, tags=tags) - self.additional_properties = additional_properties - self.identity = identity + def __init__(self, **kwargs): + super(Factory, self).__init__(**kwargs) + self.additional_properties = kwargs.get('additional_properties', None) + self.identity = kwargs.get('identity', None) self.provisioning_state = None self.create_time = None self.version = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_identity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_identity.py index e1c7644fee88..dad745424af3 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_identity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_identity.py @@ -18,8 +18,10 @@ class FactoryIdentity(Model): Variables are only populated by the server, and will be ignored when sending a request. - :ivar type: The identity type. Currently the only supported type is - 'SystemAssigned'. Default value: "SystemAssigned" . + All required parameters must be populated in order to send to Azure. + + :ivar type: Required. The identity type. Currently the only supported type + is 'SystemAssigned'. Default value: "SystemAssigned" . :vartype type: str :ivar principal_id: The principal id of the identity. :vartype principal_id: str @@ -41,7 +43,7 @@ class FactoryIdentity(Model): type = "SystemAssigned" - def __init__(self): - super(FactoryIdentity, self).__init__() + def __init__(self, **kwargs): + super(FactoryIdentity, self).__init__(**kwargs) self.principal_id = None self.tenant_id = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_identity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_identity_py3.py new file mode 100644 index 000000000000..567100d8c19e --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_identity_py3.py @@ -0,0 +1,49 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class FactoryIdentity(Model): + """Identity properties of the factory resource. + + Variables are only populated by the server, and will be ignored when + sending a request. + + All required parameters must be populated in order to send to Azure. + + :ivar type: Required. The identity type. Currently the only supported type + is 'SystemAssigned'. Default value: "SystemAssigned" . + :vartype type: str + :ivar principal_id: The principal id of the identity. + :vartype principal_id: str + :ivar tenant_id: The client tenant id of the identity. + :vartype tenant_id: str + """ + + _validation = { + 'type': {'required': True, 'constant': True}, + 'principal_id': {'readonly': True}, + 'tenant_id': {'readonly': True}, + } + + _attribute_map = { + 'type': {'key': 'type', 'type': 'str'}, + 'principal_id': {'key': 'principalId', 'type': 'str'}, + 'tenant_id': {'key': 'tenantId', 'type': 'str'}, + } + + type = "SystemAssigned" + + def __init__(self, **kwargs) -> None: + super(FactoryIdentity, self).__init__(**kwargs) + self.principal_id = None + self.tenant_id = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_py3.py new file mode 100644 index 000000000000..57bc1be61aae --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_py3.py @@ -0,0 +1,72 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .resource import Resource + + +class Factory(Resource): + """Factory resource type. + + Variables are only populated by the server, and will be ignored when + sending a request. + + :ivar id: The resource identifier. + :vartype id: str + :ivar name: The resource name. + :vartype name: str + :ivar type: The resource type. + :vartype type: str + :param location: The resource location. + :type location: str + :param tags: The resource tags. + :type tags: dict[str, str] + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param identity: Managed service identity of the factory. + :type identity: ~azure.mgmt.datafactory.models.FactoryIdentity + :ivar provisioning_state: Factory provisioning state, example Succeeded. + :vartype provisioning_state: str + :ivar create_time: Time the factory was created in ISO8601 format. + :vartype create_time: datetime + :ivar version: Version of the factory. + :vartype version: str + """ + + _validation = { + 'id': {'readonly': True}, + 'name': {'readonly': True}, + 'type': {'readonly': True}, + 'provisioning_state': {'readonly': True}, + 'create_time': {'readonly': True}, + 'version': {'readonly': True}, + } + + _attribute_map = { + 'id': {'key': 'id', 'type': 'str'}, + 'name': {'key': 'name', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'str'}, + 'location': {'key': 'location', 'type': 'str'}, + 'tags': {'key': 'tags', 'type': '{str}'}, + 'additional_properties': {'key': '', 'type': '{object}'}, + 'identity': {'key': 'identity', 'type': 'FactoryIdentity'}, + 'provisioning_state': {'key': 'properties.provisioningState', 'type': 'str'}, + 'create_time': {'key': 'properties.createTime', 'type': 'iso-8601'}, + 'version': {'key': 'properties.version', 'type': 'str'}, + } + + def __init__(self, *, location: str=None, tags=None, additional_properties=None, identity=None, **kwargs) -> None: + super(Factory, self).__init__(location=location, tags=tags, **kwargs) + self.additional_properties = additional_properties + self.identity = identity + self.provisioning_state = None + self.create_time = None + self.version = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_update_parameters.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_update_parameters.py index 0524027900dd..e9977fceff86 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_update_parameters.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_update_parameters.py @@ -26,7 +26,7 @@ class FactoryUpdateParameters(Model): 'identity': {'key': 'identity', 'type': 'FactoryIdentity'}, } - def __init__(self, tags=None, identity=None): - super(FactoryUpdateParameters, self).__init__() - self.tags = tags - self.identity = identity + def __init__(self, **kwargs): + super(FactoryUpdateParameters, self).__init__(**kwargs) + self.tags = kwargs.get('tags', None) + self.identity = kwargs.get('identity', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_update_parameters_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_update_parameters_py3.py new file mode 100644 index 000000000000..5bd523fedf3d --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_update_parameters_py3.py @@ -0,0 +1,32 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class FactoryUpdateParameters(Model): + """Parameters for updating a factory resource. + + :param tags: The resource tags. + :type tags: dict[str, str] + :param identity: Managed service identity of the factory. + :type identity: ~azure.mgmt.datafactory.models.FactoryIdentity + """ + + _attribute_map = { + 'tags': {'key': 'tags', 'type': '{str}'}, + 'identity': {'key': 'identity', 'type': 'FactoryIdentity'}, + } + + def __init__(self, *, tags=None, identity=None, **kwargs) -> None: + super(FactoryUpdateParameters, self).__init__(**kwargs) + self.tags = tags + self.identity = identity diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_server_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_server_linked_service.py index cc1aeac1d648..c3c90f30935d 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_server_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_server_linked_service.py @@ -15,6 +15,8 @@ class FileServerLinkedService(LinkedService): """File system linked service. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -29,10 +31,10 @@ class FileServerLinkedService(LinkedService): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param host: Host name of the server. Type: string (or Expression with - resultType string). + :param host: Required. Host name of the server. Type: string (or + Expression with resultType string). :type host: object :param user_id: User ID to logon the server. Type: string (or Expression with resultType string). @@ -63,10 +65,10 @@ class FileServerLinkedService(LinkedService): 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, host, additional_properties=None, connect_via=None, description=None, parameters=None, annotations=None, user_id=None, password=None, encrypted_credential=None): - super(FileServerLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations) - self.host = host - self.user_id = user_id - self.password = password - self.encrypted_credential = encrypted_credential + def __init__(self, **kwargs): + super(FileServerLinkedService, self).__init__(**kwargs) + self.host = kwargs.get('host', None) + self.user_id = kwargs.get('user_id', None) + self.password = kwargs.get('password', None) + self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'FileServer' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_server_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_server_linked_service_py3.py new file mode 100644 index 000000000000..705927000617 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_server_linked_service_py3.py @@ -0,0 +1,74 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class FileServerLinkedService(LinkedService): + """File system linked service. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param host: Required. Host name of the server. Type: string (or + Expression with resultType string). + :type host: object + :param user_id: User ID to logon the server. Type: string (or Expression + with resultType string). + :type user_id: object + :param password: Password to logon the server. + :type password: ~azure.mgmt.datafactory.models.SecretBase + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'host': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'host': {'key': 'typeProperties.host', 'type': 'object'}, + 'user_id': {'key': 'typeProperties.userId', 'type': 'object'}, + 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, *, host, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, user_id=None, password=None, encrypted_credential=None, **kwargs) -> None: + super(FileServerLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.host = host + self.user_id = user_id + self.password = password + self.encrypted_credential = encrypted_credential + self.type = 'FileServer' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_share_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_share_dataset.py index 8d53d8b797fd..0ca1dc0fd79a 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_share_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_share_dataset.py @@ -15,6 +15,8 @@ class FileShareDataset(Dataset): """An on-premises file system dataset. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -23,7 +25,7 @@ class FileShareDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object - :param linked_service_name: Linked service reference. + :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference :param parameters: Parameters for dataset. @@ -32,7 +34,7 @@ class FileShareDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param folder_path: The path of the on-premises file system. Type: string (or Expression with resultType string). @@ -70,11 +72,11 @@ class FileShareDataset(Dataset): 'compression': {'key': 'typeProperties.compression', 'type': 'DatasetCompression'}, } - def __init__(self, linked_service_name, additional_properties=None, description=None, structure=None, parameters=None, annotations=None, folder_path=None, file_name=None, format=None, file_filter=None, compression=None): - super(FileShareDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations) - self.folder_path = folder_path - self.file_name = file_name - self.format = format - self.file_filter = file_filter - self.compression = compression + def __init__(self, **kwargs): + super(FileShareDataset, self).__init__(**kwargs) + self.folder_path = kwargs.get('folder_path', None) + self.file_name = kwargs.get('file_name', None) + self.format = kwargs.get('format', None) + self.file_filter = kwargs.get('file_filter', None) + self.compression = kwargs.get('compression', None) self.type = 'FileShare' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_share_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_share_dataset_py3.py new file mode 100644 index 000000000000..ff76331de81a --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_share_dataset_py3.py @@ -0,0 +1,82 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class FileShareDataset(Dataset): + """An on-premises file system dataset. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Required. Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param folder_path: The path of the on-premises file system. Type: string + (or Expression with resultType string). + :type folder_path: object + :param file_name: The name of the on-premises file system. Type: string + (or Expression with resultType string). + :type file_name: object + :param format: The format of the files. + :type format: ~azure.mgmt.datafactory.models.DatasetStorageFormat + :param file_filter: Specify a filter to be used to select a subset of + files in the folderPath rather than all files. Type: string (or Expression + with resultType string). + :type file_filter: object + :param compression: The data compression method used for the file system. + :type compression: ~azure.mgmt.datafactory.models.DatasetCompression + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'folder_path': {'key': 'typeProperties.folderPath', 'type': 'object'}, + 'file_name': {'key': 'typeProperties.fileName', 'type': 'object'}, + 'format': {'key': 'typeProperties.format', 'type': 'DatasetStorageFormat'}, + 'file_filter': {'key': 'typeProperties.fileFilter', 'type': 'object'}, + 'compression': {'key': 'typeProperties.compression', 'type': 'DatasetCompression'}, + } + + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder_path=None, file_name=None, format=None, file_filter=None, compression=None, **kwargs) -> None: + super(FileShareDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + self.folder_path = folder_path + self.file_name = file_name + self.format = format + self.file_filter = file_filter + self.compression = compression + self.type = 'FileShare' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_system_sink.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_system_sink.py index 56b2a8fbb05d..9f33cea7a261 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_system_sink.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_system_sink.py @@ -15,6 +15,8 @@ class FileSystemSink(CopySink): """A copy activity file system sink. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -32,7 +34,7 @@ class FileSystemSink(CopySink): resultType string), pattern: ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type sink_retry_wait: object - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param copy_behavior: The type of copy behavior for copy sink. Possible values include: 'PreserveHierarchy', 'FlattenHierarchy', 'MergeFiles' @@ -54,7 +56,7 @@ class FileSystemSink(CopySink): 'copy_behavior': {'key': 'copyBehavior', 'type': 'str'}, } - def __init__(self, additional_properties=None, write_batch_size=None, write_batch_timeout=None, sink_retry_count=None, sink_retry_wait=None, copy_behavior=None): - super(FileSystemSink, self).__init__(additional_properties=additional_properties, write_batch_size=write_batch_size, write_batch_timeout=write_batch_timeout, sink_retry_count=sink_retry_count, sink_retry_wait=sink_retry_wait) - self.copy_behavior = copy_behavior + def __init__(self, **kwargs): + super(FileSystemSink, self).__init__(**kwargs) + self.copy_behavior = kwargs.get('copy_behavior', None) self.type = 'FileSystemSink' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_system_sink_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_system_sink_py3.py new file mode 100644 index 000000000000..59bd6ec12f41 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_system_sink_py3.py @@ -0,0 +1,62 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_sink import CopySink + + +class FileSystemSink(CopySink): + """A copy activity file system sink. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param write_batch_size: Write batch size. Type: integer (or Expression + with resultType integer), minimum: 0. + :type write_batch_size: object + :param write_batch_timeout: Write batch timeout. Type: string (or + Expression with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type write_batch_timeout: object + :param sink_retry_count: Sink retry count. Type: integer (or Expression + with resultType integer). + :type sink_retry_count: object + :param sink_retry_wait: Sink retry wait. Type: string (or Expression with + resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type sink_retry_wait: object + :param type: Required. Constant filled by server. + :type type: str + :param copy_behavior: The type of copy behavior for copy sink. Possible + values include: 'PreserveHierarchy', 'FlattenHierarchy', 'MergeFiles' + :type copy_behavior: str or + ~azure.mgmt.datafactory.models.CopyBehaviorType + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, + 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, + 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, + 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'copy_behavior': {'key': 'copyBehavior', 'type': 'str'}, + } + + def __init__(self, *, additional_properties=None, write_batch_size=None, write_batch_timeout=None, sink_retry_count=None, sink_retry_wait=None, copy_behavior=None, **kwargs) -> None: + super(FileSystemSink, self).__init__(additional_properties=additional_properties, write_batch_size=write_batch_size, write_batch_timeout=write_batch_timeout, sink_retry_count=sink_retry_count, sink_retry_wait=sink_retry_wait, **kwargs) + self.copy_behavior = copy_behavior + self.type = 'FileSystemSink' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_system_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_system_source.py index b52ab6d03285..1bbf97f1b31d 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_system_source.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_system_source.py @@ -15,6 +15,8 @@ class FileSystemSource(CopySource): """A copy activity file system source. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -25,7 +27,7 @@ class FileSystemSource(CopySource): with resultType string), pattern: ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type source_retry_wait: object - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param recursive: If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType @@ -45,7 +47,7 @@ class FileSystemSource(CopySource): 'recursive': {'key': 'recursive', 'type': 'object'}, } - def __init__(self, additional_properties=None, source_retry_count=None, source_retry_wait=None, recursive=None): - super(FileSystemSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait) - self.recursive = recursive + def __init__(self, **kwargs): + super(FileSystemSource, self).__init__(**kwargs) + self.recursive = kwargs.get('recursive', None) self.type = 'FileSystemSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_system_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_system_source_py3.py new file mode 100644 index 000000000000..bf6eeb2e7fd5 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_system_source_py3.py @@ -0,0 +1,53 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class FileSystemSource(CopySource): + """A copy activity file system source. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Required. Constant filled by server. + :type type: str + :param recursive: If true, files under the folder path will be read + recursively. Default is true. Type: boolean (or Expression with resultType + boolean). + :type recursive: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'recursive': {'key': 'recursive', 'type': 'object'}, + } + + def __init__(self, *, additional_properties=None, source_retry_count=None, source_retry_wait=None, recursive=None, **kwargs) -> None: + super(FileSystemSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, **kwargs) + self.recursive = recursive + self.type = 'FileSystemSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/filter_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/filter_activity.py index 74f7300b0758..409eff874d6d 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/filter_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/filter_activity.py @@ -15,20 +15,22 @@ class FilterActivity(ControlActivity): """Filter and return results from input array based on the conditions. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] - :param name: Activity name. + :param name: Required. Activity name. :type name: str :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param items: Input array on which filter should be applied. + :param items: Required. Input array on which filter should be applied. :type items: ~azure.mgmt.datafactory.models.Expression - :param condition: Condition to be used for filtering the input. + :param condition: Required. Condition to be used for filtering the input. :type condition: ~azure.mgmt.datafactory.models.Expression """ @@ -49,8 +51,8 @@ class FilterActivity(ControlActivity): 'condition': {'key': 'typeProperties.condition', 'type': 'Expression'}, } - def __init__(self, name, items, condition, additional_properties=None, description=None, depends_on=None): - super(FilterActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on) - self.items = items - self.condition = condition + def __init__(self, **kwargs): + super(FilterActivity, self).__init__(**kwargs) + self.items = kwargs.get('items', None) + self.condition = kwargs.get('condition', None) self.type = 'Filter' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/filter_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/filter_activity_py3.py new file mode 100644 index 000000000000..87b0d1ccdeb7 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/filter_activity_py3.py @@ -0,0 +1,58 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .control_activity import ControlActivity + + +class FilterActivity(ControlActivity): + """Filter and return results from input array based on the conditions. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param name: Required. Activity name. + :type name: str + :param description: Activity description. + :type description: str + :param depends_on: Activity depends on condition. + :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param type: Required. Constant filled by server. + :type type: str + :param items: Required. Input array on which filter should be applied. + :type items: ~azure.mgmt.datafactory.models.Expression + :param condition: Required. Condition to be used for filtering the input. + :type condition: ~azure.mgmt.datafactory.models.Expression + """ + + _validation = { + 'name': {'required': True}, + 'type': {'required': True}, + 'items': {'required': True}, + 'condition': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'name': {'key': 'name', 'type': 'str'}, + 'description': {'key': 'description', 'type': 'str'}, + 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'items': {'key': 'typeProperties.items', 'type': 'Expression'}, + 'condition': {'key': 'typeProperties.condition', 'type': 'Expression'}, + } + + def __init__(self, *, name: str, items, condition, additional_properties=None, description: str=None, depends_on=None, **kwargs) -> None: + super(FilterActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, **kwargs) + self.items = items + self.condition = condition + self.type = 'Filter' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/for_each_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/for_each_activity.py index f26f45af5f4a..f170cfa054ca 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/for_each_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/for_each_activity.py @@ -16,16 +16,18 @@ class ForEachActivity(ControlActivity): """This activity is used for iterating over a collection and execute given activities. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] - :param name: Activity name. + :param name: Required. Activity name. :type name: str :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param is_sequential: Should the loop be executed in sequence or in parallel (max 50) @@ -33,9 +35,9 @@ class ForEachActivity(ControlActivity): :param batch_count: Batch count to be used for controlling the number of parallel execution (when isSequential is set to false). :type batch_count: int - :param items: Collection to iterate. + :param items: Required. Collection to iterate. :type items: ~azure.mgmt.datafactory.models.Expression - :param activities: List of activities to execute . + :param activities: Required. List of activities to execute . :type activities: list[~azure.mgmt.datafactory.models.Activity] """ @@ -59,10 +61,10 @@ class ForEachActivity(ControlActivity): 'activities': {'key': 'typeProperties.activities', 'type': '[Activity]'}, } - def __init__(self, name, items, activities, additional_properties=None, description=None, depends_on=None, is_sequential=None, batch_count=None): - super(ForEachActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on) - self.is_sequential = is_sequential - self.batch_count = batch_count - self.items = items - self.activities = activities + def __init__(self, **kwargs): + super(ForEachActivity, self).__init__(**kwargs) + self.is_sequential = kwargs.get('is_sequential', None) + self.batch_count = kwargs.get('batch_count', None) + self.items = kwargs.get('items', None) + self.activities = kwargs.get('activities', None) self.type = 'ForEach' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/for_each_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/for_each_activity_py3.py new file mode 100644 index 000000000000..88d75bc00f3b --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/for_each_activity_py3.py @@ -0,0 +1,70 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .control_activity import ControlActivity + + +class ForEachActivity(ControlActivity): + """This activity is used for iterating over a collection and execute given + activities. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param name: Required. Activity name. + :type name: str + :param description: Activity description. + :type description: str + :param depends_on: Activity depends on condition. + :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param type: Required. Constant filled by server. + :type type: str + :param is_sequential: Should the loop be executed in sequence or in + parallel (max 50) + :type is_sequential: bool + :param batch_count: Batch count to be used for controlling the number of + parallel execution (when isSequential is set to false). + :type batch_count: int + :param items: Required. Collection to iterate. + :type items: ~azure.mgmt.datafactory.models.Expression + :param activities: Required. List of activities to execute . + :type activities: list[~azure.mgmt.datafactory.models.Activity] + """ + + _validation = { + 'name': {'required': True}, + 'type': {'required': True}, + 'batch_count': {'maximum': 50}, + 'items': {'required': True}, + 'activities': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'name': {'key': 'name', 'type': 'str'}, + 'description': {'key': 'description', 'type': 'str'}, + 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'is_sequential': {'key': 'typeProperties.isSequential', 'type': 'bool'}, + 'batch_count': {'key': 'typeProperties.batchCount', 'type': 'int'}, + 'items': {'key': 'typeProperties.items', 'type': 'Expression'}, + 'activities': {'key': 'typeProperties.activities', 'type': '[Activity]'}, + } + + def __init__(self, *, name: str, items, activities, additional_properties=None, description: str=None, depends_on=None, is_sequential: bool=None, batch_count: int=None, **kwargs) -> None: + super(ForEachActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, **kwargs) + self.is_sequential = is_sequential + self.batch_count = batch_count + self.items = items + self.activities = activities + self.type = 'ForEach' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/ftp_server_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/ftp_server_linked_service.py index 010d528202d8..03a09f89c13e 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/ftp_server_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/ftp_server_linked_service.py @@ -15,6 +15,8 @@ class FtpServerLinkedService(LinkedService): """A FTP server Linked Service. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -29,10 +31,10 @@ class FtpServerLinkedService(LinkedService): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param host: Host name of the FTP server. Type: string (or Expression with - resultType string). + :param host: Required. Host name of the FTP server. Type: string (or + Expression with resultType string). :type host: object :param port: The TCP port number that the FTP server uses to listen for client connections. Default value is 21. Type: integer (or Expression with @@ -83,14 +85,14 @@ class FtpServerLinkedService(LinkedService): 'enable_server_certificate_validation': {'key': 'typeProperties.enableServerCertificateValidation', 'type': 'object'}, } - def __init__(self, host, additional_properties=None, connect_via=None, description=None, parameters=None, annotations=None, port=None, authentication_type=None, user_name=None, password=None, encrypted_credential=None, enable_ssl=None, enable_server_certificate_validation=None): - super(FtpServerLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations) - self.host = host - self.port = port - self.authentication_type = authentication_type - self.user_name = user_name - self.password = password - self.encrypted_credential = encrypted_credential - self.enable_ssl = enable_ssl - self.enable_server_certificate_validation = enable_server_certificate_validation + def __init__(self, **kwargs): + super(FtpServerLinkedService, self).__init__(**kwargs) + self.host = kwargs.get('host', None) + self.port = kwargs.get('port', None) + self.authentication_type = kwargs.get('authentication_type', None) + self.user_name = kwargs.get('user_name', None) + self.password = kwargs.get('password', None) + self.encrypted_credential = kwargs.get('encrypted_credential', None) + self.enable_ssl = kwargs.get('enable_ssl', None) + self.enable_server_certificate_validation = kwargs.get('enable_server_certificate_validation', None) self.type = 'FtpServer' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/ftp_server_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/ftp_server_linked_service_py3.py new file mode 100644 index 000000000000..b80223faf92f --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/ftp_server_linked_service_py3.py @@ -0,0 +1,98 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class FtpServerLinkedService(LinkedService): + """A FTP server Linked Service. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param host: Required. Host name of the FTP server. Type: string (or + Expression with resultType string). + :type host: object + :param port: The TCP port number that the FTP server uses to listen for + client connections. Default value is 21. Type: integer (or Expression with + resultType integer), minimum: 0. + :type port: object + :param authentication_type: The authentication type to be used to connect + to the FTP server. Possible values include: 'Basic', 'Anonymous' + :type authentication_type: str or + ~azure.mgmt.datafactory.models.FtpAuthenticationType + :param user_name: Username to logon the FTP server. Type: string (or + Expression with resultType string). + :type user_name: object + :param password: Password to logon the FTP server. + :type password: ~azure.mgmt.datafactory.models.SecretBase + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + :param enable_ssl: If true, connect to the FTP server over SSL/TLS + channel. Default value is true. Type: boolean (or Expression with + resultType boolean). + :type enable_ssl: object + :param enable_server_certificate_validation: If true, validate the FTP + server SSL certificate when connect over SSL/TLS channel. Default value is + true. Type: boolean (or Expression with resultType boolean). + :type enable_server_certificate_validation: object + """ + + _validation = { + 'type': {'required': True}, + 'host': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'host': {'key': 'typeProperties.host', 'type': 'object'}, + 'port': {'key': 'typeProperties.port', 'type': 'object'}, + 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, + 'user_name': {'key': 'typeProperties.userName', 'type': 'object'}, + 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + 'enable_ssl': {'key': 'typeProperties.enableSsl', 'type': 'object'}, + 'enable_server_certificate_validation': {'key': 'typeProperties.enableServerCertificateValidation', 'type': 'object'}, + } + + def __init__(self, *, host, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, port=None, authentication_type=None, user_name=None, password=None, encrypted_credential=None, enable_ssl=None, enable_server_certificate_validation=None, **kwargs) -> None: + super(FtpServerLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.host = host + self.port = port + self.authentication_type = authentication_type + self.user_name = user_name + self.password = password + self.encrypted_credential = encrypted_credential + self.enable_ssl = enable_ssl + self.enable_server_certificate_validation = enable_server_certificate_validation + self.type = 'FtpServer' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/get_metadata_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/get_metadata_activity.py index 4d48de728e09..c1bac6e451f1 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/get_metadata_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/get_metadata_activity.py @@ -15,23 +15,25 @@ class GetMetadataActivity(ExecutionActivity): """Activity to get metadata of dataset. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] - :param name: Activity name. + :param name: Required. Activity name. :type name: str :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference :param policy: Activity policy. :type policy: ~azure.mgmt.datafactory.models.ActivityPolicy - :param dataset: GetMetadata activity dataset reference. + :param dataset: Required. GetMetadata activity dataset reference. :type dataset: ~azure.mgmt.datafactory.models.DatasetReference :param field_list: Fields of metadata to get from dataset. :type field_list: list[object] @@ -55,8 +57,8 @@ class GetMetadataActivity(ExecutionActivity): 'field_list': {'key': 'typeProperties.fieldList', 'type': '[object]'}, } - def __init__(self, name, dataset, additional_properties=None, description=None, depends_on=None, linked_service_name=None, policy=None, field_list=None): - super(GetMetadataActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, linked_service_name=linked_service_name, policy=policy) - self.dataset = dataset - self.field_list = field_list + def __init__(self, **kwargs): + super(GetMetadataActivity, self).__init__(**kwargs) + self.dataset = kwargs.get('dataset', None) + self.field_list = kwargs.get('field_list', None) self.type = 'GetMetadata' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/get_metadata_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/get_metadata_activity_py3.py new file mode 100644 index 000000000000..2db128a4dc7f --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/get_metadata_activity_py3.py @@ -0,0 +1,64 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .execution_activity import ExecutionActivity + + +class GetMetadataActivity(ExecutionActivity): + """Activity to get metadata of dataset. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param name: Required. Activity name. + :type name: str + :param description: Activity description. + :type description: str + :param depends_on: Activity depends on condition. + :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param type: Required. Constant filled by server. + :type type: str + :param linked_service_name: Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param policy: Activity policy. + :type policy: ~azure.mgmt.datafactory.models.ActivityPolicy + :param dataset: Required. GetMetadata activity dataset reference. + :type dataset: ~azure.mgmt.datafactory.models.DatasetReference + :param field_list: Fields of metadata to get from dataset. + :type field_list: list[object] + """ + + _validation = { + 'name': {'required': True}, + 'type': {'required': True}, + 'dataset': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'name': {'key': 'name', 'type': 'str'}, + 'description': {'key': 'description', 'type': 'str'}, + 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, + 'dataset': {'key': 'typeProperties.dataset', 'type': 'DatasetReference'}, + 'field_list': {'key': 'typeProperties.fieldList', 'type': '[object]'}, + } + + def __init__(self, *, name: str, dataset, additional_properties=None, description: str=None, depends_on=None, linked_service_name=None, policy=None, field_list=None, **kwargs) -> None: + super(GetMetadataActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, linked_service_name=linked_service_name, policy=policy, **kwargs) + self.dataset = dataset + self.field_list = field_list + self.type = 'GetMetadata' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_big_query_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_big_query_linked_service.py index 0fbef57855ef..c9fa8239b452 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_big_query_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_big_query_linked_service.py @@ -15,6 +15,8 @@ class GoogleBigQueryLinkedService(LinkedService): """Google BigQuery service linked service. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -29,9 +31,9 @@ class GoogleBigQueryLinkedService(LinkedService): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param project: The default BigQuery project to query against. + :param project: Required. The default BigQuery project to query against. :type project: object :param additional_projects: A comma-separated list of public BigQuery projects to access. @@ -41,9 +43,10 @@ class GoogleBigQueryLinkedService(LinkedService): that combine BigQuery data with data from Google Drive. The default value is false. :type request_google_drive_scope: object - :param authentication_type: The OAuth 2.0 authentication mechanism used - for authentication. ServiceAuthentication can only be used on self-hosted - IR. Possible values include: 'ServiceAuthentication', 'UserAuthentication' + :param authentication_type: Required. The OAuth 2.0 authentication + mechanism used for authentication. ServiceAuthentication can only be used + on self-hosted IR. Possible values include: 'ServiceAuthentication', + 'UserAuthentication' :type authentication_type: str or ~azure.mgmt.datafactory.models.GoogleBigQueryAuthenticationType :param refresh_token: The refresh token obtained from Google for @@ -104,18 +107,18 @@ class GoogleBigQueryLinkedService(LinkedService): 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, project, authentication_type, additional_properties=None, connect_via=None, description=None, parameters=None, annotations=None, additional_projects=None, request_google_drive_scope=None, refresh_token=None, client_id=None, client_secret=None, email=None, key_file_path=None, trusted_cert_path=None, use_system_trust_store=None, encrypted_credential=None): - super(GoogleBigQueryLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations) - self.project = project - self.additional_projects = additional_projects - self.request_google_drive_scope = request_google_drive_scope - self.authentication_type = authentication_type - self.refresh_token = refresh_token - self.client_id = client_id - self.client_secret = client_secret - self.email = email - self.key_file_path = key_file_path - self.trusted_cert_path = trusted_cert_path - self.use_system_trust_store = use_system_trust_store - self.encrypted_credential = encrypted_credential + def __init__(self, **kwargs): + super(GoogleBigQueryLinkedService, self).__init__(**kwargs) + self.project = kwargs.get('project', None) + self.additional_projects = kwargs.get('additional_projects', None) + self.request_google_drive_scope = kwargs.get('request_google_drive_scope', None) + self.authentication_type = kwargs.get('authentication_type', None) + self.refresh_token = kwargs.get('refresh_token', None) + self.client_id = kwargs.get('client_id', None) + self.client_secret = kwargs.get('client_secret', None) + self.email = kwargs.get('email', None) + self.key_file_path = kwargs.get('key_file_path', None) + self.trusted_cert_path = kwargs.get('trusted_cert_path', None) + self.use_system_trust_store = kwargs.get('use_system_trust_store', None) + self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'GoogleBigQuery' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_big_query_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_big_query_linked_service_py3.py new file mode 100644 index 000000000000..549dcb7abf1b --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_big_query_linked_service_py3.py @@ -0,0 +1,124 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class GoogleBigQueryLinkedService(LinkedService): + """Google BigQuery service linked service. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param project: Required. The default BigQuery project to query against. + :type project: object + :param additional_projects: A comma-separated list of public BigQuery + projects to access. + :type additional_projects: object + :param request_google_drive_scope: Whether to request access to Google + Drive. Allowing Google Drive access enables support for federated tables + that combine BigQuery data with data from Google Drive. The default value + is false. + :type request_google_drive_scope: object + :param authentication_type: Required. The OAuth 2.0 authentication + mechanism used for authentication. ServiceAuthentication can only be used + on self-hosted IR. Possible values include: 'ServiceAuthentication', + 'UserAuthentication' + :type authentication_type: str or + ~azure.mgmt.datafactory.models.GoogleBigQueryAuthenticationType + :param refresh_token: The refresh token obtained from Google for + authorizing access to BigQuery for UserAuthentication. + :type refresh_token: ~azure.mgmt.datafactory.models.SecretBase + :param client_id: The client id of the google application used to acquire + the refresh token. + :type client_id: ~azure.mgmt.datafactory.models.SecretBase + :param client_secret: The client secret of the google application used to + acquire the refresh token. + :type client_secret: ~azure.mgmt.datafactory.models.SecretBase + :param email: The service account email ID that is used for + ServiceAuthentication and can only be used on self-hosted IR. + :type email: object + :param key_file_path: The full path to the .p12 key file that is used to + authenticate the service account email address and can only be used on + self-hosted IR. + :type key_file_path: object + :param trusted_cert_path: The full path of the .pem file containing + trusted CA certificates for verifying the server when connecting over SSL. + This property can only be set when using SSL on self-hosted IR. The + default value is the cacerts.pem file installed with the IR. + :type trusted_cert_path: object + :param use_system_trust_store: Specifies whether to use a CA certificate + from the system trust store or from a specified PEM file. The default + value is false. + :type use_system_trust_store: object + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'project': {'required': True}, + 'authentication_type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'project': {'key': 'typeProperties.project', 'type': 'object'}, + 'additional_projects': {'key': 'typeProperties.additionalProjects', 'type': 'object'}, + 'request_google_drive_scope': {'key': 'typeProperties.requestGoogleDriveScope', 'type': 'object'}, + 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, + 'refresh_token': {'key': 'typeProperties.refreshToken', 'type': 'SecretBase'}, + 'client_id': {'key': 'typeProperties.clientId', 'type': 'SecretBase'}, + 'client_secret': {'key': 'typeProperties.clientSecret', 'type': 'SecretBase'}, + 'email': {'key': 'typeProperties.email', 'type': 'object'}, + 'key_file_path': {'key': 'typeProperties.keyFilePath', 'type': 'object'}, + 'trusted_cert_path': {'key': 'typeProperties.trustedCertPath', 'type': 'object'}, + 'use_system_trust_store': {'key': 'typeProperties.useSystemTrustStore', 'type': 'object'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, *, project, authentication_type, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, additional_projects=None, request_google_drive_scope=None, refresh_token=None, client_id=None, client_secret=None, email=None, key_file_path=None, trusted_cert_path=None, use_system_trust_store=None, encrypted_credential=None, **kwargs) -> None: + super(GoogleBigQueryLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.project = project + self.additional_projects = additional_projects + self.request_google_drive_scope = request_google_drive_scope + self.authentication_type = authentication_type + self.refresh_token = refresh_token + self.client_id = client_id + self.client_secret = client_secret + self.email = email + self.key_file_path = key_file_path + self.trusted_cert_path = trusted_cert_path + self.use_system_trust_store = use_system_trust_store + self.encrypted_credential = encrypted_credential + self.type = 'GoogleBigQuery' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_big_query_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_big_query_object_dataset.py index ce49eb0321c1..e089870c4d1a 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_big_query_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_big_query_object_dataset.py @@ -15,6 +15,8 @@ class GoogleBigQueryObjectDataset(Dataset): """Google BigQuery service dataset. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -23,7 +25,7 @@ class GoogleBigQueryObjectDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object - :param linked_service_name: Linked service reference. + :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference :param parameters: Parameters for dataset. @@ -32,7 +34,7 @@ class GoogleBigQueryObjectDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str """ @@ -41,6 +43,16 @@ class GoogleBigQueryObjectDataset(Dataset): 'type': {'required': True}, } - def __init__(self, linked_service_name, additional_properties=None, description=None, structure=None, parameters=None, annotations=None): - super(GoogleBigQueryObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations) + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, **kwargs): + super(GoogleBigQueryObjectDataset, self).__init__(**kwargs) self.type = 'GoogleBigQueryObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_big_query_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_big_query_object_dataset_py3.py new file mode 100644 index 000000000000..1bd794bc719c --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_big_query_object_dataset_py3.py @@ -0,0 +1,58 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class GoogleBigQueryObjectDataset(Dataset): + """Google BigQuery service dataset. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Required. Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: + super(GoogleBigQueryObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + self.type = 'GoogleBigQueryObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_big_query_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_big_query_source.py index 938df771fdb2..c0598d88a6ed 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_big_query_source.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_big_query_source.py @@ -15,6 +15,8 @@ class GoogleBigQuerySource(CopySource): """A copy activity Google BigQuery service source. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -25,7 +27,7 @@ class GoogleBigQuerySource(CopySource): with resultType string), pattern: ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type source_retry_wait: object - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). @@ -44,7 +46,7 @@ class GoogleBigQuerySource(CopySource): 'query': {'key': 'query', 'type': 'object'}, } - def __init__(self, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None): - super(GoogleBigQuerySource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait) - self.query = query + def __init__(self, **kwargs): + super(GoogleBigQuerySource, self).__init__(**kwargs) + self.query = kwargs.get('query', None) self.type = 'GoogleBigQuerySource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_big_query_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_big_query_source_py3.py new file mode 100644 index 000000000000..7bd1a5256c19 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_big_query_source_py3.py @@ -0,0 +1,52 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class GoogleBigQuerySource(CopySource): + """A copy activity Google BigQuery service source. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Required. Constant filled by server. + :type type: str + :param query: A query to retrieve data from source. Type: string (or + Expression with resultType string). + :type query: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'query': {'key': 'query', 'type': 'object'}, + } + + def __init__(self, *, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None, **kwargs) -> None: + super(GoogleBigQuerySource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, **kwargs) + self.query = query + self.type = 'GoogleBigQuerySource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_linked_service.py index a24878a71d40..cedb1a371923 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_linked_service.py @@ -15,6 +15,8 @@ class GreenplumLinkedService(LinkedService): """Greenplum Database linked service. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -29,7 +31,7 @@ class GreenplumLinkedService(LinkedService): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param connection_string: An ODBC connection string. :type connection_string: ~azure.mgmt.datafactory.models.SecretBase @@ -54,8 +56,8 @@ class GreenplumLinkedService(LinkedService): 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, additional_properties=None, connect_via=None, description=None, parameters=None, annotations=None, connection_string=None, encrypted_credential=None): - super(GreenplumLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations) - self.connection_string = connection_string - self.encrypted_credential = encrypted_credential + def __init__(self, **kwargs): + super(GreenplumLinkedService, self).__init__(**kwargs) + self.connection_string = kwargs.get('connection_string', None) + self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'Greenplum' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_linked_service_py3.py new file mode 100644 index 000000000000..bf0d5a8f9150 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_linked_service_py3.py @@ -0,0 +1,63 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class GreenplumLinkedService(LinkedService): + """Greenplum Database linked service. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param connection_string: An ODBC connection string. + :type connection_string: ~azure.mgmt.datafactory.models.SecretBase + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'SecretBase'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, *, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, connection_string=None, encrypted_credential=None, **kwargs) -> None: + super(GreenplumLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.connection_string = connection_string + self.encrypted_credential = encrypted_credential + self.type = 'Greenplum' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_source.py index 25a669e2f402..a463ff2c3482 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_source.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_source.py @@ -15,6 +15,8 @@ class GreenplumSource(CopySource): """A copy activity Greenplum Database source. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -25,7 +27,7 @@ class GreenplumSource(CopySource): with resultType string), pattern: ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type source_retry_wait: object - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). @@ -44,7 +46,7 @@ class GreenplumSource(CopySource): 'query': {'key': 'query', 'type': 'object'}, } - def __init__(self, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None): - super(GreenplumSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait) - self.query = query + def __init__(self, **kwargs): + super(GreenplumSource, self).__init__(**kwargs) + self.query = kwargs.get('query', None) self.type = 'GreenplumSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_source_py3.py new file mode 100644 index 000000000000..9adb06fb16ce --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_source_py3.py @@ -0,0 +1,52 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class GreenplumSource(CopySource): + """A copy activity Greenplum Database source. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Required. Constant filled by server. + :type type: str + :param query: A query to retrieve data from source. Type: string (or + Expression with resultType string). + :type query: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'query': {'key': 'query', 'type': 'object'}, + } + + def __init__(self, *, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None, **kwargs) -> None: + super(GreenplumSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, **kwargs) + self.query = query + self.type = 'GreenplumSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_table_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_table_dataset.py index 27bcb145ea28..7c031f285b00 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_table_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_table_dataset.py @@ -15,6 +15,8 @@ class GreenplumTableDataset(Dataset): """Greenplum Database dataset. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -23,7 +25,7 @@ class GreenplumTableDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object - :param linked_service_name: Linked service reference. + :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference :param parameters: Parameters for dataset. @@ -32,7 +34,7 @@ class GreenplumTableDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str """ @@ -41,6 +43,16 @@ class GreenplumTableDataset(Dataset): 'type': {'required': True}, } - def __init__(self, linked_service_name, additional_properties=None, description=None, structure=None, parameters=None, annotations=None): - super(GreenplumTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations) + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, **kwargs): + super(GreenplumTableDataset, self).__init__(**kwargs) self.type = 'GreenplumTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_table_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_table_dataset_py3.py new file mode 100644 index 000000000000..e68a07b2d05b --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_table_dataset_py3.py @@ -0,0 +1,58 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class GreenplumTableDataset(Dataset): + """Greenplum Database dataset. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Required. Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: + super(GreenplumTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + self.type = 'GreenplumTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hbase_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hbase_linked_service.py index 27060fa9e91a..4d7f3bf5ccb6 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hbase_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hbase_linked_service.py @@ -15,6 +15,8 @@ class HBaseLinkedService(LinkedService): """HBase server linked service. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -29,10 +31,10 @@ class HBaseLinkedService(LinkedService): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param host: The IP address or host name of the HBase server. (i.e. - 192.168.222.160) + :param host: Required. The IP address or host name of the HBase server. + (i.e. 192.168.222.160) :type host: object :param port: The TCP port that the HBase instance uses to listen for client connections. The default value is 9090. @@ -40,8 +42,9 @@ class HBaseLinkedService(LinkedService): :param http_path: The partial URL corresponding to the HBase server. (i.e. /gateway/sandbox/hbase/version) :type http_path: object - :param authentication_type: The authentication mechanism to use to connect - to the HBase server. Possible values include: 'Anonymous', 'Basic' + :param authentication_type: Required. The authentication mechanism to use + to connect to the HBase server. Possible values include: 'Anonymous', + 'Basic' :type authentication_type: str or ~azure.mgmt.datafactory.models.HBaseAuthenticationType :param username: The user name used to connect to the HBase instance. @@ -95,17 +98,17 @@ class HBaseLinkedService(LinkedService): 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, host, authentication_type, additional_properties=None, connect_via=None, description=None, parameters=None, annotations=None, port=None, http_path=None, username=None, password=None, enable_ssl=None, trusted_cert_path=None, allow_host_name_cn_mismatch=None, allow_self_signed_server_cert=None, encrypted_credential=None): - super(HBaseLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations) - self.host = host - self.port = port - self.http_path = http_path - self.authentication_type = authentication_type - self.username = username - self.password = password - self.enable_ssl = enable_ssl - self.trusted_cert_path = trusted_cert_path - self.allow_host_name_cn_mismatch = allow_host_name_cn_mismatch - self.allow_self_signed_server_cert = allow_self_signed_server_cert - self.encrypted_credential = encrypted_credential + def __init__(self, **kwargs): + super(HBaseLinkedService, self).__init__(**kwargs) + self.host = kwargs.get('host', None) + self.port = kwargs.get('port', None) + self.http_path = kwargs.get('http_path', None) + self.authentication_type = kwargs.get('authentication_type', None) + self.username = kwargs.get('username', None) + self.password = kwargs.get('password', None) + self.enable_ssl = kwargs.get('enable_ssl', None) + self.trusted_cert_path = kwargs.get('trusted_cert_path', None) + self.allow_host_name_cn_mismatch = kwargs.get('allow_host_name_cn_mismatch', None) + self.allow_self_signed_server_cert = kwargs.get('allow_self_signed_server_cert', None) + self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'HBase' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hbase_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hbase_linked_service_py3.py new file mode 100644 index 000000000000..e8bc89f401b7 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hbase_linked_service_py3.py @@ -0,0 +1,114 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class HBaseLinkedService(LinkedService): + """HBase server linked service. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param host: Required. The IP address or host name of the HBase server. + (i.e. 192.168.222.160) + :type host: object + :param port: The TCP port that the HBase instance uses to listen for + client connections. The default value is 9090. + :type port: object + :param http_path: The partial URL corresponding to the HBase server. (i.e. + /gateway/sandbox/hbase/version) + :type http_path: object + :param authentication_type: Required. The authentication mechanism to use + to connect to the HBase server. Possible values include: 'Anonymous', + 'Basic' + :type authentication_type: str or + ~azure.mgmt.datafactory.models.HBaseAuthenticationType + :param username: The user name used to connect to the HBase instance. + :type username: object + :param password: The password corresponding to the user name. + :type password: ~azure.mgmt.datafactory.models.SecretBase + :param enable_ssl: Specifies whether the connections to the server are + encrypted using SSL. The default value is false. + :type enable_ssl: object + :param trusted_cert_path: The full path of the .pem file containing + trusted CA certificates for verifying the server when connecting over SSL. + This property can only be set when using SSL on self-hosted IR. The + default value is the cacerts.pem file installed with the IR. + :type trusted_cert_path: object + :param allow_host_name_cn_mismatch: Specifies whether to require a + CA-issued SSL certificate name to match the host name of the server when + connecting over SSL. The default value is false. + :type allow_host_name_cn_mismatch: object + :param allow_self_signed_server_cert: Specifies whether to allow + self-signed certificates from the server. The default value is false. + :type allow_self_signed_server_cert: object + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'host': {'required': True}, + 'authentication_type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'host': {'key': 'typeProperties.host', 'type': 'object'}, + 'port': {'key': 'typeProperties.port', 'type': 'object'}, + 'http_path': {'key': 'typeProperties.httpPath', 'type': 'object'}, + 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, + 'username': {'key': 'typeProperties.username', 'type': 'object'}, + 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, + 'enable_ssl': {'key': 'typeProperties.enableSsl', 'type': 'object'}, + 'trusted_cert_path': {'key': 'typeProperties.trustedCertPath', 'type': 'object'}, + 'allow_host_name_cn_mismatch': {'key': 'typeProperties.allowHostNameCNMismatch', 'type': 'object'}, + 'allow_self_signed_server_cert': {'key': 'typeProperties.allowSelfSignedServerCert', 'type': 'object'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, *, host, authentication_type, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, port=None, http_path=None, username=None, password=None, enable_ssl=None, trusted_cert_path=None, allow_host_name_cn_mismatch=None, allow_self_signed_server_cert=None, encrypted_credential=None, **kwargs) -> None: + super(HBaseLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.host = host + self.port = port + self.http_path = http_path + self.authentication_type = authentication_type + self.username = username + self.password = password + self.enable_ssl = enable_ssl + self.trusted_cert_path = trusted_cert_path + self.allow_host_name_cn_mismatch = allow_host_name_cn_mismatch + self.allow_self_signed_server_cert = allow_self_signed_server_cert + self.encrypted_credential = encrypted_credential + self.type = 'HBase' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hbase_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hbase_object_dataset.py index d5ed5ed80ab7..6e78856a1468 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hbase_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hbase_object_dataset.py @@ -15,6 +15,8 @@ class HBaseObjectDataset(Dataset): """HBase server dataset. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -23,7 +25,7 @@ class HBaseObjectDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object - :param linked_service_name: Linked service reference. + :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference :param parameters: Parameters for dataset. @@ -32,7 +34,7 @@ class HBaseObjectDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str """ @@ -41,6 +43,16 @@ class HBaseObjectDataset(Dataset): 'type': {'required': True}, } - def __init__(self, linked_service_name, additional_properties=None, description=None, structure=None, parameters=None, annotations=None): - super(HBaseObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations) + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, **kwargs): + super(HBaseObjectDataset, self).__init__(**kwargs) self.type = 'HBaseObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hbase_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hbase_object_dataset_py3.py new file mode 100644 index 000000000000..30c2ed7de587 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hbase_object_dataset_py3.py @@ -0,0 +1,58 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class HBaseObjectDataset(Dataset): + """HBase server dataset. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Required. Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: + super(HBaseObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + self.type = 'HBaseObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hbase_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hbase_source.py index 09b14046b414..cc2c4fd1a843 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hbase_source.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hbase_source.py @@ -15,6 +15,8 @@ class HBaseSource(CopySource): """A copy activity HBase server source. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -25,7 +27,7 @@ class HBaseSource(CopySource): with resultType string), pattern: ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type source_retry_wait: object - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). @@ -44,7 +46,7 @@ class HBaseSource(CopySource): 'query': {'key': 'query', 'type': 'object'}, } - def __init__(self, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None): - super(HBaseSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait) - self.query = query + def __init__(self, **kwargs): + super(HBaseSource, self).__init__(**kwargs) + self.query = kwargs.get('query', None) self.type = 'HBaseSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hbase_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hbase_source_py3.py new file mode 100644 index 000000000000..a6834f23afec --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hbase_source_py3.py @@ -0,0 +1,52 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class HBaseSource(CopySource): + """A copy activity HBase server source. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Required. Constant filled by server. + :type type: str + :param query: A query to retrieve data from source. Type: string (or + Expression with resultType string). + :type query: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'query': {'key': 'query', 'type': 'object'}, + } + + def __init__(self, *, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None, **kwargs) -> None: + super(HBaseSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, **kwargs) + self.query = query + self.type = 'HBaseSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_hive_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_hive_activity.py index 32c58c55004e..21b78dbc4d24 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_hive_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_hive_activity.py @@ -15,16 +15,18 @@ class HDInsightHiveActivity(ExecutionActivity): """HDInsight Hive activity type. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] - :param name: Activity name. + :param name: Required. Activity name. :type name: str :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. :type linked_service_name: @@ -71,12 +73,12 @@ class HDInsightHiveActivity(ExecutionActivity): 'defines': {'key': 'typeProperties.defines', 'type': '{object}'}, } - def __init__(self, name, additional_properties=None, description=None, depends_on=None, linked_service_name=None, policy=None, storage_linked_services=None, arguments=None, get_debug_info=None, script_path=None, script_linked_service=None, defines=None): - super(HDInsightHiveActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, linked_service_name=linked_service_name, policy=policy) - self.storage_linked_services = storage_linked_services - self.arguments = arguments - self.get_debug_info = get_debug_info - self.script_path = script_path - self.script_linked_service = script_linked_service - self.defines = defines + def __init__(self, **kwargs): + super(HDInsightHiveActivity, self).__init__(**kwargs) + self.storage_linked_services = kwargs.get('storage_linked_services', None) + self.arguments = kwargs.get('arguments', None) + self.get_debug_info = kwargs.get('get_debug_info', None) + self.script_path = kwargs.get('script_path', None) + self.script_linked_service = kwargs.get('script_linked_service', None) + self.defines = kwargs.get('defines', None) self.type = 'HDInsightHive' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_hive_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_hive_activity_py3.py new file mode 100644 index 000000000000..2bcb4b1c5df4 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_hive_activity_py3.py @@ -0,0 +1,84 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .execution_activity import ExecutionActivity + + +class HDInsightHiveActivity(ExecutionActivity): + """HDInsight Hive activity type. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param name: Required. Activity name. + :type name: str + :param description: Activity description. + :type description: str + :param depends_on: Activity depends on condition. + :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param type: Required. Constant filled by server. + :type type: str + :param linked_service_name: Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param policy: Activity policy. + :type policy: ~azure.mgmt.datafactory.models.ActivityPolicy + :param storage_linked_services: Storage linked service references. + :type storage_linked_services: + list[~azure.mgmt.datafactory.models.LinkedServiceReference] + :param arguments: User specified arguments to HDInsightActivity. + :type arguments: list[object] + :param get_debug_info: Debug info option. Possible values include: 'None', + 'Always', 'Failure' + :type get_debug_info: str or + ~azure.mgmt.datafactory.models.HDInsightActivityDebugInfoOption + :param script_path: Script path. Type: string (or Expression with + resultType string). + :type script_path: object + :param script_linked_service: Script linked service reference. + :type script_linked_service: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param defines: Allows user to specify defines for Hive job request. + :type defines: dict[str, object] + """ + + _validation = { + 'name': {'required': True}, + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'name': {'key': 'name', 'type': 'str'}, + 'description': {'key': 'description', 'type': 'str'}, + 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, + 'storage_linked_services': {'key': 'typeProperties.storageLinkedServices', 'type': '[LinkedServiceReference]'}, + 'arguments': {'key': 'typeProperties.arguments', 'type': '[object]'}, + 'get_debug_info': {'key': 'typeProperties.getDebugInfo', 'type': 'str'}, + 'script_path': {'key': 'typeProperties.scriptPath', 'type': 'object'}, + 'script_linked_service': {'key': 'typeProperties.scriptLinkedService', 'type': 'LinkedServiceReference'}, + 'defines': {'key': 'typeProperties.defines', 'type': '{object}'}, + } + + def __init__(self, *, name: str, additional_properties=None, description: str=None, depends_on=None, linked_service_name=None, policy=None, storage_linked_services=None, arguments=None, get_debug_info=None, script_path=None, script_linked_service=None, defines=None, **kwargs) -> None: + super(HDInsightHiveActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, linked_service_name=linked_service_name, policy=policy, **kwargs) + self.storage_linked_services = storage_linked_services + self.arguments = arguments + self.get_debug_info = get_debug_info + self.script_path = script_path + self.script_linked_service = script_linked_service + self.defines = defines + self.type = 'HDInsightHive' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_linked_service.py index 017d86784e54..db7e14837305 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_linked_service.py @@ -15,6 +15,8 @@ class HDInsightLinkedService(LinkedService): """HDInsight linked service. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -29,10 +31,10 @@ class HDInsightLinkedService(LinkedService): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param cluster_uri: HDInsight cluster URI. Type: string (or Expression - with resultType string). + :param cluster_uri: Required. HDInsight cluster URI. Type: string (or + Expression with resultType string). :type cluster_uri: object :param user_name: HDInsight cluster user name. Type: string (or Expression with resultType string). @@ -72,12 +74,12 @@ class HDInsightLinkedService(LinkedService): 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, cluster_uri, additional_properties=None, connect_via=None, description=None, parameters=None, annotations=None, user_name=None, password=None, linked_service_name=None, hcatalog_linked_service_name=None, encrypted_credential=None): - super(HDInsightLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations) - self.cluster_uri = cluster_uri - self.user_name = user_name - self.password = password - self.linked_service_name = linked_service_name - self.hcatalog_linked_service_name = hcatalog_linked_service_name - self.encrypted_credential = encrypted_credential + def __init__(self, **kwargs): + super(HDInsightLinkedService, self).__init__(**kwargs) + self.cluster_uri = kwargs.get('cluster_uri', None) + self.user_name = kwargs.get('user_name', None) + self.password = kwargs.get('password', None) + self.linked_service_name = kwargs.get('linked_service_name', None) + self.hcatalog_linked_service_name = kwargs.get('hcatalog_linked_service_name', None) + self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'HDInsight' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_linked_service_py3.py new file mode 100644 index 000000000000..0b5374553827 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_linked_service_py3.py @@ -0,0 +1,85 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class HDInsightLinkedService(LinkedService): + """HDInsight linked service. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param cluster_uri: Required. HDInsight cluster URI. Type: string (or + Expression with resultType string). + :type cluster_uri: object + :param user_name: HDInsight cluster user name. Type: string (or Expression + with resultType string). + :type user_name: object + :param password: HDInsight cluster password. + :type password: ~azure.mgmt.datafactory.models.SecretBase + :param linked_service_name: The Azure Storage linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param hcatalog_linked_service_name: A reference to the Azure SQL linked + service that points to the HCatalog database. + :type hcatalog_linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'cluster_uri': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'cluster_uri': {'key': 'typeProperties.clusterUri', 'type': 'object'}, + 'user_name': {'key': 'typeProperties.userName', 'type': 'object'}, + 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, + 'linked_service_name': {'key': 'typeProperties.linkedServiceName', 'type': 'LinkedServiceReference'}, + 'hcatalog_linked_service_name': {'key': 'typeProperties.hcatalogLinkedServiceName', 'type': 'LinkedServiceReference'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, *, cluster_uri, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, user_name=None, password=None, linked_service_name=None, hcatalog_linked_service_name=None, encrypted_credential=None, **kwargs) -> None: + super(HDInsightLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.cluster_uri = cluster_uri + self.user_name = user_name + self.password = password + self.linked_service_name = linked_service_name + self.hcatalog_linked_service_name = hcatalog_linked_service_name + self.encrypted_credential = encrypted_credential + self.type = 'HDInsight' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_map_reduce_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_map_reduce_activity.py index 38d2a7241563..2f7661fd79de 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_map_reduce_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_map_reduce_activity.py @@ -15,16 +15,18 @@ class HDInsightMapReduceActivity(ExecutionActivity): """HDInsight MapReduce activity type. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] - :param name: Activity name. + :param name: Required. Activity name. :type name: str :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. :type linked_service_name: @@ -40,10 +42,10 @@ class HDInsightMapReduceActivity(ExecutionActivity): 'Always', 'Failure' :type get_debug_info: str or ~azure.mgmt.datafactory.models.HDInsightActivityDebugInfoOption - :param class_name: Class name. Type: string (or Expression with resultType - string). + :param class_name: Required. Class name. Type: string (or Expression with + resultType string). :type class_name: object - :param jar_file_path: Jar path. Type: string (or Expression with + :param jar_file_path: Required. Jar path. Type: string (or Expression with resultType string). :type jar_file_path: object :param jar_linked_service: Jar linked service reference. @@ -81,14 +83,14 @@ class HDInsightMapReduceActivity(ExecutionActivity): 'defines': {'key': 'typeProperties.defines', 'type': '{object}'}, } - def __init__(self, name, class_name, jar_file_path, additional_properties=None, description=None, depends_on=None, linked_service_name=None, policy=None, storage_linked_services=None, arguments=None, get_debug_info=None, jar_linked_service=None, jar_libs=None, defines=None): - super(HDInsightMapReduceActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, linked_service_name=linked_service_name, policy=policy) - self.storage_linked_services = storage_linked_services - self.arguments = arguments - self.get_debug_info = get_debug_info - self.class_name = class_name - self.jar_file_path = jar_file_path - self.jar_linked_service = jar_linked_service - self.jar_libs = jar_libs - self.defines = defines + def __init__(self, **kwargs): + super(HDInsightMapReduceActivity, self).__init__(**kwargs) + self.storage_linked_services = kwargs.get('storage_linked_services', None) + self.arguments = kwargs.get('arguments', None) + self.get_debug_info = kwargs.get('get_debug_info', None) + self.class_name = kwargs.get('class_name', None) + self.jar_file_path = kwargs.get('jar_file_path', None) + self.jar_linked_service = kwargs.get('jar_linked_service', None) + self.jar_libs = kwargs.get('jar_libs', None) + self.defines = kwargs.get('defines', None) self.type = 'HDInsightMapReduce' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_map_reduce_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_map_reduce_activity_py3.py new file mode 100644 index 000000000000..1b781fd08bd9 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_map_reduce_activity_py3.py @@ -0,0 +1,96 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .execution_activity import ExecutionActivity + + +class HDInsightMapReduceActivity(ExecutionActivity): + """HDInsight MapReduce activity type. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param name: Required. Activity name. + :type name: str + :param description: Activity description. + :type description: str + :param depends_on: Activity depends on condition. + :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param type: Required. Constant filled by server. + :type type: str + :param linked_service_name: Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param policy: Activity policy. + :type policy: ~azure.mgmt.datafactory.models.ActivityPolicy + :param storage_linked_services: Storage linked service references. + :type storage_linked_services: + list[~azure.mgmt.datafactory.models.LinkedServiceReference] + :param arguments: User specified arguments to HDInsightActivity. + :type arguments: list[object] + :param get_debug_info: Debug info option. Possible values include: 'None', + 'Always', 'Failure' + :type get_debug_info: str or + ~azure.mgmt.datafactory.models.HDInsightActivityDebugInfoOption + :param class_name: Required. Class name. Type: string (or Expression with + resultType string). + :type class_name: object + :param jar_file_path: Required. Jar path. Type: string (or Expression with + resultType string). + :type jar_file_path: object + :param jar_linked_service: Jar linked service reference. + :type jar_linked_service: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param jar_libs: Jar libs. + :type jar_libs: list[object] + :param defines: Allows user to specify defines for the MapReduce job + request. + :type defines: dict[str, object] + """ + + _validation = { + 'name': {'required': True}, + 'type': {'required': True}, + 'class_name': {'required': True}, + 'jar_file_path': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'name': {'key': 'name', 'type': 'str'}, + 'description': {'key': 'description', 'type': 'str'}, + 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, + 'storage_linked_services': {'key': 'typeProperties.storageLinkedServices', 'type': '[LinkedServiceReference]'}, + 'arguments': {'key': 'typeProperties.arguments', 'type': '[object]'}, + 'get_debug_info': {'key': 'typeProperties.getDebugInfo', 'type': 'str'}, + 'class_name': {'key': 'typeProperties.className', 'type': 'object'}, + 'jar_file_path': {'key': 'typeProperties.jarFilePath', 'type': 'object'}, + 'jar_linked_service': {'key': 'typeProperties.jarLinkedService', 'type': 'LinkedServiceReference'}, + 'jar_libs': {'key': 'typeProperties.jarLibs', 'type': '[object]'}, + 'defines': {'key': 'typeProperties.defines', 'type': '{object}'}, + } + + def __init__(self, *, name: str, class_name, jar_file_path, additional_properties=None, description: str=None, depends_on=None, linked_service_name=None, policy=None, storage_linked_services=None, arguments=None, get_debug_info=None, jar_linked_service=None, jar_libs=None, defines=None, **kwargs) -> None: + super(HDInsightMapReduceActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, linked_service_name=linked_service_name, policy=policy, **kwargs) + self.storage_linked_services = storage_linked_services + self.arguments = arguments + self.get_debug_info = get_debug_info + self.class_name = class_name + self.jar_file_path = jar_file_path + self.jar_linked_service = jar_linked_service + self.jar_libs = jar_libs + self.defines = defines + self.type = 'HDInsightMapReduce' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_on_demand_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_on_demand_linked_service.py index fc7eb34cb00e..8069bd2d2d26 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_on_demand_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_on_demand_linked_service.py @@ -15,6 +15,8 @@ class HDInsightOnDemandLinkedService(LinkedService): """HDInsight ondemand linked service. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -29,37 +31,37 @@ class HDInsightOnDemandLinkedService(LinkedService): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param cluster_size: Number of worker/data nodes in the cluster. + :param cluster_size: Required. Number of worker/data nodes in the cluster. Suggestion value: 4. Type: string (or Expression with resultType string). :type cluster_size: object - :param time_to_live: The allowed idle time for the on-demand HDInsight - cluster. Specifies how long the on-demand HDInsight cluster stays alive - after completion of an activity run if there are no other active jobs in - the cluster. The minimum value is 5 mins. Type: string (or Expression with - resultType string). - :type time_to_live: object - :param version: Version of the HDInsight cluster.  Type: string (or + :param time_to_live: Required. The allowed idle time for the on-demand + HDInsight cluster. Specifies how long the on-demand HDInsight cluster + stays alive after completion of an activity run if there are no other + active jobs in the cluster. The minimum value is 5 mins. Type: string (or Expression with resultType string). + :type time_to_live: object + :param version: Required. Version of the HDInsight cluster.  Type: string + (or Expression with resultType string). :type version: object - :param linked_service_name: Azure Storage linked service to be used by the - on-demand cluster for storing and processing data. + :param linked_service_name: Required. Azure Storage linked service to be + used by the on-demand cluster for storing and processing data. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference - :param host_subscription_id: The customer’s subscription to host the - cluster. Type: string (or Expression with resultType string). + :param host_subscription_id: Required. The customer’s subscription to host + the cluster. Type: string (or Expression with resultType string). :type host_subscription_id: object :param service_principal_id: The service principal id for the hostSubscriptionId. Type: string (or Expression with resultType string). :type service_principal_id: object :param service_principal_key: The key for the service principal id. :type service_principal_key: ~azure.mgmt.datafactory.models.SecretBase - :param tenant: The Tenant id/name to which the service principal belongs. - Type: string (or Expression with resultType string). - :type tenant: object - :param cluster_resource_group: The resource group where the cluster + :param tenant: Required. The Tenant id/name to which the service principal belongs. Type: string (or Expression with resultType string). + :type tenant: object + :param cluster_resource_group: Required. The resource group where the + cluster belongs. Type: string (or Expression with resultType string). :type cluster_resource_group: object :param cluster_name_prefix: The prefix of cluster name, postfix will be distinct with timestamp. Type: string (or Expression with resultType @@ -182,36 +184,36 @@ class HDInsightOnDemandLinkedService(LinkedService): 'zookeeper_node_size': {'key': 'typeProperties.zookeeperNodeSize', 'type': 'object'}, } - def __init__(self, cluster_size, time_to_live, version, linked_service_name, host_subscription_id, tenant, cluster_resource_group, additional_properties=None, connect_via=None, description=None, parameters=None, annotations=None, service_principal_id=None, service_principal_key=None, cluster_name_prefix=None, cluster_user_name=None, cluster_password=None, cluster_ssh_user_name=None, cluster_ssh_password=None, additional_linked_service_names=None, hcatalog_linked_service_name=None, cluster_type=None, spark_version=None, core_configuration=None, h_base_configuration=None, hdfs_configuration=None, hive_configuration=None, map_reduce_configuration=None, oozie_configuration=None, storm_configuration=None, yarn_configuration=None, encrypted_credential=None, head_node_size=None, data_node_size=None, zookeeper_node_size=None): - super(HDInsightOnDemandLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations) - self.cluster_size = cluster_size - self.time_to_live = time_to_live - self.version = version - self.linked_service_name = linked_service_name - self.host_subscription_id = host_subscription_id - self.service_principal_id = service_principal_id - self.service_principal_key = service_principal_key - self.tenant = tenant - self.cluster_resource_group = cluster_resource_group - self.cluster_name_prefix = cluster_name_prefix - self.cluster_user_name = cluster_user_name - self.cluster_password = cluster_password - self.cluster_ssh_user_name = cluster_ssh_user_name - self.cluster_ssh_password = cluster_ssh_password - self.additional_linked_service_names = additional_linked_service_names - self.hcatalog_linked_service_name = hcatalog_linked_service_name - self.cluster_type = cluster_type - self.spark_version = spark_version - self.core_configuration = core_configuration - self.h_base_configuration = h_base_configuration - self.hdfs_configuration = hdfs_configuration - self.hive_configuration = hive_configuration - self.map_reduce_configuration = map_reduce_configuration - self.oozie_configuration = oozie_configuration - self.storm_configuration = storm_configuration - self.yarn_configuration = yarn_configuration - self.encrypted_credential = encrypted_credential - self.head_node_size = head_node_size - self.data_node_size = data_node_size - self.zookeeper_node_size = zookeeper_node_size + def __init__(self, **kwargs): + super(HDInsightOnDemandLinkedService, self).__init__(**kwargs) + self.cluster_size = kwargs.get('cluster_size', None) + self.time_to_live = kwargs.get('time_to_live', None) + self.version = kwargs.get('version', None) + self.linked_service_name = kwargs.get('linked_service_name', None) + self.host_subscription_id = kwargs.get('host_subscription_id', None) + self.service_principal_id = kwargs.get('service_principal_id', None) + self.service_principal_key = kwargs.get('service_principal_key', None) + self.tenant = kwargs.get('tenant', None) + self.cluster_resource_group = kwargs.get('cluster_resource_group', None) + self.cluster_name_prefix = kwargs.get('cluster_name_prefix', None) + self.cluster_user_name = kwargs.get('cluster_user_name', None) + self.cluster_password = kwargs.get('cluster_password', None) + self.cluster_ssh_user_name = kwargs.get('cluster_ssh_user_name', None) + self.cluster_ssh_password = kwargs.get('cluster_ssh_password', None) + self.additional_linked_service_names = kwargs.get('additional_linked_service_names', None) + self.hcatalog_linked_service_name = kwargs.get('hcatalog_linked_service_name', None) + self.cluster_type = kwargs.get('cluster_type', None) + self.spark_version = kwargs.get('spark_version', None) + self.core_configuration = kwargs.get('core_configuration', None) + self.h_base_configuration = kwargs.get('h_base_configuration', None) + self.hdfs_configuration = kwargs.get('hdfs_configuration', None) + self.hive_configuration = kwargs.get('hive_configuration', None) + self.map_reduce_configuration = kwargs.get('map_reduce_configuration', None) + self.oozie_configuration = kwargs.get('oozie_configuration', None) + self.storm_configuration = kwargs.get('storm_configuration', None) + self.yarn_configuration = kwargs.get('yarn_configuration', None) + self.encrypted_credential = kwargs.get('encrypted_credential', None) + self.head_node_size = kwargs.get('head_node_size', None) + self.data_node_size = kwargs.get('data_node_size', None) + self.zookeeper_node_size = kwargs.get('zookeeper_node_size', None) self.type = 'HDInsightOnDemand' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_on_demand_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_on_demand_linked_service_py3.py new file mode 100644 index 000000000000..8bb2cfdb1e91 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_on_demand_linked_service_py3.py @@ -0,0 +1,219 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class HDInsightOnDemandLinkedService(LinkedService): + """HDInsight ondemand linked service. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param cluster_size: Required. Number of worker/data nodes in the cluster. + Suggestion value: 4. Type: string (or Expression with resultType string). + :type cluster_size: object + :param time_to_live: Required. The allowed idle time for the on-demand + HDInsight cluster. Specifies how long the on-demand HDInsight cluster + stays alive after completion of an activity run if there are no other + active jobs in the cluster. The minimum value is 5 mins. Type: string (or + Expression with resultType string). + :type time_to_live: object + :param version: Required. Version of the HDInsight cluster.  Type: string + (or Expression with resultType string). + :type version: object + :param linked_service_name: Required. Azure Storage linked service to be + used by the on-demand cluster for storing and processing data. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param host_subscription_id: Required. The customer’s subscription to host + the cluster. Type: string (or Expression with resultType string). + :type host_subscription_id: object + :param service_principal_id: The service principal id for the + hostSubscriptionId. Type: string (or Expression with resultType string). + :type service_principal_id: object + :param service_principal_key: The key for the service principal id. + :type service_principal_key: ~azure.mgmt.datafactory.models.SecretBase + :param tenant: Required. The Tenant id/name to which the service principal + belongs. Type: string (or Expression with resultType string). + :type tenant: object + :param cluster_resource_group: Required. The resource group where the + cluster belongs. Type: string (or Expression with resultType string). + :type cluster_resource_group: object + :param cluster_name_prefix: The prefix of cluster name, postfix will be + distinct with timestamp. Type: string (or Expression with resultType + string). + :type cluster_name_prefix: object + :param cluster_user_name: The username to access the cluster. Type: string + (or Expression with resultType string). + :type cluster_user_name: object + :param cluster_password: The password to access the cluster. + :type cluster_password: ~azure.mgmt.datafactory.models.SecretBase + :param cluster_ssh_user_name: The username to SSH remotely connect to + cluster’s node (for Linux). Type: string (or Expression with resultType + string). + :type cluster_ssh_user_name: object + :param cluster_ssh_password: The password to SSH remotely connect + cluster’s node (for Linux). + :type cluster_ssh_password: ~azure.mgmt.datafactory.models.SecretBase + :param additional_linked_service_names: Specifies additional storage + accounts for the HDInsight linked service so that the Data Factory service + can register them on your behalf. + :type additional_linked_service_names: + list[~azure.mgmt.datafactory.models.LinkedServiceReference] + :param hcatalog_linked_service_name: The name of Azure SQL linked service + that point to the HCatalog database. The on-demand HDInsight cluster is + created by using the Azure SQL database as the metastore. + :type hcatalog_linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param cluster_type: The cluster type. Type: string (or Expression with + resultType string). + :type cluster_type: object + :param spark_version: The version of spark if the cluster type is 'spark'. + Type: string (or Expression with resultType string). + :type spark_version: object + :param core_configuration: Specifies the core configuration parameters (as + in core-site.xml) for the HDInsight cluster to be created. + :type core_configuration: object + :param h_base_configuration: Specifies the HBase configuration parameters + (hbase-site.xml) for the HDInsight cluster. + :type h_base_configuration: object + :param hdfs_configuration: Specifies the HDFS configuration parameters + (hdfs-site.xml) for the HDInsight cluster. + :type hdfs_configuration: object + :param hive_configuration: Specifies the hive configuration parameters + (hive-site.xml) for the HDInsight cluster. + :type hive_configuration: object + :param map_reduce_configuration: Specifies the MapReduce configuration + parameters (mapred-site.xml) for the HDInsight cluster. + :type map_reduce_configuration: object + :param oozie_configuration: Specifies the Oozie configuration parameters + (oozie-site.xml) for the HDInsight cluster. + :type oozie_configuration: object + :param storm_configuration: Specifies the Storm configuration parameters + (storm-site.xml) for the HDInsight cluster. + :type storm_configuration: object + :param yarn_configuration: Specifies the Yarn configuration parameters + (yarn-site.xml) for the HDInsight cluster. + :type yarn_configuration: object + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + :param head_node_size: Specifies the size of the head node for the + HDInsight cluster. + :type head_node_size: object + :param data_node_size: Specifies the size of the data node for the + HDInsight cluster. + :type data_node_size: object + :param zookeeper_node_size: Specifies the size of the Zoo Keeper node for + the HDInsight cluster. + :type zookeeper_node_size: object + """ + + _validation = { + 'type': {'required': True}, + 'cluster_size': {'required': True}, + 'time_to_live': {'required': True}, + 'version': {'required': True}, + 'linked_service_name': {'required': True}, + 'host_subscription_id': {'required': True}, + 'tenant': {'required': True}, + 'cluster_resource_group': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'cluster_size': {'key': 'typeProperties.clusterSize', 'type': 'object'}, + 'time_to_live': {'key': 'typeProperties.timeToLive', 'type': 'object'}, + 'version': {'key': 'typeProperties.version', 'type': 'object'}, + 'linked_service_name': {'key': 'typeProperties.linkedServiceName', 'type': 'LinkedServiceReference'}, + 'host_subscription_id': {'key': 'typeProperties.hostSubscriptionId', 'type': 'object'}, + 'service_principal_id': {'key': 'typeProperties.servicePrincipalId', 'type': 'object'}, + 'service_principal_key': {'key': 'typeProperties.servicePrincipalKey', 'type': 'SecretBase'}, + 'tenant': {'key': 'typeProperties.tenant', 'type': 'object'}, + 'cluster_resource_group': {'key': 'typeProperties.clusterResourceGroup', 'type': 'object'}, + 'cluster_name_prefix': {'key': 'typeProperties.clusterNamePrefix', 'type': 'object'}, + 'cluster_user_name': {'key': 'typeProperties.clusterUserName', 'type': 'object'}, + 'cluster_password': {'key': 'typeProperties.clusterPassword', 'type': 'SecretBase'}, + 'cluster_ssh_user_name': {'key': 'typeProperties.clusterSshUserName', 'type': 'object'}, + 'cluster_ssh_password': {'key': 'typeProperties.clusterSshPassword', 'type': 'SecretBase'}, + 'additional_linked_service_names': {'key': 'typeProperties.additionalLinkedServiceNames', 'type': '[LinkedServiceReference]'}, + 'hcatalog_linked_service_name': {'key': 'typeProperties.hcatalogLinkedServiceName', 'type': 'LinkedServiceReference'}, + 'cluster_type': {'key': 'typeProperties.clusterType', 'type': 'object'}, + 'spark_version': {'key': 'typeProperties.sparkVersion', 'type': 'object'}, + 'core_configuration': {'key': 'typeProperties.coreConfiguration', 'type': 'object'}, + 'h_base_configuration': {'key': 'typeProperties.hBaseConfiguration', 'type': 'object'}, + 'hdfs_configuration': {'key': 'typeProperties.hdfsConfiguration', 'type': 'object'}, + 'hive_configuration': {'key': 'typeProperties.hiveConfiguration', 'type': 'object'}, + 'map_reduce_configuration': {'key': 'typeProperties.mapReduceConfiguration', 'type': 'object'}, + 'oozie_configuration': {'key': 'typeProperties.oozieConfiguration', 'type': 'object'}, + 'storm_configuration': {'key': 'typeProperties.stormConfiguration', 'type': 'object'}, + 'yarn_configuration': {'key': 'typeProperties.yarnConfiguration', 'type': 'object'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + 'head_node_size': {'key': 'typeProperties.headNodeSize', 'type': 'object'}, + 'data_node_size': {'key': 'typeProperties.dataNodeSize', 'type': 'object'}, + 'zookeeper_node_size': {'key': 'typeProperties.zookeeperNodeSize', 'type': 'object'}, + } + + def __init__(self, *, cluster_size, time_to_live, version, linked_service_name, host_subscription_id, tenant, cluster_resource_group, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, service_principal_id=None, service_principal_key=None, cluster_name_prefix=None, cluster_user_name=None, cluster_password=None, cluster_ssh_user_name=None, cluster_ssh_password=None, additional_linked_service_names=None, hcatalog_linked_service_name=None, cluster_type=None, spark_version=None, core_configuration=None, h_base_configuration=None, hdfs_configuration=None, hive_configuration=None, map_reduce_configuration=None, oozie_configuration=None, storm_configuration=None, yarn_configuration=None, encrypted_credential=None, head_node_size=None, data_node_size=None, zookeeper_node_size=None, **kwargs) -> None: + super(HDInsightOnDemandLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.cluster_size = cluster_size + self.time_to_live = time_to_live + self.version = version + self.linked_service_name = linked_service_name + self.host_subscription_id = host_subscription_id + self.service_principal_id = service_principal_id + self.service_principal_key = service_principal_key + self.tenant = tenant + self.cluster_resource_group = cluster_resource_group + self.cluster_name_prefix = cluster_name_prefix + self.cluster_user_name = cluster_user_name + self.cluster_password = cluster_password + self.cluster_ssh_user_name = cluster_ssh_user_name + self.cluster_ssh_password = cluster_ssh_password + self.additional_linked_service_names = additional_linked_service_names + self.hcatalog_linked_service_name = hcatalog_linked_service_name + self.cluster_type = cluster_type + self.spark_version = spark_version + self.core_configuration = core_configuration + self.h_base_configuration = h_base_configuration + self.hdfs_configuration = hdfs_configuration + self.hive_configuration = hive_configuration + self.map_reduce_configuration = map_reduce_configuration + self.oozie_configuration = oozie_configuration + self.storm_configuration = storm_configuration + self.yarn_configuration = yarn_configuration + self.encrypted_credential = encrypted_credential + self.head_node_size = head_node_size + self.data_node_size = data_node_size + self.zookeeper_node_size = zookeeper_node_size + self.type = 'HDInsightOnDemand' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_pig_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_pig_activity.py index 228857b1926d..a552b1d9ae52 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_pig_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_pig_activity.py @@ -15,16 +15,18 @@ class HDInsightPigActivity(ExecutionActivity): """HDInsight Pig activity type. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] - :param name: Activity name. + :param name: Required. Activity name. :type name: str :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. :type linked_service_name: @@ -71,12 +73,12 @@ class HDInsightPigActivity(ExecutionActivity): 'defines': {'key': 'typeProperties.defines', 'type': '{object}'}, } - def __init__(self, name, additional_properties=None, description=None, depends_on=None, linked_service_name=None, policy=None, storage_linked_services=None, arguments=None, get_debug_info=None, script_path=None, script_linked_service=None, defines=None): - super(HDInsightPigActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, linked_service_name=linked_service_name, policy=policy) - self.storage_linked_services = storage_linked_services - self.arguments = arguments - self.get_debug_info = get_debug_info - self.script_path = script_path - self.script_linked_service = script_linked_service - self.defines = defines + def __init__(self, **kwargs): + super(HDInsightPigActivity, self).__init__(**kwargs) + self.storage_linked_services = kwargs.get('storage_linked_services', None) + self.arguments = kwargs.get('arguments', None) + self.get_debug_info = kwargs.get('get_debug_info', None) + self.script_path = kwargs.get('script_path', None) + self.script_linked_service = kwargs.get('script_linked_service', None) + self.defines = kwargs.get('defines', None) self.type = 'HDInsightPig' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_pig_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_pig_activity_py3.py new file mode 100644 index 000000000000..a2cbf90ac518 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_pig_activity_py3.py @@ -0,0 +1,84 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .execution_activity import ExecutionActivity + + +class HDInsightPigActivity(ExecutionActivity): + """HDInsight Pig activity type. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param name: Required. Activity name. + :type name: str + :param description: Activity description. + :type description: str + :param depends_on: Activity depends on condition. + :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param type: Required. Constant filled by server. + :type type: str + :param linked_service_name: Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param policy: Activity policy. + :type policy: ~azure.mgmt.datafactory.models.ActivityPolicy + :param storage_linked_services: Storage linked service references. + :type storage_linked_services: + list[~azure.mgmt.datafactory.models.LinkedServiceReference] + :param arguments: User specified arguments to HDInsightActivity. + :type arguments: list[object] + :param get_debug_info: Debug info option. Possible values include: 'None', + 'Always', 'Failure' + :type get_debug_info: str or + ~azure.mgmt.datafactory.models.HDInsightActivityDebugInfoOption + :param script_path: Script path. Type: string (or Expression with + resultType string). + :type script_path: object + :param script_linked_service: Script linked service reference. + :type script_linked_service: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param defines: Allows user to specify defines for Pig job request. + :type defines: dict[str, object] + """ + + _validation = { + 'name': {'required': True}, + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'name': {'key': 'name', 'type': 'str'}, + 'description': {'key': 'description', 'type': 'str'}, + 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, + 'storage_linked_services': {'key': 'typeProperties.storageLinkedServices', 'type': '[LinkedServiceReference]'}, + 'arguments': {'key': 'typeProperties.arguments', 'type': '[object]'}, + 'get_debug_info': {'key': 'typeProperties.getDebugInfo', 'type': 'str'}, + 'script_path': {'key': 'typeProperties.scriptPath', 'type': 'object'}, + 'script_linked_service': {'key': 'typeProperties.scriptLinkedService', 'type': 'LinkedServiceReference'}, + 'defines': {'key': 'typeProperties.defines', 'type': '{object}'}, + } + + def __init__(self, *, name: str, additional_properties=None, description: str=None, depends_on=None, linked_service_name=None, policy=None, storage_linked_services=None, arguments=None, get_debug_info=None, script_path=None, script_linked_service=None, defines=None, **kwargs) -> None: + super(HDInsightPigActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, linked_service_name=linked_service_name, policy=policy, **kwargs) + self.storage_linked_services = storage_linked_services + self.arguments = arguments + self.get_debug_info = get_debug_info + self.script_path = script_path + self.script_linked_service = script_linked_service + self.defines = defines + self.type = 'HDInsightPig' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_spark_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_spark_activity.py index c7847b612628..4984385f5d26 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_spark_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_spark_activity.py @@ -15,28 +15,30 @@ class HDInsightSparkActivity(ExecutionActivity): """HDInsight Spark activity. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] - :param name: Activity name. + :param name: Required. Activity name. :type name: str :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference :param policy: Activity policy. :type policy: ~azure.mgmt.datafactory.models.ActivityPolicy - :param root_path: The root path in 'sparkJobLinkedService' for all the - job’s files. Type: string (or Expression with resultType string). + :param root_path: Required. The root path in 'sparkJobLinkedService' for + all the job’s files. Type: string (or Expression with resultType string). :type root_path: object - :param entry_file_path: The relative path to the root folder of the - code/package to be executed. Type: string (or Expression with resultType - string). + :param entry_file_path: Required. The relative path to the root folder of + the code/package to be executed. Type: string (or Expression with + resultType string). :type entry_file_path: object :param arguments: The user-specified arguments to HDInsightSparkActivity. :type arguments: list[object] @@ -82,14 +84,14 @@ class HDInsightSparkActivity(ExecutionActivity): 'spark_config': {'key': 'typeProperties.sparkConfig', 'type': '{object}'}, } - def __init__(self, name, root_path, entry_file_path, additional_properties=None, description=None, depends_on=None, linked_service_name=None, policy=None, arguments=None, get_debug_info=None, spark_job_linked_service=None, class_name=None, proxy_user=None, spark_config=None): - super(HDInsightSparkActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, linked_service_name=linked_service_name, policy=policy) - self.root_path = root_path - self.entry_file_path = entry_file_path - self.arguments = arguments - self.get_debug_info = get_debug_info - self.spark_job_linked_service = spark_job_linked_service - self.class_name = class_name - self.proxy_user = proxy_user - self.spark_config = spark_config + def __init__(self, **kwargs): + super(HDInsightSparkActivity, self).__init__(**kwargs) + self.root_path = kwargs.get('root_path', None) + self.entry_file_path = kwargs.get('entry_file_path', None) + self.arguments = kwargs.get('arguments', None) + self.get_debug_info = kwargs.get('get_debug_info', None) + self.spark_job_linked_service = kwargs.get('spark_job_linked_service', None) + self.class_name = kwargs.get('class_name', None) + self.proxy_user = kwargs.get('proxy_user', None) + self.spark_config = kwargs.get('spark_config', None) self.type = 'HDInsightSpark' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_spark_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_spark_activity_py3.py new file mode 100644 index 000000000000..118772db8539 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_spark_activity_py3.py @@ -0,0 +1,97 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .execution_activity import ExecutionActivity + + +class HDInsightSparkActivity(ExecutionActivity): + """HDInsight Spark activity. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param name: Required. Activity name. + :type name: str + :param description: Activity description. + :type description: str + :param depends_on: Activity depends on condition. + :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param type: Required. Constant filled by server. + :type type: str + :param linked_service_name: Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param policy: Activity policy. + :type policy: ~azure.mgmt.datafactory.models.ActivityPolicy + :param root_path: Required. The root path in 'sparkJobLinkedService' for + all the job’s files. Type: string (or Expression with resultType string). + :type root_path: object + :param entry_file_path: Required. The relative path to the root folder of + the code/package to be executed. Type: string (or Expression with + resultType string). + :type entry_file_path: object + :param arguments: The user-specified arguments to HDInsightSparkActivity. + :type arguments: list[object] + :param get_debug_info: Debug info option. Possible values include: 'None', + 'Always', 'Failure' + :type get_debug_info: str or + ~azure.mgmt.datafactory.models.HDInsightActivityDebugInfoOption + :param spark_job_linked_service: The storage linked service for uploading + the entry file and dependencies, and for receiving logs. + :type spark_job_linked_service: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param class_name: The application's Java/Spark main class. + :type class_name: str + :param proxy_user: The user to impersonate that will execute the job. + Type: string (or Expression with resultType string). + :type proxy_user: object + :param spark_config: Spark configuration property. + :type spark_config: dict[str, object] + """ + + _validation = { + 'name': {'required': True}, + 'type': {'required': True}, + 'root_path': {'required': True}, + 'entry_file_path': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'name': {'key': 'name', 'type': 'str'}, + 'description': {'key': 'description', 'type': 'str'}, + 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, + 'root_path': {'key': 'typeProperties.rootPath', 'type': 'object'}, + 'entry_file_path': {'key': 'typeProperties.entryFilePath', 'type': 'object'}, + 'arguments': {'key': 'typeProperties.arguments', 'type': '[object]'}, + 'get_debug_info': {'key': 'typeProperties.getDebugInfo', 'type': 'str'}, + 'spark_job_linked_service': {'key': 'typeProperties.sparkJobLinkedService', 'type': 'LinkedServiceReference'}, + 'class_name': {'key': 'typeProperties.className', 'type': 'str'}, + 'proxy_user': {'key': 'typeProperties.proxyUser', 'type': 'object'}, + 'spark_config': {'key': 'typeProperties.sparkConfig', 'type': '{object}'}, + } + + def __init__(self, *, name: str, root_path, entry_file_path, additional_properties=None, description: str=None, depends_on=None, linked_service_name=None, policy=None, arguments=None, get_debug_info=None, spark_job_linked_service=None, class_name: str=None, proxy_user=None, spark_config=None, **kwargs) -> None: + super(HDInsightSparkActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, linked_service_name=linked_service_name, policy=policy, **kwargs) + self.root_path = root_path + self.entry_file_path = entry_file_path + self.arguments = arguments + self.get_debug_info = get_debug_info + self.spark_job_linked_service = spark_job_linked_service + self.class_name = class_name + self.proxy_user = proxy_user + self.spark_config = spark_config + self.type = 'HDInsightSpark' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_streaming_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_streaming_activity.py index bb10b7578dce..80469577315f 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_streaming_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_streaming_activity.py @@ -15,16 +15,18 @@ class HDInsightStreamingActivity(ExecutionActivity): """HDInsight streaming activity type. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] - :param name: Activity name. + :param name: Required. Activity name. :type name: str :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. :type linked_service_name: @@ -40,19 +42,20 @@ class HDInsightStreamingActivity(ExecutionActivity): 'Always', 'Failure' :type get_debug_info: str or ~azure.mgmt.datafactory.models.HDInsightActivityDebugInfoOption - :param mapper: Mapper executable name. Type: string (or Expression with - resultType string). + :param mapper: Required. Mapper executable name. Type: string (or + Expression with resultType string). :type mapper: object - :param reducer: Reducer executable name. Type: string (or Expression with - resultType string). + :param reducer: Required. Reducer executable name. Type: string (or + Expression with resultType string). :type reducer: object - :param input: Input blob path. Type: string (or Expression with resultType - string). - :type input: object - :param output: Output blob path. Type: string (or Expression with + :param input: Required. Input blob path. Type: string (or Expression with resultType string). + :type input: object + :param output: Required. Output blob path. Type: string (or Expression + with resultType string). :type output: object - :param file_paths: Paths to streaming job files. Can be directories. + :param file_paths: Required. Paths to streaming job files. Can be + directories. :type file_paths: list[object] :param file_linked_service: Linked service reference where the files are located. @@ -99,18 +102,18 @@ class HDInsightStreamingActivity(ExecutionActivity): 'defines': {'key': 'typeProperties.defines', 'type': '{object}'}, } - def __init__(self, name, mapper, reducer, input, output, file_paths, additional_properties=None, description=None, depends_on=None, linked_service_name=None, policy=None, storage_linked_services=None, arguments=None, get_debug_info=None, file_linked_service=None, combiner=None, command_environment=None, defines=None): - super(HDInsightStreamingActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, linked_service_name=linked_service_name, policy=policy) - self.storage_linked_services = storage_linked_services - self.arguments = arguments - self.get_debug_info = get_debug_info - self.mapper = mapper - self.reducer = reducer - self.input = input - self.output = output - self.file_paths = file_paths - self.file_linked_service = file_linked_service - self.combiner = combiner - self.command_environment = command_environment - self.defines = defines + def __init__(self, **kwargs): + super(HDInsightStreamingActivity, self).__init__(**kwargs) + self.storage_linked_services = kwargs.get('storage_linked_services', None) + self.arguments = kwargs.get('arguments', None) + self.get_debug_info = kwargs.get('get_debug_info', None) + self.mapper = kwargs.get('mapper', None) + self.reducer = kwargs.get('reducer', None) + self.input = kwargs.get('input', None) + self.output = kwargs.get('output', None) + self.file_paths = kwargs.get('file_paths', None) + self.file_linked_service = kwargs.get('file_linked_service', None) + self.combiner = kwargs.get('combiner', None) + self.command_environment = kwargs.get('command_environment', None) + self.defines = kwargs.get('defines', None) self.type = 'HDInsightStreaming' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_streaming_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_streaming_activity_py3.py new file mode 100644 index 000000000000..e42e543bebac --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_streaming_activity_py3.py @@ -0,0 +1,119 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .execution_activity import ExecutionActivity + + +class HDInsightStreamingActivity(ExecutionActivity): + """HDInsight streaming activity type. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param name: Required. Activity name. + :type name: str + :param description: Activity description. + :type description: str + :param depends_on: Activity depends on condition. + :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param type: Required. Constant filled by server. + :type type: str + :param linked_service_name: Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param policy: Activity policy. + :type policy: ~azure.mgmt.datafactory.models.ActivityPolicy + :param storage_linked_services: Storage linked service references. + :type storage_linked_services: + list[~azure.mgmt.datafactory.models.LinkedServiceReference] + :param arguments: User specified arguments to HDInsightActivity. + :type arguments: list[object] + :param get_debug_info: Debug info option. Possible values include: 'None', + 'Always', 'Failure' + :type get_debug_info: str or + ~azure.mgmt.datafactory.models.HDInsightActivityDebugInfoOption + :param mapper: Required. Mapper executable name. Type: string (or + Expression with resultType string). + :type mapper: object + :param reducer: Required. Reducer executable name. Type: string (or + Expression with resultType string). + :type reducer: object + :param input: Required. Input blob path. Type: string (or Expression with + resultType string). + :type input: object + :param output: Required. Output blob path. Type: string (or Expression + with resultType string). + :type output: object + :param file_paths: Required. Paths to streaming job files. Can be + directories. + :type file_paths: list[object] + :param file_linked_service: Linked service reference where the files are + located. + :type file_linked_service: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param combiner: Combiner executable name. Type: string (or Expression + with resultType string). + :type combiner: object + :param command_environment: Command line environment values. + :type command_environment: list[object] + :param defines: Allows user to specify defines for streaming job request. + :type defines: dict[str, object] + """ + + _validation = { + 'name': {'required': True}, + 'type': {'required': True}, + 'mapper': {'required': True}, + 'reducer': {'required': True}, + 'input': {'required': True}, + 'output': {'required': True}, + 'file_paths': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'name': {'key': 'name', 'type': 'str'}, + 'description': {'key': 'description', 'type': 'str'}, + 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, + 'storage_linked_services': {'key': 'typeProperties.storageLinkedServices', 'type': '[LinkedServiceReference]'}, + 'arguments': {'key': 'typeProperties.arguments', 'type': '[object]'}, + 'get_debug_info': {'key': 'typeProperties.getDebugInfo', 'type': 'str'}, + 'mapper': {'key': 'typeProperties.mapper', 'type': 'object'}, + 'reducer': {'key': 'typeProperties.reducer', 'type': 'object'}, + 'input': {'key': 'typeProperties.input', 'type': 'object'}, + 'output': {'key': 'typeProperties.output', 'type': 'object'}, + 'file_paths': {'key': 'typeProperties.filePaths', 'type': '[object]'}, + 'file_linked_service': {'key': 'typeProperties.fileLinkedService', 'type': 'LinkedServiceReference'}, + 'combiner': {'key': 'typeProperties.combiner', 'type': 'object'}, + 'command_environment': {'key': 'typeProperties.commandEnvironment', 'type': '[object]'}, + 'defines': {'key': 'typeProperties.defines', 'type': '{object}'}, + } + + def __init__(self, *, name: str, mapper, reducer, input, output, file_paths, additional_properties=None, description: str=None, depends_on=None, linked_service_name=None, policy=None, storage_linked_services=None, arguments=None, get_debug_info=None, file_linked_service=None, combiner=None, command_environment=None, defines=None, **kwargs) -> None: + super(HDInsightStreamingActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, linked_service_name=linked_service_name, policy=policy, **kwargs) + self.storage_linked_services = storage_linked_services + self.arguments = arguments + self.get_debug_info = get_debug_info + self.mapper = mapper + self.reducer = reducer + self.input = input + self.output = output + self.file_paths = file_paths + self.file_linked_service = file_linked_service + self.combiner = combiner + self.command_environment = command_environment + self.defines = defines + self.type = 'HDInsightStreaming' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hdfs_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hdfs_linked_service.py index d0253c69ee5d..ab26ae10fe8c 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hdfs_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hdfs_linked_service.py @@ -15,6 +15,8 @@ class HdfsLinkedService(LinkedService): """Hadoop Distributed File System (HDFS) linked service. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -29,9 +31,9 @@ class HdfsLinkedService(LinkedService): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param url: The URL of the HDFS service endpoint, e.g. + :param url: Required. The URL of the HDFS service endpoint, e.g. http://myhostname:50070/webhdfs/v1 . Type: string (or Expression with resultType string). :type url: object @@ -69,11 +71,11 @@ class HdfsLinkedService(LinkedService): 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, } - def __init__(self, url, additional_properties=None, connect_via=None, description=None, parameters=None, annotations=None, authentication_type=None, encrypted_credential=None, user_name=None, password=None): - super(HdfsLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations) - self.url = url - self.authentication_type = authentication_type - self.encrypted_credential = encrypted_credential - self.user_name = user_name - self.password = password + def __init__(self, **kwargs): + super(HdfsLinkedService, self).__init__(**kwargs) + self.url = kwargs.get('url', None) + self.authentication_type = kwargs.get('authentication_type', None) + self.encrypted_credential = kwargs.get('encrypted_credential', None) + self.user_name = kwargs.get('user_name', None) + self.password = kwargs.get('password', None) self.type = 'Hdfs' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hdfs_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hdfs_linked_service_py3.py new file mode 100644 index 000000000000..550b4824c181 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hdfs_linked_service_py3.py @@ -0,0 +1,81 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class HdfsLinkedService(LinkedService): + """Hadoop Distributed File System (HDFS) linked service. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param url: Required. The URL of the HDFS service endpoint, e.g. + http://myhostname:50070/webhdfs/v1 . Type: string (or Expression with + resultType string). + :type url: object + :param authentication_type: Type of authentication used to connect to the + HDFS. Possible values are: Anonymous and Windows. Type: string (or + Expression with resultType string). + :type authentication_type: object + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + :param user_name: User name for Windows authentication. Type: string (or + Expression with resultType string). + :type user_name: object + :param password: Password for Windows authentication. + :type password: ~azure.mgmt.datafactory.models.SecretBase + """ + + _validation = { + 'type': {'required': True}, + 'url': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'url': {'key': 'typeProperties.url', 'type': 'object'}, + 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'object'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + 'user_name': {'key': 'typeProperties.userName', 'type': 'object'}, + 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, + } + + def __init__(self, *, url, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, authentication_type=None, encrypted_credential=None, user_name=None, password=None, **kwargs) -> None: + super(HdfsLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.url = url + self.authentication_type = authentication_type + self.encrypted_credential = encrypted_credential + self.user_name = user_name + self.password = password + self.type = 'Hdfs' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hdfs_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hdfs_source.py index 6517f02d2531..1322a0e68cea 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hdfs_source.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hdfs_source.py @@ -15,6 +15,8 @@ class HdfsSource(CopySource): """A copy activity HDFS source. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -25,7 +27,7 @@ class HdfsSource(CopySource): with resultType string), pattern: ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type source_retry_wait: object - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param recursive: If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType @@ -48,8 +50,8 @@ class HdfsSource(CopySource): 'distcp_settings': {'key': 'distcpSettings', 'type': 'DistcpSettings'}, } - def __init__(self, additional_properties=None, source_retry_count=None, source_retry_wait=None, recursive=None, distcp_settings=None): - super(HdfsSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait) - self.recursive = recursive - self.distcp_settings = distcp_settings + def __init__(self, **kwargs): + super(HdfsSource, self).__init__(**kwargs) + self.recursive = kwargs.get('recursive', None) + self.distcp_settings = kwargs.get('distcp_settings', None) self.type = 'HdfsSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hdfs_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hdfs_source_py3.py new file mode 100644 index 000000000000..e7125240a2ac --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hdfs_source_py3.py @@ -0,0 +1,57 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class HdfsSource(CopySource): + """A copy activity HDFS source. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Required. Constant filled by server. + :type type: str + :param recursive: If true, files under the folder path will be read + recursively. Default is true. Type: boolean (or Expression with resultType + boolean). + :type recursive: object + :param distcp_settings: Specifies Distcp-related settings. + :type distcp_settings: ~azure.mgmt.datafactory.models.DistcpSettings + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'recursive': {'key': 'recursive', 'type': 'object'}, + 'distcp_settings': {'key': 'distcpSettings', 'type': 'DistcpSettings'}, + } + + def __init__(self, *, additional_properties=None, source_retry_count=None, source_retry_wait=None, recursive=None, distcp_settings=None, **kwargs) -> None: + super(HdfsSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, **kwargs) + self.recursive = recursive + self.distcp_settings = distcp_settings + self.type = 'HdfsSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hive_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hive_linked_service.py index a3068d5fc5f4..57b40c30304a 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hive_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hive_linked_service.py @@ -15,6 +15,8 @@ class HiveLinkedService(LinkedService): """Hive Server linked service. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -29,10 +31,11 @@ class HiveLinkedService(LinkedService): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param host: IP address or host name of the Hive server, separated by ';' - for multiple hosts (only when serviceDiscoveryMode is enable). + :param host: Required. IP address or host name of the Hive server, + separated by ';' for multiple hosts (only when serviceDiscoveryMode is + enable). :type host: object :param port: The TCP port that the Hive server uses to listen for client connections. @@ -44,8 +47,8 @@ class HiveLinkedService(LinkedService): Thrift layer. Possible values include: 'Binary', 'SASL', 'HTTP ' :type thrift_transport_protocol: str or ~azure.mgmt.datafactory.models.HiveThriftTransportProtocol - :param authentication_type: The authentication method used to access the - Hive server. Possible values include: 'Anonymous', 'Username', + :param authentication_type: Required. The authentication method used to + access the Hive server. Possible values include: 'Anonymous', 'Username', 'UsernameAndPassword', 'WindowsAzureHDInsightService' :type authentication_type: str or ~azure.mgmt.datafactory.models.HiveAuthenticationType @@ -122,23 +125,23 @@ class HiveLinkedService(LinkedService): 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, host, authentication_type, additional_properties=None, connect_via=None, description=None, parameters=None, annotations=None, port=None, server_type=None, thrift_transport_protocol=None, service_discovery_mode=None, zoo_keeper_name_space=None, use_native_query=None, username=None, password=None, http_path=None, enable_ssl=None, trusted_cert_path=None, use_system_trust_store=None, allow_host_name_cn_mismatch=None, allow_self_signed_server_cert=None, encrypted_credential=None): - super(HiveLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations) - self.host = host - self.port = port - self.server_type = server_type - self.thrift_transport_protocol = thrift_transport_protocol - self.authentication_type = authentication_type - self.service_discovery_mode = service_discovery_mode - self.zoo_keeper_name_space = zoo_keeper_name_space - self.use_native_query = use_native_query - self.username = username - self.password = password - self.http_path = http_path - self.enable_ssl = enable_ssl - self.trusted_cert_path = trusted_cert_path - self.use_system_trust_store = use_system_trust_store - self.allow_host_name_cn_mismatch = allow_host_name_cn_mismatch - self.allow_self_signed_server_cert = allow_self_signed_server_cert - self.encrypted_credential = encrypted_credential + def __init__(self, **kwargs): + super(HiveLinkedService, self).__init__(**kwargs) + self.host = kwargs.get('host', None) + self.port = kwargs.get('port', None) + self.server_type = kwargs.get('server_type', None) + self.thrift_transport_protocol = kwargs.get('thrift_transport_protocol', None) + self.authentication_type = kwargs.get('authentication_type', None) + self.service_discovery_mode = kwargs.get('service_discovery_mode', None) + self.zoo_keeper_name_space = kwargs.get('zoo_keeper_name_space', None) + self.use_native_query = kwargs.get('use_native_query', None) + self.username = kwargs.get('username', None) + self.password = kwargs.get('password', None) + self.http_path = kwargs.get('http_path', None) + self.enable_ssl = kwargs.get('enable_ssl', None) + self.trusted_cert_path = kwargs.get('trusted_cert_path', None) + self.use_system_trust_store = kwargs.get('use_system_trust_store', None) + self.allow_host_name_cn_mismatch = kwargs.get('allow_host_name_cn_mismatch', None) + self.allow_self_signed_server_cert = kwargs.get('allow_self_signed_server_cert', None) + self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'Hive' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hive_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hive_linked_service_py3.py new file mode 100644 index 000000000000..d6e1a46cbe7e --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hive_linked_service_py3.py @@ -0,0 +1,147 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class HiveLinkedService(LinkedService): + """Hive Server linked service. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param host: Required. IP address or host name of the Hive server, + separated by ';' for multiple hosts (only when serviceDiscoveryMode is + enable). + :type host: object + :param port: The TCP port that the Hive server uses to listen for client + connections. + :type port: object + :param server_type: The type of Hive server. Possible values include: + 'HiveServer1', 'HiveServer2', 'HiveThriftServer' + :type server_type: str or ~azure.mgmt.datafactory.models.HiveServerType + :param thrift_transport_protocol: The transport protocol to use in the + Thrift layer. Possible values include: 'Binary', 'SASL', 'HTTP ' + :type thrift_transport_protocol: str or + ~azure.mgmt.datafactory.models.HiveThriftTransportProtocol + :param authentication_type: Required. The authentication method used to + access the Hive server. Possible values include: 'Anonymous', 'Username', + 'UsernameAndPassword', 'WindowsAzureHDInsightService' + :type authentication_type: str or + ~azure.mgmt.datafactory.models.HiveAuthenticationType + :param service_discovery_mode: true to indicate using the ZooKeeper + service, false not. + :type service_discovery_mode: object + :param zoo_keeper_name_space: The namespace on ZooKeeper under which Hive + Server 2 nodes are added. + :type zoo_keeper_name_space: object + :param use_native_query: Specifies whether the driver uses native HiveQL + queries,or converts them into an equivalent form in HiveQL. + :type use_native_query: object + :param username: The user name that you use to access Hive Server. + :type username: object + :param password: The password corresponding to the user name that you + provided in the Username field + :type password: ~azure.mgmt.datafactory.models.SecretBase + :param http_path: The partial URL corresponding to the Hive server. + :type http_path: object + :param enable_ssl: Specifies whether the connections to the server are + encrypted using SSL. The default value is false. + :type enable_ssl: object + :param trusted_cert_path: The full path of the .pem file containing + trusted CA certificates for verifying the server when connecting over SSL. + This property can only be set when using SSL on self-hosted IR. The + default value is the cacerts.pem file installed with the IR. + :type trusted_cert_path: object + :param use_system_trust_store: Specifies whether to use a CA certificate + from the system trust store or from a specified PEM file. The default + value is false. + :type use_system_trust_store: object + :param allow_host_name_cn_mismatch: Specifies whether to require a + CA-issued SSL certificate name to match the host name of the server when + connecting over SSL. The default value is false. + :type allow_host_name_cn_mismatch: object + :param allow_self_signed_server_cert: Specifies whether to allow + self-signed certificates from the server. The default value is false. + :type allow_self_signed_server_cert: object + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'host': {'required': True}, + 'authentication_type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'host': {'key': 'typeProperties.host', 'type': 'object'}, + 'port': {'key': 'typeProperties.port', 'type': 'object'}, + 'server_type': {'key': 'typeProperties.serverType', 'type': 'str'}, + 'thrift_transport_protocol': {'key': 'typeProperties.thriftTransportProtocol', 'type': 'str'}, + 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, + 'service_discovery_mode': {'key': 'typeProperties.serviceDiscoveryMode', 'type': 'object'}, + 'zoo_keeper_name_space': {'key': 'typeProperties.zooKeeperNameSpace', 'type': 'object'}, + 'use_native_query': {'key': 'typeProperties.useNativeQuery', 'type': 'object'}, + 'username': {'key': 'typeProperties.username', 'type': 'object'}, + 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, + 'http_path': {'key': 'typeProperties.httpPath', 'type': 'object'}, + 'enable_ssl': {'key': 'typeProperties.enableSsl', 'type': 'object'}, + 'trusted_cert_path': {'key': 'typeProperties.trustedCertPath', 'type': 'object'}, + 'use_system_trust_store': {'key': 'typeProperties.useSystemTrustStore', 'type': 'object'}, + 'allow_host_name_cn_mismatch': {'key': 'typeProperties.allowHostNameCNMismatch', 'type': 'object'}, + 'allow_self_signed_server_cert': {'key': 'typeProperties.allowSelfSignedServerCert', 'type': 'object'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, *, host, authentication_type, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, port=None, server_type=None, thrift_transport_protocol=None, service_discovery_mode=None, zoo_keeper_name_space=None, use_native_query=None, username=None, password=None, http_path=None, enable_ssl=None, trusted_cert_path=None, use_system_trust_store=None, allow_host_name_cn_mismatch=None, allow_self_signed_server_cert=None, encrypted_credential=None, **kwargs) -> None: + super(HiveLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.host = host + self.port = port + self.server_type = server_type + self.thrift_transport_protocol = thrift_transport_protocol + self.authentication_type = authentication_type + self.service_discovery_mode = service_discovery_mode + self.zoo_keeper_name_space = zoo_keeper_name_space + self.use_native_query = use_native_query + self.username = username + self.password = password + self.http_path = http_path + self.enable_ssl = enable_ssl + self.trusted_cert_path = trusted_cert_path + self.use_system_trust_store = use_system_trust_store + self.allow_host_name_cn_mismatch = allow_host_name_cn_mismatch + self.allow_self_signed_server_cert = allow_self_signed_server_cert + self.encrypted_credential = encrypted_credential + self.type = 'Hive' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hive_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hive_object_dataset.py index b88fd992a0a8..cfc5620fd776 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hive_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hive_object_dataset.py @@ -15,6 +15,8 @@ class HiveObjectDataset(Dataset): """Hive Server dataset. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -23,7 +25,7 @@ class HiveObjectDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object - :param linked_service_name: Linked service reference. + :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference :param parameters: Parameters for dataset. @@ -32,7 +34,7 @@ class HiveObjectDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str """ @@ -41,6 +43,16 @@ class HiveObjectDataset(Dataset): 'type': {'required': True}, } - def __init__(self, linked_service_name, additional_properties=None, description=None, structure=None, parameters=None, annotations=None): - super(HiveObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations) + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, **kwargs): + super(HiveObjectDataset, self).__init__(**kwargs) self.type = 'HiveObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hive_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hive_object_dataset_py3.py new file mode 100644 index 000000000000..5b245c3543f8 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hive_object_dataset_py3.py @@ -0,0 +1,58 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class HiveObjectDataset(Dataset): + """Hive Server dataset. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Required. Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: + super(HiveObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + self.type = 'HiveObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hive_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hive_source.py index 92f007ea7f26..ad7cd5dc5a8a 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hive_source.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hive_source.py @@ -15,6 +15,8 @@ class HiveSource(CopySource): """A copy activity Hive Server source. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -25,7 +27,7 @@ class HiveSource(CopySource): with resultType string), pattern: ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type source_retry_wait: object - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). @@ -44,7 +46,7 @@ class HiveSource(CopySource): 'query': {'key': 'query', 'type': 'object'}, } - def __init__(self, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None): - super(HiveSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait) - self.query = query + def __init__(self, **kwargs): + super(HiveSource, self).__init__(**kwargs) + self.query = kwargs.get('query', None) self.type = 'HiveSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hive_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hive_source_py3.py new file mode 100644 index 000000000000..aa52f335d530 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hive_source_py3.py @@ -0,0 +1,52 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class HiveSource(CopySource): + """A copy activity Hive Server source. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Required. Constant filled by server. + :type type: str + :param query: A query to retrieve data from source. Type: string (or + Expression with resultType string). + :type query: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'query': {'key': 'query', 'type': 'object'}, + } + + def __init__(self, *, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None, **kwargs) -> None: + super(HiveSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, **kwargs) + self.query = query + self.type = 'HiveSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/http_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/http_dataset.py index 7b84ce8051cc..e2b2325b5294 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/http_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/http_dataset.py @@ -15,6 +15,8 @@ class HttpDataset(Dataset): """A file in an HTTP web server. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -23,7 +25,7 @@ class HttpDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object - :param linked_service_name: Linked service reference. + :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference :param parameters: Parameters for dataset. @@ -32,7 +34,7 @@ class HttpDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param relative_url: The relative URL based on the URL in the HttpLinkedService refers to an HTTP file Type: string (or Expression with @@ -77,12 +79,12 @@ class HttpDataset(Dataset): 'compression': {'key': 'typeProperties.compression', 'type': 'DatasetCompression'}, } - def __init__(self, linked_service_name, additional_properties=None, description=None, structure=None, parameters=None, annotations=None, relative_url=None, request_method=None, request_body=None, additional_headers=None, format=None, compression=None): - super(HttpDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations) - self.relative_url = relative_url - self.request_method = request_method - self.request_body = request_body - self.additional_headers = additional_headers - self.format = format - self.compression = compression + def __init__(self, **kwargs): + super(HttpDataset, self).__init__(**kwargs) + self.relative_url = kwargs.get('relative_url', None) + self.request_method = kwargs.get('request_method', None) + self.request_body = kwargs.get('request_body', None) + self.additional_headers = kwargs.get('additional_headers', None) + self.format = kwargs.get('format', None) + self.compression = kwargs.get('compression', None) self.type = 'HttpFile' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/http_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/http_dataset_py3.py new file mode 100644 index 000000000000..a3e7a17b95d3 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/http_dataset_py3.py @@ -0,0 +1,90 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class HttpDataset(Dataset): + """A file in an HTTP web server. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Required. Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param relative_url: The relative URL based on the URL in the + HttpLinkedService refers to an HTTP file Type: string (or Expression with + resultType string). + :type relative_url: object + :param request_method: The HTTP method for the HTTP request. Type: string + (or Expression with resultType string). + :type request_method: object + :param request_body: The body for the HTTP request. Type: string (or + Expression with resultType string). + :type request_body: object + :param additional_headers: The headers for the HTTP Request. e.g. + request-header-name-1:request-header-value-1 + ... + request-header-name-n:request-header-value-n Type: string (or Expression + with resultType string). + :type additional_headers: object + :param format: The format of files. + :type format: ~azure.mgmt.datafactory.models.DatasetStorageFormat + :param compression: The data compression method used on files. + :type compression: ~azure.mgmt.datafactory.models.DatasetCompression + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'relative_url': {'key': 'typeProperties.relativeUrl', 'type': 'object'}, + 'request_method': {'key': 'typeProperties.requestMethod', 'type': 'object'}, + 'request_body': {'key': 'typeProperties.requestBody', 'type': 'object'}, + 'additional_headers': {'key': 'typeProperties.additionalHeaders', 'type': 'object'}, + 'format': {'key': 'typeProperties.format', 'type': 'DatasetStorageFormat'}, + 'compression': {'key': 'typeProperties.compression', 'type': 'DatasetCompression'}, + } + + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, relative_url=None, request_method=None, request_body=None, additional_headers=None, format=None, compression=None, **kwargs) -> None: + super(HttpDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + self.relative_url = relative_url + self.request_method = request_method + self.request_body = request_body + self.additional_headers = additional_headers + self.format = format + self.compression = compression + self.type = 'HttpFile' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/http_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/http_linked_service.py index 07ad90ad5596..86d6a072925e 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/http_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/http_linked_service.py @@ -15,6 +15,8 @@ class HttpLinkedService(LinkedService): """Linked service for an HTTP source. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -29,9 +31,9 @@ class HttpLinkedService(LinkedService): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param url: The base URL of the HTTP endpoint, e.g. + :param url: Required. The base URL of the HTTP endpoint, e.g. http://www.microsoft.com. Type: string (or Expression with resultType string). :type url: object @@ -90,14 +92,14 @@ class HttpLinkedService(LinkedService): 'enable_server_certificate_validation': {'key': 'typeProperties.enableServerCertificateValidation', 'type': 'object'}, } - def __init__(self, url, additional_properties=None, connect_via=None, description=None, parameters=None, annotations=None, authentication_type=None, user_name=None, password=None, embedded_cert_data=None, cert_thumbprint=None, encrypted_credential=None, enable_server_certificate_validation=None): - super(HttpLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations) - self.url = url - self.authentication_type = authentication_type - self.user_name = user_name - self.password = password - self.embedded_cert_data = embedded_cert_data - self.cert_thumbprint = cert_thumbprint - self.encrypted_credential = encrypted_credential - self.enable_server_certificate_validation = enable_server_certificate_validation + def __init__(self, **kwargs): + super(HttpLinkedService, self).__init__(**kwargs) + self.url = kwargs.get('url', None) + self.authentication_type = kwargs.get('authentication_type', None) + self.user_name = kwargs.get('user_name', None) + self.password = kwargs.get('password', None) + self.embedded_cert_data = kwargs.get('embedded_cert_data', None) + self.cert_thumbprint = kwargs.get('cert_thumbprint', None) + self.encrypted_credential = kwargs.get('encrypted_credential', None) + self.enable_server_certificate_validation = kwargs.get('enable_server_certificate_validation', None) self.type = 'HttpServer' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/http_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/http_linked_service_py3.py new file mode 100644 index 000000000000..044ab5352683 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/http_linked_service_py3.py @@ -0,0 +1,105 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class HttpLinkedService(LinkedService): + """Linked service for an HTTP source. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param url: Required. The base URL of the HTTP endpoint, e.g. + http://www.microsoft.com. Type: string (or Expression with resultType + string). + :type url: object + :param authentication_type: The authentication type to be used to connect + to the HTTP server. Possible values include: 'Basic', 'Anonymous', + 'Digest', 'Windows', 'ClientCertificate' + :type authentication_type: str or + ~azure.mgmt.datafactory.models.HttpAuthenticationType + :param user_name: User name for Basic, Digest, or Windows authentication. + Type: string (or Expression with resultType string). + :type user_name: object + :param password: Password for Basic, Digest, Windows, or ClientCertificate + with EmbeddedCertData authentication. + :type password: ~azure.mgmt.datafactory.models.SecretBase + :param embedded_cert_data: Base64 encoded certificate data for + ClientCertificate authentication. For on-premises copy with + ClientCertificate authentication, either CertThumbprint or + EmbeddedCertData/Password should be specified. Type: string (or Expression + with resultType string). + :type embedded_cert_data: object + :param cert_thumbprint: Thumbprint of certificate for ClientCertificate + authentication. Only valid for on-premises copy. For on-premises copy with + ClientCertificate authentication, either CertThumbprint or + EmbeddedCertData/Password should be specified. Type: string (or Expression + with resultType string). + :type cert_thumbprint: object + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + :param enable_server_certificate_validation: If true, validate the HTTPS + server SSL certificate. Default value is true. Type: boolean (or + Expression with resultType boolean). + :type enable_server_certificate_validation: object + """ + + _validation = { + 'type': {'required': True}, + 'url': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'url': {'key': 'typeProperties.url', 'type': 'object'}, + 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, + 'user_name': {'key': 'typeProperties.userName', 'type': 'object'}, + 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, + 'embedded_cert_data': {'key': 'typeProperties.embeddedCertData', 'type': 'object'}, + 'cert_thumbprint': {'key': 'typeProperties.certThumbprint', 'type': 'object'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + 'enable_server_certificate_validation': {'key': 'typeProperties.enableServerCertificateValidation', 'type': 'object'}, + } + + def __init__(self, *, url, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, authentication_type=None, user_name=None, password=None, embedded_cert_data=None, cert_thumbprint=None, encrypted_credential=None, enable_server_certificate_validation=None, **kwargs) -> None: + super(HttpLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.url = url + self.authentication_type = authentication_type + self.user_name = user_name + self.password = password + self.embedded_cert_data = embedded_cert_data + self.cert_thumbprint = cert_thumbprint + self.encrypted_credential = encrypted_credential + self.enable_server_certificate_validation = enable_server_certificate_validation + self.type = 'HttpServer' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/http_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/http_source.py index 994e992dbd38..8c4a6ef6b8d7 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/http_source.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/http_source.py @@ -15,6 +15,8 @@ class HttpSource(CopySource): """A copy activity source for an HTTP file. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -25,7 +27,7 @@ class HttpSource(CopySource): with resultType string), pattern: ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type source_retry_wait: object - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param http_request_timeout: Specifies the timeout for a HTTP client to get HTTP response from HTTP server. The default value is equivalent to @@ -47,7 +49,7 @@ class HttpSource(CopySource): 'http_request_timeout': {'key': 'httpRequestTimeout', 'type': 'object'}, } - def __init__(self, additional_properties=None, source_retry_count=None, source_retry_wait=None, http_request_timeout=None): - super(HttpSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait) - self.http_request_timeout = http_request_timeout + def __init__(self, **kwargs): + super(HttpSource, self).__init__(**kwargs) + self.http_request_timeout = kwargs.get('http_request_timeout', None) self.type = 'HttpSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/http_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/http_source_py3.py new file mode 100644 index 000000000000..53fd82025b11 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/http_source_py3.py @@ -0,0 +1,55 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class HttpSource(CopySource): + """A copy activity source for an HTTP file. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Required. Constant filled by server. + :type type: str + :param http_request_timeout: Specifies the timeout for a HTTP client to + get HTTP response from HTTP server. The default value is equivalent to + System.Net.HttpWebRequest.Timeout. Type: string (or Expression with + resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type http_request_timeout: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'http_request_timeout': {'key': 'httpRequestTimeout', 'type': 'object'}, + } + + def __init__(self, *, additional_properties=None, source_retry_count=None, source_retry_wait=None, http_request_timeout=None, **kwargs) -> None: + super(HttpSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, **kwargs) + self.http_request_timeout = http_request_timeout + self.type = 'HttpSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_linked_service.py index 7af6e0637ff0..bd446c19c6da 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_linked_service.py @@ -15,6 +15,8 @@ class HubspotLinkedService(LinkedService): """Hubspot Serivce linked service. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -29,9 +31,10 @@ class HubspotLinkedService(LinkedService): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param client_id: The client ID associated with your Hubspot application. + :param client_id: Required. The client ID associated with your Hubspot + application. :type client_id: object :param client_secret: The client secret associated with your Hubspot application. @@ -80,14 +83,14 @@ class HubspotLinkedService(LinkedService): 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, client_id, additional_properties=None, connect_via=None, description=None, parameters=None, annotations=None, client_secret=None, access_token=None, refresh_token=None, use_encrypted_endpoints=None, use_host_verification=None, use_peer_verification=None, encrypted_credential=None): - super(HubspotLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations) - self.client_id = client_id - self.client_secret = client_secret - self.access_token = access_token - self.refresh_token = refresh_token - self.use_encrypted_endpoints = use_encrypted_endpoints - self.use_host_verification = use_host_verification - self.use_peer_verification = use_peer_verification - self.encrypted_credential = encrypted_credential + def __init__(self, **kwargs): + super(HubspotLinkedService, self).__init__(**kwargs) + self.client_id = kwargs.get('client_id', None) + self.client_secret = kwargs.get('client_secret', None) + self.access_token = kwargs.get('access_token', None) + self.refresh_token = kwargs.get('refresh_token', None) + self.use_encrypted_endpoints = kwargs.get('use_encrypted_endpoints', None) + self.use_host_verification = kwargs.get('use_host_verification', None) + self.use_peer_verification = kwargs.get('use_peer_verification', None) + self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'Hubspot' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_linked_service_py3.py new file mode 100644 index 000000000000..216adb7fe228 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_linked_service_py3.py @@ -0,0 +1,96 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class HubspotLinkedService(LinkedService): + """Hubspot Serivce linked service. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param client_id: Required. The client ID associated with your Hubspot + application. + :type client_id: object + :param client_secret: The client secret associated with your Hubspot + application. + :type client_secret: ~azure.mgmt.datafactory.models.SecretBase + :param access_token: The access token obtained when initially + authenticating your OAuth integration. + :type access_token: ~azure.mgmt.datafactory.models.SecretBase + :param refresh_token: The refresh token obtained when initially + authenticating your OAuth integration. + :type refresh_token: ~azure.mgmt.datafactory.models.SecretBase + :param use_encrypted_endpoints: Specifies whether the data source + endpoints are encrypted using HTTPS. The default value is true. + :type use_encrypted_endpoints: object + :param use_host_verification: Specifies whether to require the host name + in the server's certificate to match the host name of the server when + connecting over SSL. The default value is true. + :type use_host_verification: object + :param use_peer_verification: Specifies whether to verify the identity of + the server when connecting over SSL. The default value is true. + :type use_peer_verification: object + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'client_id': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'client_id': {'key': 'typeProperties.clientId', 'type': 'object'}, + 'client_secret': {'key': 'typeProperties.clientSecret', 'type': 'SecretBase'}, + 'access_token': {'key': 'typeProperties.accessToken', 'type': 'SecretBase'}, + 'refresh_token': {'key': 'typeProperties.refreshToken', 'type': 'SecretBase'}, + 'use_encrypted_endpoints': {'key': 'typeProperties.useEncryptedEndpoints', 'type': 'object'}, + 'use_host_verification': {'key': 'typeProperties.useHostVerification', 'type': 'object'}, + 'use_peer_verification': {'key': 'typeProperties.usePeerVerification', 'type': 'object'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, *, client_id, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, client_secret=None, access_token=None, refresh_token=None, use_encrypted_endpoints=None, use_host_verification=None, use_peer_verification=None, encrypted_credential=None, **kwargs) -> None: + super(HubspotLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.client_id = client_id + self.client_secret = client_secret + self.access_token = access_token + self.refresh_token = refresh_token + self.use_encrypted_endpoints = use_encrypted_endpoints + self.use_host_verification = use_host_verification + self.use_peer_verification = use_peer_verification + self.encrypted_credential = encrypted_credential + self.type = 'Hubspot' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_object_dataset.py index 72374f68d90e..a03b68a4f38b 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_object_dataset.py @@ -15,6 +15,8 @@ class HubspotObjectDataset(Dataset): """Hubspot Serivce dataset. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -23,7 +25,7 @@ class HubspotObjectDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object - :param linked_service_name: Linked service reference. + :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference :param parameters: Parameters for dataset. @@ -32,7 +34,7 @@ class HubspotObjectDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str """ @@ -41,6 +43,16 @@ class HubspotObjectDataset(Dataset): 'type': {'required': True}, } - def __init__(self, linked_service_name, additional_properties=None, description=None, structure=None, parameters=None, annotations=None): - super(HubspotObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations) + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, **kwargs): + super(HubspotObjectDataset, self).__init__(**kwargs) self.type = 'HubspotObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_object_dataset_py3.py new file mode 100644 index 000000000000..b8591962805b --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_object_dataset_py3.py @@ -0,0 +1,58 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class HubspotObjectDataset(Dataset): + """Hubspot Serivce dataset. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Required. Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: + super(HubspotObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + self.type = 'HubspotObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_source.py index 5633cbdd31b2..267a6ff3797c 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_source.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_source.py @@ -15,6 +15,8 @@ class HubspotSource(CopySource): """A copy activity Hubspot Serivce source. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -25,7 +27,7 @@ class HubspotSource(CopySource): with resultType string), pattern: ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type source_retry_wait: object - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). @@ -44,7 +46,7 @@ class HubspotSource(CopySource): 'query': {'key': 'query', 'type': 'object'}, } - def __init__(self, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None): - super(HubspotSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait) - self.query = query + def __init__(self, **kwargs): + super(HubspotSource, self).__init__(**kwargs) + self.query = kwargs.get('query', None) self.type = 'HubspotSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_source_py3.py new file mode 100644 index 000000000000..3efc8e70fdd1 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_source_py3.py @@ -0,0 +1,52 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class HubspotSource(CopySource): + """A copy activity Hubspot Serivce source. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Required. Constant filled by server. + :type type: str + :param query: A query to retrieve data from source. Type: string (or + Expression with resultType string). + :type query: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'query': {'key': 'query', 'type': 'object'}, + } + + def __init__(self, *, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None, **kwargs) -> None: + super(HubspotSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, **kwargs) + self.query = query + self.type = 'HubspotSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/if_condition_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/if_condition_activity.py index 9b71b0defd0f..78eae5911550 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/if_condition_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/if_condition_activity.py @@ -17,19 +17,21 @@ class IfConditionActivity(ControlActivity): activities under the ifTrueActivities property or the ifFalseActivities property depending on the result of the expression. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] - :param name: Activity name. + :param name: Required. Activity name. :type name: str :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param expression: An expression that would evaluate to Boolean. This is - used to determine the block of activities (ifTrueActivities or + :param expression: Required. An expression that would evaluate to Boolean. + This is used to determine the block of activities (ifTrueActivities or ifFalseActivities) that will be executed. :type expression: ~azure.mgmt.datafactory.models.Expression :param if_true_activities: List of activities to execute if expression is @@ -59,9 +61,9 @@ class IfConditionActivity(ControlActivity): 'if_false_activities': {'key': 'typeProperties.ifFalseActivities', 'type': '[Activity]'}, } - def __init__(self, name, expression, additional_properties=None, description=None, depends_on=None, if_true_activities=None, if_false_activities=None): - super(IfConditionActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on) - self.expression = expression - self.if_true_activities = if_true_activities - self.if_false_activities = if_false_activities + def __init__(self, **kwargs): + super(IfConditionActivity, self).__init__(**kwargs) + self.expression = kwargs.get('expression', None) + self.if_true_activities = kwargs.get('if_true_activities', None) + self.if_false_activities = kwargs.get('if_false_activities', None) self.type = 'IfCondition' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/if_condition_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/if_condition_activity_py3.py new file mode 100644 index 000000000000..8f63f627ed98 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/if_condition_activity_py3.py @@ -0,0 +1,69 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .control_activity import ControlActivity + + +class IfConditionActivity(ControlActivity): + """This activity evaluates a boolean expression and executes either the + activities under the ifTrueActivities property or the ifFalseActivities + property depending on the result of the expression. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param name: Required. Activity name. + :type name: str + :param description: Activity description. + :type description: str + :param depends_on: Activity depends on condition. + :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param type: Required. Constant filled by server. + :type type: str + :param expression: Required. An expression that would evaluate to Boolean. + This is used to determine the block of activities (ifTrueActivities or + ifFalseActivities) that will be executed. + :type expression: ~azure.mgmt.datafactory.models.Expression + :param if_true_activities: List of activities to execute if expression is + evaluated to true. This is an optional property and if not provided, the + activity will exit without any action. + :type if_true_activities: list[~azure.mgmt.datafactory.models.Activity] + :param if_false_activities: List of activities to execute if expression is + evaluated to false. This is an optional property and if not provided, the + activity will exit without any action. + :type if_false_activities: list[~azure.mgmt.datafactory.models.Activity] + """ + + _validation = { + 'name': {'required': True}, + 'type': {'required': True}, + 'expression': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'name': {'key': 'name', 'type': 'str'}, + 'description': {'key': 'description', 'type': 'str'}, + 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'expression': {'key': 'typeProperties.expression', 'type': 'Expression'}, + 'if_true_activities': {'key': 'typeProperties.ifTrueActivities', 'type': '[Activity]'}, + 'if_false_activities': {'key': 'typeProperties.ifFalseActivities', 'type': '[Activity]'}, + } + + def __init__(self, *, name: str, expression, additional_properties=None, description: str=None, depends_on=None, if_true_activities=None, if_false_activities=None, **kwargs) -> None: + super(IfConditionActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, **kwargs) + self.expression = expression + self.if_true_activities = if_true_activities + self.if_false_activities = if_false_activities + self.type = 'IfCondition' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/impala_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/impala_linked_service.py index 312842a46b92..fdc471ea225f 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/impala_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/impala_linked_service.py @@ -15,6 +15,8 @@ class ImpalaLinkedService(LinkedService): """Impala server linked service. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -29,16 +31,17 @@ class ImpalaLinkedService(LinkedService): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param host: The IP address or host name of the Impala server. (i.e. - 192.168.222.160) + :param host: Required. The IP address or host name of the Impala server. + (i.e. 192.168.222.160) :type host: object :param port: The TCP port that the Impala server uses to listen for client connections. The default value is 21050. :type port: object - :param authentication_type: The authentication type to use. Possible - values include: 'Anonymous', 'SASLUsername', 'UsernameAndPassword' + :param authentication_type: Required. The authentication type to use. + Possible values include: 'Anonymous', 'SASLUsername', + 'UsernameAndPassword' :type authentication_type: str or ~azure.mgmt.datafactory.models.ImpalaAuthenticationType :param username: The user name used to access the Impala server. The @@ -98,17 +101,17 @@ class ImpalaLinkedService(LinkedService): 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, host, authentication_type, additional_properties=None, connect_via=None, description=None, parameters=None, annotations=None, port=None, username=None, password=None, enable_ssl=None, trusted_cert_path=None, use_system_trust_store=None, allow_host_name_cn_mismatch=None, allow_self_signed_server_cert=None, encrypted_credential=None): - super(ImpalaLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations) - self.host = host - self.port = port - self.authentication_type = authentication_type - self.username = username - self.password = password - self.enable_ssl = enable_ssl - self.trusted_cert_path = trusted_cert_path - self.use_system_trust_store = use_system_trust_store - self.allow_host_name_cn_mismatch = allow_host_name_cn_mismatch - self.allow_self_signed_server_cert = allow_self_signed_server_cert - self.encrypted_credential = encrypted_credential + def __init__(self, **kwargs): + super(ImpalaLinkedService, self).__init__(**kwargs) + self.host = kwargs.get('host', None) + self.port = kwargs.get('port', None) + self.authentication_type = kwargs.get('authentication_type', None) + self.username = kwargs.get('username', None) + self.password = kwargs.get('password', None) + self.enable_ssl = kwargs.get('enable_ssl', None) + self.trusted_cert_path = kwargs.get('trusted_cert_path', None) + self.use_system_trust_store = kwargs.get('use_system_trust_store', None) + self.allow_host_name_cn_mismatch = kwargs.get('allow_host_name_cn_mismatch', None) + self.allow_self_signed_server_cert = kwargs.get('allow_self_signed_server_cert', None) + self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'Impala' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/impala_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/impala_linked_service_py3.py new file mode 100644 index 000000000000..3605a3443e04 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/impala_linked_service_py3.py @@ -0,0 +1,117 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class ImpalaLinkedService(LinkedService): + """Impala server linked service. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param host: Required. The IP address or host name of the Impala server. + (i.e. 192.168.222.160) + :type host: object + :param port: The TCP port that the Impala server uses to listen for client + connections. The default value is 21050. + :type port: object + :param authentication_type: Required. The authentication type to use. + Possible values include: 'Anonymous', 'SASLUsername', + 'UsernameAndPassword' + :type authentication_type: str or + ~azure.mgmt.datafactory.models.ImpalaAuthenticationType + :param username: The user name used to access the Impala server. The + default value is anonymous when using SASLUsername. + :type username: object + :param password: The password corresponding to the user name when using + UsernameAndPassword. + :type password: ~azure.mgmt.datafactory.models.SecretBase + :param enable_ssl: Specifies whether the connections to the server are + encrypted using SSL. The default value is false. + :type enable_ssl: object + :param trusted_cert_path: The full path of the .pem file containing + trusted CA certificates for verifying the server when connecting over SSL. + This property can only be set when using SSL on self-hosted IR. The + default value is the cacerts.pem file installed with the IR. + :type trusted_cert_path: object + :param use_system_trust_store: Specifies whether to use a CA certificate + from the system trust store or from a specified PEM file. The default + value is false. + :type use_system_trust_store: object + :param allow_host_name_cn_mismatch: Specifies whether to require a + CA-issued SSL certificate name to match the host name of the server when + connecting over SSL. The default value is false. + :type allow_host_name_cn_mismatch: object + :param allow_self_signed_server_cert: Specifies whether to allow + self-signed certificates from the server. The default value is false. + :type allow_self_signed_server_cert: object + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'host': {'required': True}, + 'authentication_type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'host': {'key': 'typeProperties.host', 'type': 'object'}, + 'port': {'key': 'typeProperties.port', 'type': 'object'}, + 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, + 'username': {'key': 'typeProperties.username', 'type': 'object'}, + 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, + 'enable_ssl': {'key': 'typeProperties.enableSsl', 'type': 'object'}, + 'trusted_cert_path': {'key': 'typeProperties.trustedCertPath', 'type': 'object'}, + 'use_system_trust_store': {'key': 'typeProperties.useSystemTrustStore', 'type': 'object'}, + 'allow_host_name_cn_mismatch': {'key': 'typeProperties.allowHostNameCNMismatch', 'type': 'object'}, + 'allow_self_signed_server_cert': {'key': 'typeProperties.allowSelfSignedServerCert', 'type': 'object'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, *, host, authentication_type, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, port=None, username=None, password=None, enable_ssl=None, trusted_cert_path=None, use_system_trust_store=None, allow_host_name_cn_mismatch=None, allow_self_signed_server_cert=None, encrypted_credential=None, **kwargs) -> None: + super(ImpalaLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.host = host + self.port = port + self.authentication_type = authentication_type + self.username = username + self.password = password + self.enable_ssl = enable_ssl + self.trusted_cert_path = trusted_cert_path + self.use_system_trust_store = use_system_trust_store + self.allow_host_name_cn_mismatch = allow_host_name_cn_mismatch + self.allow_self_signed_server_cert = allow_self_signed_server_cert + self.encrypted_credential = encrypted_credential + self.type = 'Impala' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/impala_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/impala_object_dataset.py index 01e13a4c5838..634a63f99be7 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/impala_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/impala_object_dataset.py @@ -15,6 +15,8 @@ class ImpalaObjectDataset(Dataset): """Impala server dataset. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -23,7 +25,7 @@ class ImpalaObjectDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object - :param linked_service_name: Linked service reference. + :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference :param parameters: Parameters for dataset. @@ -32,7 +34,7 @@ class ImpalaObjectDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str """ @@ -41,6 +43,16 @@ class ImpalaObjectDataset(Dataset): 'type': {'required': True}, } - def __init__(self, linked_service_name, additional_properties=None, description=None, structure=None, parameters=None, annotations=None): - super(ImpalaObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations) + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, **kwargs): + super(ImpalaObjectDataset, self).__init__(**kwargs) self.type = 'ImpalaObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/impala_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/impala_object_dataset_py3.py new file mode 100644 index 000000000000..840e4082b726 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/impala_object_dataset_py3.py @@ -0,0 +1,58 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class ImpalaObjectDataset(Dataset): + """Impala server dataset. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Required. Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: + super(ImpalaObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + self.type = 'ImpalaObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/impala_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/impala_source.py index 6eace84db950..dec8e843d0c6 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/impala_source.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/impala_source.py @@ -15,6 +15,8 @@ class ImpalaSource(CopySource): """A copy activity Impala server source. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -25,7 +27,7 @@ class ImpalaSource(CopySource): with resultType string), pattern: ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type source_retry_wait: object - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). @@ -44,7 +46,7 @@ class ImpalaSource(CopySource): 'query': {'key': 'query', 'type': 'object'}, } - def __init__(self, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None): - super(ImpalaSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait) - self.query = query + def __init__(self, **kwargs): + super(ImpalaSource, self).__init__(**kwargs) + self.query = kwargs.get('query', None) self.type = 'ImpalaSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/impala_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/impala_source_py3.py new file mode 100644 index 000000000000..63c3e8a3492d --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/impala_source_py3.py @@ -0,0 +1,52 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class ImpalaSource(CopySource): + """A copy activity Impala server source. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Required. Constant filled by server. + :type type: str + :param query: A query to retrieve data from source. Type: string (or + Expression with resultType string). + :type query: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'query': {'key': 'query', 'type': 'object'}, + } + + def __init__(self, *, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None, **kwargs) -> None: + super(ImpalaSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, **kwargs) + self.query = query + self.type = 'ImpalaSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime.py index 69e2792fda46..5dd45d16f76e 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime.py @@ -19,12 +19,14 @@ class IntegrationRuntime(Model): You probably want to use the sub-classes and not this class directly. Known sub-classes are: SelfHostedIntegrationRuntime, ManagedIntegrationRuntime + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] :param description: Integration runtime description. :type description: str - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str """ @@ -42,8 +44,8 @@ class IntegrationRuntime(Model): 'type': {'SelfHosted': 'SelfHostedIntegrationRuntime', 'Managed': 'ManagedIntegrationRuntime'} } - def __init__(self, additional_properties=None, description=None): - super(IntegrationRuntime, self).__init__() - self.additional_properties = additional_properties - self.description = description + def __init__(self, **kwargs): + super(IntegrationRuntime, self).__init__(**kwargs) + self.additional_properties = kwargs.get('additional_properties', None) + self.description = kwargs.get('description', None) self.type = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_auth_keys.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_auth_keys.py index e0582ea5cdf7..12ed6925585e 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_auth_keys.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_auth_keys.py @@ -26,7 +26,7 @@ class IntegrationRuntimeAuthKeys(Model): 'auth_key2': {'key': 'authKey2', 'type': 'str'}, } - def __init__(self, auth_key1=None, auth_key2=None): - super(IntegrationRuntimeAuthKeys, self).__init__() - self.auth_key1 = auth_key1 - self.auth_key2 = auth_key2 + def __init__(self, **kwargs): + super(IntegrationRuntimeAuthKeys, self).__init__(**kwargs) + self.auth_key1 = kwargs.get('auth_key1', None) + self.auth_key2 = kwargs.get('auth_key2', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_auth_keys_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_auth_keys_py3.py new file mode 100644 index 000000000000..b807d4cd5b55 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_auth_keys_py3.py @@ -0,0 +1,32 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class IntegrationRuntimeAuthKeys(Model): + """The integration runtime authentication keys. + + :param auth_key1: The primary integration runtime authentication key. + :type auth_key1: str + :param auth_key2: The secondary integration runtime authentication key. + :type auth_key2: str + """ + + _attribute_map = { + 'auth_key1': {'key': 'authKey1', 'type': 'str'}, + 'auth_key2': {'key': 'authKey2', 'type': 'str'}, + } + + def __init__(self, *, auth_key1: str=None, auth_key2: str=None, **kwargs) -> None: + super(IntegrationRuntimeAuthKeys, self).__init__(**kwargs) + self.auth_key1 = auth_key1 + self.auth_key2 = auth_key2 diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_compute_properties.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_compute_properties.py index 1315d632c234..71723579c8aa 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_compute_properties.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_compute_properties.py @@ -50,11 +50,11 @@ class IntegrationRuntimeComputeProperties(Model): 'v_net_properties': {'key': 'vNetProperties', 'type': 'IntegrationRuntimeVNetProperties'}, } - def __init__(self, additional_properties=None, location=None, node_size=None, number_of_nodes=None, max_parallel_executions_per_node=None, v_net_properties=None): - super(IntegrationRuntimeComputeProperties, self).__init__() - self.additional_properties = additional_properties - self.location = location - self.node_size = node_size - self.number_of_nodes = number_of_nodes - self.max_parallel_executions_per_node = max_parallel_executions_per_node - self.v_net_properties = v_net_properties + def __init__(self, **kwargs): + super(IntegrationRuntimeComputeProperties, self).__init__(**kwargs) + self.additional_properties = kwargs.get('additional_properties', None) + self.location = kwargs.get('location', None) + self.node_size = kwargs.get('node_size', None) + self.number_of_nodes = kwargs.get('number_of_nodes', None) + self.max_parallel_executions_per_node = kwargs.get('max_parallel_executions_per_node', None) + self.v_net_properties = kwargs.get('v_net_properties', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_compute_properties_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_compute_properties_py3.py new file mode 100644 index 000000000000..6b9212633ee1 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_compute_properties_py3.py @@ -0,0 +1,60 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class IntegrationRuntimeComputeProperties(Model): + """The compute resource properties for managed integration runtime. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param location: The location for managed integration runtime. The + supported regions could be found on + https://docs.microsoft.com/en-us/azure/data-factory/data-factory-data-movement-activities + :type location: str + :param node_size: The node size requirement to managed integration + runtime. + :type node_size: str + :param number_of_nodes: The required number of nodes for managed + integration runtime. + :type number_of_nodes: int + :param max_parallel_executions_per_node: Maximum parallel executions count + per node for managed integration runtime. + :type max_parallel_executions_per_node: int + :param v_net_properties: VNet properties for managed integration runtime. + :type v_net_properties: + ~azure.mgmt.datafactory.models.IntegrationRuntimeVNetProperties + """ + + _validation = { + 'number_of_nodes': {'minimum': 1}, + 'max_parallel_executions_per_node': {'maximum': 8, 'minimum': 1}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'location': {'key': 'location', 'type': 'str'}, + 'node_size': {'key': 'nodeSize', 'type': 'str'}, + 'number_of_nodes': {'key': 'numberOfNodes', 'type': 'int'}, + 'max_parallel_executions_per_node': {'key': 'maxParallelExecutionsPerNode', 'type': 'int'}, + 'v_net_properties': {'key': 'vNetProperties', 'type': 'IntegrationRuntimeVNetProperties'}, + } + + def __init__(self, *, additional_properties=None, location: str=None, node_size: str=None, number_of_nodes: int=None, max_parallel_executions_per_node: int=None, v_net_properties=None, **kwargs) -> None: + super(IntegrationRuntimeComputeProperties, self).__init__(**kwargs) + self.additional_properties = additional_properties + self.location = location + self.node_size = node_size + self.number_of_nodes = number_of_nodes + self.max_parallel_executions_per_node = max_parallel_executions_per_node + self.v_net_properties = v_net_properties diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_connection_info.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_connection_info.py index 3bdb02304d52..c185f916e8e5 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_connection_info.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_connection_info.py @@ -59,9 +59,9 @@ class IntegrationRuntimeConnectionInfo(Model): 'is_identity_cert_exprired': {'key': 'isIdentityCertExprired', 'type': 'bool'}, } - def __init__(self, additional_properties=None): - super(IntegrationRuntimeConnectionInfo, self).__init__() - self.additional_properties = additional_properties + def __init__(self, **kwargs): + super(IntegrationRuntimeConnectionInfo, self).__init__(**kwargs) + self.additional_properties = kwargs.get('additional_properties', None) self.service_token = None self.identity_cert_thumbprint = None self.host_service_uri = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_connection_info_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_connection_info_py3.py new file mode 100644 index 000000000000..8cc5aceb16d7 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_connection_info_py3.py @@ -0,0 +1,70 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class IntegrationRuntimeConnectionInfo(Model): + """Connection information for encrypting the on-premises data source + credentials. + + Variables are only populated by the server, and will be ignored when + sending a request. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :ivar service_token: The token generated in service. Callers use this + token to authenticate to integration runtime. + :vartype service_token: str + :ivar identity_cert_thumbprint: The integration runtime SSL certificate + thumbprint. Click-Once application uses it to do server validation. + :vartype identity_cert_thumbprint: str + :ivar host_service_uri: The on-premises integration runtime host URL. + :vartype host_service_uri: str + :ivar version: The integration runtime version. + :vartype version: str + :ivar public_key: The public key for encrypting a credential when + transferring the credential to the integration runtime. + :vartype public_key: str + :ivar is_identity_cert_exprired: Whether the identity certificate is + expired. + :vartype is_identity_cert_exprired: bool + """ + + _validation = { + 'service_token': {'readonly': True}, + 'identity_cert_thumbprint': {'readonly': True}, + 'host_service_uri': {'readonly': True}, + 'version': {'readonly': True}, + 'public_key': {'readonly': True}, + 'is_identity_cert_exprired': {'readonly': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'service_token': {'key': 'serviceToken', 'type': 'str'}, + 'identity_cert_thumbprint': {'key': 'identityCertThumbprint', 'type': 'str'}, + 'host_service_uri': {'key': 'hostServiceUri', 'type': 'str'}, + 'version': {'key': 'version', 'type': 'str'}, + 'public_key': {'key': 'publicKey', 'type': 'str'}, + 'is_identity_cert_exprired': {'key': 'isIdentityCertExprired', 'type': 'bool'}, + } + + def __init__(self, *, additional_properties=None, **kwargs) -> None: + super(IntegrationRuntimeConnectionInfo, self).__init__(**kwargs) + self.additional_properties = additional_properties + self.service_token = None + self.identity_cert_thumbprint = None + self.host_service_uri = None + self.version = None + self.public_key = None + self.is_identity_cert_exprired = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_custom_setup_script_properties.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_custom_setup_script_properties.py index b76cc5e39078..44cd5fe5979b 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_custom_setup_script_properties.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_custom_setup_script_properties.py @@ -27,7 +27,7 @@ class IntegrationRuntimeCustomSetupScriptProperties(Model): 'sas_token': {'key': 'sasToken', 'type': 'SecureString'}, } - def __init__(self, blob_container_uri=None, sas_token=None): - super(IntegrationRuntimeCustomSetupScriptProperties, self).__init__() - self.blob_container_uri = blob_container_uri - self.sas_token = sas_token + def __init__(self, **kwargs): + super(IntegrationRuntimeCustomSetupScriptProperties, self).__init__(**kwargs) + self.blob_container_uri = kwargs.get('blob_container_uri', None) + self.sas_token = kwargs.get('sas_token', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_custom_setup_script_properties_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_custom_setup_script_properties_py3.py new file mode 100644 index 000000000000..7f3c08c0b339 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_custom_setup_script_properties_py3.py @@ -0,0 +1,33 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class IntegrationRuntimeCustomSetupScriptProperties(Model): + """Custom setup script properties for a managed dedicated integration runtime. + + :param blob_container_uri: The URI of the Azure blob container that + contains the custom setup script. + :type blob_container_uri: str + :param sas_token: The SAS token of the Azure blob container. + :type sas_token: ~azure.mgmt.datafactory.models.SecureString + """ + + _attribute_map = { + 'blob_container_uri': {'key': 'blobContainerUri', 'type': 'str'}, + 'sas_token': {'key': 'sasToken', 'type': 'SecureString'}, + } + + def __init__(self, *, blob_container_uri: str=None, sas_token=None, **kwargs) -> None: + super(IntegrationRuntimeCustomSetupScriptProperties, self).__init__(**kwargs) + self.blob_container_uri = blob_container_uri + self.sas_token = sas_token diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_monitoring_data.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_monitoring_data.py index aa1feac333d5..f7b695729403 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_monitoring_data.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_monitoring_data.py @@ -27,7 +27,7 @@ class IntegrationRuntimeMonitoringData(Model): 'nodes': {'key': 'nodes', 'type': '[IntegrationRuntimeNodeMonitoringData]'}, } - def __init__(self, name=None, nodes=None): - super(IntegrationRuntimeMonitoringData, self).__init__() - self.name = name - self.nodes = nodes + def __init__(self, **kwargs): + super(IntegrationRuntimeMonitoringData, self).__init__(**kwargs) + self.name = kwargs.get('name', None) + self.nodes = kwargs.get('nodes', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_monitoring_data_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_monitoring_data_py3.py new file mode 100644 index 000000000000..16f3b656c9cc --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_monitoring_data_py3.py @@ -0,0 +1,33 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class IntegrationRuntimeMonitoringData(Model): + """Get monitoring data response. + + :param name: Integration runtime name. + :type name: str + :param nodes: Integration runtime node monitoring data. + :type nodes: + list[~azure.mgmt.datafactory.models.IntegrationRuntimeNodeMonitoringData] + """ + + _attribute_map = { + 'name': {'key': 'name', 'type': 'str'}, + 'nodes': {'key': 'nodes', 'type': '[IntegrationRuntimeNodeMonitoringData]'}, + } + + def __init__(self, *, name: str=None, nodes=None, **kwargs) -> None: + super(IntegrationRuntimeMonitoringData, self).__init__(**kwargs) + self.name = name + self.nodes = nodes diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_node_ip_address.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_node_ip_address.py index a260924f1f16..2edabd3e2472 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_node_ip_address.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_node_ip_address.py @@ -30,6 +30,6 @@ class IntegrationRuntimeNodeIpAddress(Model): 'ip_address': {'key': 'ipAddress', 'type': 'str'}, } - def __init__(self): - super(IntegrationRuntimeNodeIpAddress, self).__init__() + def __init__(self, **kwargs): + super(IntegrationRuntimeNodeIpAddress, self).__init__(**kwargs) self.ip_address = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_node_ip_address_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_node_ip_address_py3.py new file mode 100644 index 000000000000..476be9815984 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_node_ip_address_py3.py @@ -0,0 +1,35 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class IntegrationRuntimeNodeIpAddress(Model): + """The IP address of self-hosted integration runtime node. + + Variables are only populated by the server, and will be ignored when + sending a request. + + :ivar ip_address: The IP address of self-hosted integration runtime node. + :vartype ip_address: str + """ + + _validation = { + 'ip_address': {'readonly': True}, + } + + _attribute_map = { + 'ip_address': {'key': 'ipAddress', 'type': 'str'}, + } + + def __init__(self, **kwargs) -> None: + super(IntegrationRuntimeNodeIpAddress, self).__init__(**kwargs) + self.ip_address = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_node_monitoring_data.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_node_monitoring_data.py index 16c7ffba4b96..93dc31785912 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_node_monitoring_data.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_node_monitoring_data.py @@ -66,9 +66,9 @@ class IntegrationRuntimeNodeMonitoringData(Model): 'received_bytes': {'key': 'receivedBytes', 'type': 'float'}, } - def __init__(self, additional_properties=None): - super(IntegrationRuntimeNodeMonitoringData, self).__init__() - self.additional_properties = additional_properties + def __init__(self, **kwargs): + super(IntegrationRuntimeNodeMonitoringData, self).__init__(**kwargs) + self.additional_properties = kwargs.get('additional_properties', None) self.node_name = None self.available_memory_in_mb = None self.cpu_utilization = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_node_monitoring_data_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_node_monitoring_data_py3.py new file mode 100644 index 000000000000..8c6482817e46 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_node_monitoring_data_py3.py @@ -0,0 +1,79 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class IntegrationRuntimeNodeMonitoringData(Model): + """Monitoring data for integration runtime node. + + Variables are only populated by the server, and will be ignored when + sending a request. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :ivar node_name: Name of the integration runtime node. + :vartype node_name: str + :ivar available_memory_in_mb: Available memory (MB) on the integration + runtime node. + :vartype available_memory_in_mb: int + :ivar cpu_utilization: CPU percentage on the integration runtime node. + :vartype cpu_utilization: float + :ivar concurrent_jobs_limit: Maximum concurrent jobs on the integration + runtime node. + :vartype concurrent_jobs_limit: int + :ivar concurrent_jobs_running: The number of jobs currently running on the + integration runtime node. + :vartype concurrent_jobs_running: int + :ivar max_concurrent_jobs: The maximum concurrent jobs in this integration + runtime. + :vartype max_concurrent_jobs: int + :ivar sent_bytes: Sent bytes on the integration runtime node. + :vartype sent_bytes: float + :ivar received_bytes: Received bytes on the integration runtime node. + :vartype received_bytes: float + """ + + _validation = { + 'node_name': {'readonly': True}, + 'available_memory_in_mb': {'readonly': True}, + 'cpu_utilization': {'readonly': True}, + 'concurrent_jobs_limit': {'readonly': True}, + 'concurrent_jobs_running': {'readonly': True}, + 'max_concurrent_jobs': {'readonly': True}, + 'sent_bytes': {'readonly': True}, + 'received_bytes': {'readonly': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'node_name': {'key': 'nodeName', 'type': 'str'}, + 'available_memory_in_mb': {'key': 'availableMemoryInMB', 'type': 'int'}, + 'cpu_utilization': {'key': 'cpuUtilization', 'type': 'float'}, + 'concurrent_jobs_limit': {'key': 'concurrentJobsLimit', 'type': 'int'}, + 'concurrent_jobs_running': {'key': 'concurrentJobsRunning', 'type': 'int'}, + 'max_concurrent_jobs': {'key': 'maxConcurrentJobs', 'type': 'int'}, + 'sent_bytes': {'key': 'sentBytes', 'type': 'float'}, + 'received_bytes': {'key': 'receivedBytes', 'type': 'float'}, + } + + def __init__(self, *, additional_properties=None, **kwargs) -> None: + super(IntegrationRuntimeNodeMonitoringData, self).__init__(**kwargs) + self.additional_properties = additional_properties + self.node_name = None + self.available_memory_in_mb = None + self.cpu_utilization = None + self.concurrent_jobs_limit = None + self.concurrent_jobs_running = None + self.max_concurrent_jobs = None + self.sent_bytes = None + self.received_bytes = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_py3.py new file mode 100644 index 000000000000..b4056a07591b --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_py3.py @@ -0,0 +1,51 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class IntegrationRuntime(Model): + """Azure Data Factory nested object which serves as a compute resource for + activities. + + You probably want to use the sub-classes and not this class directly. Known + sub-classes are: SelfHostedIntegrationRuntime, ManagedIntegrationRuntime + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param description: Integration runtime description. + :type description: str + :param type: Required. Constant filled by server. + :type type: str + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + _subtype_map = { + 'type': {'SelfHosted': 'SelfHostedIntegrationRuntime', 'Managed': 'ManagedIntegrationRuntime'} + } + + def __init__(self, *, additional_properties=None, description: str=None, **kwargs) -> None: + super(IntegrationRuntime, self).__init__(**kwargs) + self.additional_properties = additional_properties + self.description = description + self.type = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_reference.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_reference.py index 507b578a2cd8..7461d29de284 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_reference.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_reference.py @@ -18,10 +18,12 @@ class IntegrationRuntimeReference(Model): Variables are only populated by the server, and will be ignored when sending a request. - :ivar type: Type of integration runtime. Default value: + All required parameters must be populated in order to send to Azure. + + :ivar type: Required. Type of integration runtime. Default value: "IntegrationRuntimeReference" . :vartype type: str - :param reference_name: Reference integration runtime name. + :param reference_name: Required. Reference integration runtime name. :type reference_name: str :param parameters: Arguments for integration runtime. :type parameters: dict[str, object] @@ -40,7 +42,7 @@ class IntegrationRuntimeReference(Model): type = "IntegrationRuntimeReference" - def __init__(self, reference_name, parameters=None): - super(IntegrationRuntimeReference, self).__init__() - self.reference_name = reference_name - self.parameters = parameters + def __init__(self, **kwargs): + super(IntegrationRuntimeReference, self).__init__(**kwargs) + self.reference_name = kwargs.get('reference_name', None) + self.parameters = kwargs.get('parameters', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_reference_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_reference_py3.py new file mode 100644 index 000000000000..56fd3608ba61 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_reference_py3.py @@ -0,0 +1,48 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class IntegrationRuntimeReference(Model): + """Integration runtime reference type. + + Variables are only populated by the server, and will be ignored when + sending a request. + + All required parameters must be populated in order to send to Azure. + + :ivar type: Required. Type of integration runtime. Default value: + "IntegrationRuntimeReference" . + :vartype type: str + :param reference_name: Required. Reference integration runtime name. + :type reference_name: str + :param parameters: Arguments for integration runtime. + :type parameters: dict[str, object] + """ + + _validation = { + 'type': {'required': True, 'constant': True}, + 'reference_name': {'required': True}, + } + + _attribute_map = { + 'type': {'key': 'type', 'type': 'str'}, + 'reference_name': {'key': 'referenceName', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{object}'}, + } + + type = "IntegrationRuntimeReference" + + def __init__(self, *, reference_name: str, parameters=None, **kwargs) -> None: + super(IntegrationRuntimeReference, self).__init__(**kwargs) + self.reference_name = reference_name + self.parameters = parameters diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_regenerate_key_parameters.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_regenerate_key_parameters.py index f8b4a57d8ff0..3cd91195af1b 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_regenerate_key_parameters.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_regenerate_key_parameters.py @@ -25,6 +25,6 @@ class IntegrationRuntimeRegenerateKeyParameters(Model): 'key_name': {'key': 'keyName', 'type': 'str'}, } - def __init__(self, key_name=None): - super(IntegrationRuntimeRegenerateKeyParameters, self).__init__() - self.key_name = key_name + def __init__(self, **kwargs): + super(IntegrationRuntimeRegenerateKeyParameters, self).__init__(**kwargs) + self.key_name = kwargs.get('key_name', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_regenerate_key_parameters_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_regenerate_key_parameters_py3.py new file mode 100644 index 000000000000..f3846cf8ec55 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_regenerate_key_parameters_py3.py @@ -0,0 +1,30 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class IntegrationRuntimeRegenerateKeyParameters(Model): + """Parameters to regenerate the authentication key. + + :param key_name: The name of the authentication key to regenerate. + Possible values include: 'authKey1', 'authKey2' + :type key_name: str or + ~azure.mgmt.datafactory.models.IntegrationRuntimeAuthKeyName + """ + + _attribute_map = { + 'key_name': {'key': 'keyName', 'type': 'str'}, + } + + def __init__(self, *, key_name=None, **kwargs) -> None: + super(IntegrationRuntimeRegenerateKeyParameters, self).__init__(**kwargs) + self.key_name = key_name diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_remove_node_request.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_remove_node_request.py index afbd29f0424c..9080f93ffbdd 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_remove_node_request.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_remove_node_request.py @@ -27,7 +27,7 @@ class IntegrationRuntimeRemoveNodeRequest(Model): 'node_name': {'key': 'nodeName', 'type': 'str'}, } - def __init__(self, additional_properties=None, node_name=None): - super(IntegrationRuntimeRemoveNodeRequest, self).__init__() - self.additional_properties = additional_properties - self.node_name = node_name + def __init__(self, **kwargs): + super(IntegrationRuntimeRemoveNodeRequest, self).__init__(**kwargs) + self.additional_properties = kwargs.get('additional_properties', None) + self.node_name = kwargs.get('node_name', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_remove_node_request_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_remove_node_request_py3.py new file mode 100644 index 000000000000..57453c633fc5 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_remove_node_request_py3.py @@ -0,0 +1,33 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class IntegrationRuntimeRemoveNodeRequest(Model): + """Request to remove a node. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param node_name: The name of the node to be removed. + :type node_name: str + """ + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'node_name': {'key': 'nodeName', 'type': 'str'}, + } + + def __init__(self, *, additional_properties=None, node_name: str=None, **kwargs) -> None: + super(IntegrationRuntimeRemoveNodeRequest, self).__init__(**kwargs) + self.additional_properties = additional_properties + self.node_name = node_name diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_resource.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_resource.py index 8568ed26cb1e..b18f376d3698 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_resource.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_resource.py @@ -18,6 +18,8 @@ class IntegrationRuntimeResource(SubResource): Variables are only populated by the server, and will be ignored when sending a request. + All required parameters must be populated in order to send to Azure. + :ivar id: The resource identifier. :vartype id: str :ivar name: The resource name. @@ -26,7 +28,7 @@ class IntegrationRuntimeResource(SubResource): :vartype type: str :ivar etag: Etag identifies change in the resource. :vartype etag: str - :param properties: Integration runtime properties. + :param properties: Required. Integration runtime properties. :type properties: ~azure.mgmt.datafactory.models.IntegrationRuntime """ @@ -46,6 +48,6 @@ class IntegrationRuntimeResource(SubResource): 'properties': {'key': 'properties', 'type': 'IntegrationRuntime'}, } - def __init__(self, properties): - super(IntegrationRuntimeResource, self).__init__() - self.properties = properties + def __init__(self, **kwargs): + super(IntegrationRuntimeResource, self).__init__(**kwargs) + self.properties = kwargs.get('properties', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_resource_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_resource_py3.py new file mode 100644 index 000000000000..acf162de2562 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_resource_py3.py @@ -0,0 +1,53 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .sub_resource import SubResource + + +class IntegrationRuntimeResource(SubResource): + """Integration runtime resource type. + + Variables are only populated by the server, and will be ignored when + sending a request. + + All required parameters must be populated in order to send to Azure. + + :ivar id: The resource identifier. + :vartype id: str + :ivar name: The resource name. + :vartype name: str + :ivar type: The resource type. + :vartype type: str + :ivar etag: Etag identifies change in the resource. + :vartype etag: str + :param properties: Required. Integration runtime properties. + :type properties: ~azure.mgmt.datafactory.models.IntegrationRuntime + """ + + _validation = { + 'id': {'readonly': True}, + 'name': {'readonly': True}, + 'type': {'readonly': True}, + 'etag': {'readonly': True}, + 'properties': {'required': True}, + } + + _attribute_map = { + 'id': {'key': 'id', 'type': 'str'}, + 'name': {'key': 'name', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'str'}, + 'etag': {'key': 'etag', 'type': 'str'}, + 'properties': {'key': 'properties', 'type': 'IntegrationRuntime'}, + } + + def __init__(self, *, properties, **kwargs) -> None: + super(IntegrationRuntimeResource, self).__init__(, **kwargs) + self.properties = properties diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_ssis_catalog_info.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_ssis_catalog_info.py index 0c7e9dc74878..3399f8f38300 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_ssis_catalog_info.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_ssis_catalog_info.py @@ -46,10 +46,10 @@ class IntegrationRuntimeSsisCatalogInfo(Model): 'catalog_pricing_tier': {'key': 'catalogPricingTier', 'type': 'str'}, } - def __init__(self, additional_properties=None, catalog_server_endpoint=None, catalog_admin_user_name=None, catalog_admin_password=None, catalog_pricing_tier=None): - super(IntegrationRuntimeSsisCatalogInfo, self).__init__() - self.additional_properties = additional_properties - self.catalog_server_endpoint = catalog_server_endpoint - self.catalog_admin_user_name = catalog_admin_user_name - self.catalog_admin_password = catalog_admin_password - self.catalog_pricing_tier = catalog_pricing_tier + def __init__(self, **kwargs): + super(IntegrationRuntimeSsisCatalogInfo, self).__init__(**kwargs) + self.additional_properties = kwargs.get('additional_properties', None) + self.catalog_server_endpoint = kwargs.get('catalog_server_endpoint', None) + self.catalog_admin_user_name = kwargs.get('catalog_admin_user_name', None) + self.catalog_admin_password = kwargs.get('catalog_admin_password', None) + self.catalog_pricing_tier = kwargs.get('catalog_pricing_tier', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_ssis_catalog_info_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_ssis_catalog_info_py3.py new file mode 100644 index 000000000000..27996bb4aeb5 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_ssis_catalog_info_py3.py @@ -0,0 +1,55 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class IntegrationRuntimeSsisCatalogInfo(Model): + """Catalog information for managed dedicated integration runtime. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param catalog_server_endpoint: The catalog database server URL. + :type catalog_server_endpoint: str + :param catalog_admin_user_name: The administrator user name of catalog + database. + :type catalog_admin_user_name: str + :param catalog_admin_password: The password of the administrator user + account of the catalog database. + :type catalog_admin_password: ~azure.mgmt.datafactory.models.SecureString + :param catalog_pricing_tier: The pricing tier for the catalog database. + The valid values could be found in + https://azure.microsoft.com/en-us/pricing/details/sql-database/. Possible + values include: 'Basic', 'Standard', 'Premium', 'PremiumRS' + :type catalog_pricing_tier: str or + ~azure.mgmt.datafactory.models.IntegrationRuntimeSsisCatalogPricingTier + """ + + _validation = { + 'catalog_admin_user_name': {'max_length': 128, 'min_length': 1}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'catalog_server_endpoint': {'key': 'catalogServerEndpoint', 'type': 'str'}, + 'catalog_admin_user_name': {'key': 'catalogAdminUserName', 'type': 'str'}, + 'catalog_admin_password': {'key': 'catalogAdminPassword', 'type': 'SecureString'}, + 'catalog_pricing_tier': {'key': 'catalogPricingTier', 'type': 'str'}, + } + + def __init__(self, *, additional_properties=None, catalog_server_endpoint: str=None, catalog_admin_user_name: str=None, catalog_admin_password=None, catalog_pricing_tier=None, **kwargs) -> None: + super(IntegrationRuntimeSsisCatalogInfo, self).__init__(**kwargs) + self.additional_properties = additional_properties + self.catalog_server_endpoint = catalog_server_endpoint + self.catalog_admin_user_name = catalog_admin_user_name + self.catalog_admin_password = catalog_admin_password + self.catalog_pricing_tier = catalog_pricing_tier diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_ssis_properties.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_ssis_properties.py index 562e74c78ffe..e1a091166529 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_ssis_properties.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_ssis_properties.py @@ -44,10 +44,10 @@ class IntegrationRuntimeSsisProperties(Model): 'edition': {'key': 'edition', 'type': 'str'}, } - def __init__(self, additional_properties=None, catalog_info=None, license_type=None, custom_setup_script_properties=None, edition=None): - super(IntegrationRuntimeSsisProperties, self).__init__() - self.additional_properties = additional_properties - self.catalog_info = catalog_info - self.license_type = license_type - self.custom_setup_script_properties = custom_setup_script_properties - self.edition = edition + def __init__(self, **kwargs): + super(IntegrationRuntimeSsisProperties, self).__init__(**kwargs) + self.additional_properties = kwargs.get('additional_properties', None) + self.catalog_info = kwargs.get('catalog_info', None) + self.license_type = kwargs.get('license_type', None) + self.custom_setup_script_properties = kwargs.get('custom_setup_script_properties', None) + self.edition = kwargs.get('edition', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_ssis_properties_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_ssis_properties_py3.py new file mode 100644 index 000000000000..eb70dd23ddb7 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_ssis_properties_py3.py @@ -0,0 +1,53 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class IntegrationRuntimeSsisProperties(Model): + """SSIS properties for managed integration runtime. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param catalog_info: Catalog information for managed dedicated integration + runtime. + :type catalog_info: + ~azure.mgmt.datafactory.models.IntegrationRuntimeSsisCatalogInfo + :param license_type: License type for bringing your own license scenario. + Possible values include: 'BasePrice', 'LicenseIncluded' + :type license_type: str or + ~azure.mgmt.datafactory.models.IntegrationRuntimeLicenseType + :param custom_setup_script_properties: Custom setup script properties for + a managed dedicated integration runtime. + :type custom_setup_script_properties: + ~azure.mgmt.datafactory.models.IntegrationRuntimeCustomSetupScriptProperties + :param edition: The edition for the SSIS Integration Runtime. Possible + values include: 'Standard', 'Enterprise' + :type edition: str or + ~azure.mgmt.datafactory.models.IntegrationRuntimeEdition + """ + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'catalog_info': {'key': 'catalogInfo', 'type': 'IntegrationRuntimeSsisCatalogInfo'}, + 'license_type': {'key': 'licenseType', 'type': 'str'}, + 'custom_setup_script_properties': {'key': 'customSetupScriptProperties', 'type': 'IntegrationRuntimeCustomSetupScriptProperties'}, + 'edition': {'key': 'edition', 'type': 'str'}, + } + + def __init__(self, *, additional_properties=None, catalog_info=None, license_type=None, custom_setup_script_properties=None, edition=None, **kwargs) -> None: + super(IntegrationRuntimeSsisProperties, self).__init__(**kwargs) + self.additional_properties = additional_properties + self.catalog_info = catalog_info + self.license_type = license_type + self.custom_setup_script_properties = custom_setup_script_properties + self.edition = edition diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_status.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_status.py index e6f7010a0bdf..987fd8c7e0c4 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_status.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_status.py @@ -22,6 +22,8 @@ class IntegrationRuntimeStatus(Model): Variables are only populated by the server, and will be ignored when sending a request. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -33,7 +35,7 @@ class IntegrationRuntimeStatus(Model): 'NeedRegistration', 'Online', 'Limited', 'Offline' :vartype state: str or ~azure.mgmt.datafactory.models.IntegrationRuntimeState - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str """ @@ -54,9 +56,9 @@ class IntegrationRuntimeStatus(Model): 'type': {'SelfHosted': 'SelfHostedIntegrationRuntimeStatus', 'Managed': 'ManagedIntegrationRuntimeStatus'} } - def __init__(self, additional_properties=None): - super(IntegrationRuntimeStatus, self).__init__() - self.additional_properties = additional_properties + def __init__(self, **kwargs): + super(IntegrationRuntimeStatus, self).__init__(**kwargs) + self.additional_properties = kwargs.get('additional_properties', None) self.data_factory_name = None self.state = None self.type = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_status_list_response.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_status_list_response.py index 23d5c95fcd28..9382b4b08fde 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_status_list_response.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_status_list_response.py @@ -15,7 +15,9 @@ class IntegrationRuntimeStatusListResponse(Model): """A list of integration runtime status. - :param value: List of integration runtime status. + All required parameters must be populated in order to send to Azure. + + :param value: Required. List of integration runtime status. :type value: list[~azure.mgmt.datafactory.models.IntegrationRuntimeStatusResponse] :param next_link: The link to the next page of results, if any remaining @@ -32,7 +34,7 @@ class IntegrationRuntimeStatusListResponse(Model): 'next_link': {'key': 'nextLink', 'type': 'str'}, } - def __init__(self, value, next_link=None): - super(IntegrationRuntimeStatusListResponse, self).__init__() - self.value = value - self.next_link = next_link + def __init__(self, **kwargs): + super(IntegrationRuntimeStatusListResponse, self).__init__(**kwargs) + self.value = kwargs.get('value', None) + self.next_link = kwargs.get('next_link', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_status_list_response_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_status_list_response_py3.py new file mode 100644 index 000000000000..bed71f74ffc6 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_status_list_response_py3.py @@ -0,0 +1,40 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class IntegrationRuntimeStatusListResponse(Model): + """A list of integration runtime status. + + All required parameters must be populated in order to send to Azure. + + :param value: Required. List of integration runtime status. + :type value: + list[~azure.mgmt.datafactory.models.IntegrationRuntimeStatusResponse] + :param next_link: The link to the next page of results, if any remaining + results exist. + :type next_link: str + """ + + _validation = { + 'value': {'required': True}, + } + + _attribute_map = { + 'value': {'key': 'value', 'type': '[IntegrationRuntimeStatusResponse]'}, + 'next_link': {'key': 'nextLink', 'type': 'str'}, + } + + def __init__(self, *, value, next_link: str=None, **kwargs) -> None: + super(IntegrationRuntimeStatusListResponse, self).__init__(**kwargs) + self.value = value + self.next_link = next_link diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_status_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_status_py3.py new file mode 100644 index 000000000000..0f41e6f73f72 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_status_py3.py @@ -0,0 +1,64 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class IntegrationRuntimeStatus(Model): + """Integration runtime status. + + You probably want to use the sub-classes and not this class directly. Known + sub-classes are: SelfHostedIntegrationRuntimeStatus, + ManagedIntegrationRuntimeStatus + + Variables are only populated by the server, and will be ignored when + sending a request. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :ivar data_factory_name: The data factory name which the integration + runtime belong to. + :vartype data_factory_name: str + :ivar state: The state of integration runtime. Possible values include: + 'Initial', 'Stopped', 'Started', 'Starting', 'Stopping', + 'NeedRegistration', 'Online', 'Limited', 'Offline' + :vartype state: str or + ~azure.mgmt.datafactory.models.IntegrationRuntimeState + :param type: Required. Constant filled by server. + :type type: str + """ + + _validation = { + 'data_factory_name': {'readonly': True}, + 'state': {'readonly': True}, + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'data_factory_name': {'key': 'dataFactoryName', 'type': 'str'}, + 'state': {'key': 'state', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + _subtype_map = { + 'type': {'SelfHosted': 'SelfHostedIntegrationRuntimeStatus', 'Managed': 'ManagedIntegrationRuntimeStatus'} + } + + def __init__(self, *, additional_properties=None, **kwargs) -> None: + super(IntegrationRuntimeStatus, self).__init__(**kwargs) + self.additional_properties = additional_properties + self.data_factory_name = None + self.state = None + self.type = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_status_response.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_status_response.py index 89b045642459..901b4d8b7442 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_status_response.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_status_response.py @@ -18,9 +18,11 @@ class IntegrationRuntimeStatusResponse(Model): Variables are only populated by the server, and will be ignored when sending a request. + All required parameters must be populated in order to send to Azure. + :ivar name: The integration runtime name. :vartype name: str - :param properties: Integration runtime properties. + :param properties: Required. Integration runtime properties. :type properties: ~azure.mgmt.datafactory.models.IntegrationRuntimeStatus """ @@ -34,7 +36,7 @@ class IntegrationRuntimeStatusResponse(Model): 'properties': {'key': 'properties', 'type': 'IntegrationRuntimeStatus'}, } - def __init__(self, properties): - super(IntegrationRuntimeStatusResponse, self).__init__() + def __init__(self, **kwargs): + super(IntegrationRuntimeStatusResponse, self).__init__(**kwargs) self.name = None - self.properties = properties + self.properties = kwargs.get('properties', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_status_response_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_status_response_py3.py new file mode 100644 index 000000000000..64d84a1e4f19 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_status_response_py3.py @@ -0,0 +1,42 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class IntegrationRuntimeStatusResponse(Model): + """Integration runtime status response. + + Variables are only populated by the server, and will be ignored when + sending a request. + + All required parameters must be populated in order to send to Azure. + + :ivar name: The integration runtime name. + :vartype name: str + :param properties: Required. Integration runtime properties. + :type properties: ~azure.mgmt.datafactory.models.IntegrationRuntimeStatus + """ + + _validation = { + 'name': {'readonly': True}, + 'properties': {'required': True}, + } + + _attribute_map = { + 'name': {'key': 'name', 'type': 'str'}, + 'properties': {'key': 'properties', 'type': 'IntegrationRuntimeStatus'}, + } + + def __init__(self, *, properties, **kwargs) -> None: + super(IntegrationRuntimeStatusResponse, self).__init__(**kwargs) + self.name = None + self.properties = properties diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_vnet_properties.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_vnet_properties.py index 702723a2f067..752b5b99eb60 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_vnet_properties.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_vnet_properties.py @@ -31,8 +31,8 @@ class IntegrationRuntimeVNetProperties(Model): 'subnet': {'key': 'subnet', 'type': 'str'}, } - def __init__(self, additional_properties=None, v_net_id=None, subnet=None): - super(IntegrationRuntimeVNetProperties, self).__init__() - self.additional_properties = additional_properties - self.v_net_id = v_net_id - self.subnet = subnet + def __init__(self, **kwargs): + super(IntegrationRuntimeVNetProperties, self).__init__(**kwargs) + self.additional_properties = kwargs.get('additional_properties', None) + self.v_net_id = kwargs.get('v_net_id', None) + self.subnet = kwargs.get('subnet', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_vnet_properties_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_vnet_properties_py3.py new file mode 100644 index 000000000000..32e8beb31ea1 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_vnet_properties_py3.py @@ -0,0 +1,38 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class IntegrationRuntimeVNetProperties(Model): + """VNet properties for managed integration runtime. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param v_net_id: The ID of the VNet that this integration runtime will + join. + :type v_net_id: str + :param subnet: The name of the subnet this integration runtime will join. + :type subnet: str + """ + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'v_net_id': {'key': 'vNetId', 'type': 'str'}, + 'subnet': {'key': 'subnet', 'type': 'str'}, + } + + def __init__(self, *, additional_properties=None, v_net_id: str=None, subnet: str=None, **kwargs) -> None: + super(IntegrationRuntimeVNetProperties, self).__init__(**kwargs) + self.additional_properties = additional_properties + self.v_net_id = v_net_id + self.subnet = subnet diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_linked_service.py index f42602398af7..4ea04a954079 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_linked_service.py @@ -15,6 +15,8 @@ class JiraLinkedService(LinkedService): """Jira Serivce linked service. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -29,16 +31,17 @@ class JiraLinkedService(LinkedService): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param host: The IP address or host name of the Jira service. (e.g. - jira.example.com) + :param host: Required. The IP address or host name of the Jira service. + (e.g. jira.example.com) :type host: object :param port: The TCP port that the Jira server uses to listen for client connections. The default value is 443 if connecting through HTTPS, or 8080 if connecting through HTTP. :type port: object - :param username: The user name that you use to access Jira Service. + :param username: Required. The user name that you use to access Jira + Service. :type username: object :param password: The password corresponding to the user name that you provided in the username field. @@ -82,14 +85,14 @@ class JiraLinkedService(LinkedService): 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, host, username, additional_properties=None, connect_via=None, description=None, parameters=None, annotations=None, port=None, password=None, use_encrypted_endpoints=None, use_host_verification=None, use_peer_verification=None, encrypted_credential=None): - super(JiraLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations) - self.host = host - self.port = port - self.username = username - self.password = password - self.use_encrypted_endpoints = use_encrypted_endpoints - self.use_host_verification = use_host_verification - self.use_peer_verification = use_peer_verification - self.encrypted_credential = encrypted_credential + def __init__(self, **kwargs): + super(JiraLinkedService, self).__init__(**kwargs) + self.host = kwargs.get('host', None) + self.port = kwargs.get('port', None) + self.username = kwargs.get('username', None) + self.password = kwargs.get('password', None) + self.use_encrypted_endpoints = kwargs.get('use_encrypted_endpoints', None) + self.use_host_verification = kwargs.get('use_host_verification', None) + self.use_peer_verification = kwargs.get('use_peer_verification', None) + self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'Jira' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_linked_service_py3.py new file mode 100644 index 000000000000..824a30ea1b2b --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_linked_service_py3.py @@ -0,0 +1,98 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class JiraLinkedService(LinkedService): + """Jira Serivce linked service. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param host: Required. The IP address or host name of the Jira service. + (e.g. jira.example.com) + :type host: object + :param port: The TCP port that the Jira server uses to listen for client + connections. The default value is 443 if connecting through HTTPS, or 8080 + if connecting through HTTP. + :type port: object + :param username: Required. The user name that you use to access Jira + Service. + :type username: object + :param password: The password corresponding to the user name that you + provided in the username field. + :type password: ~azure.mgmt.datafactory.models.SecretBase + :param use_encrypted_endpoints: Specifies whether the data source + endpoints are encrypted using HTTPS. The default value is true. + :type use_encrypted_endpoints: object + :param use_host_verification: Specifies whether to require the host name + in the server's certificate to match the host name of the server when + connecting over SSL. The default value is true. + :type use_host_verification: object + :param use_peer_verification: Specifies whether to verify the identity of + the server when connecting over SSL. The default value is true. + :type use_peer_verification: object + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'host': {'required': True}, + 'username': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'host': {'key': 'typeProperties.host', 'type': 'object'}, + 'port': {'key': 'typeProperties.port', 'type': 'object'}, + 'username': {'key': 'typeProperties.username', 'type': 'object'}, + 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, + 'use_encrypted_endpoints': {'key': 'typeProperties.useEncryptedEndpoints', 'type': 'object'}, + 'use_host_verification': {'key': 'typeProperties.useHostVerification', 'type': 'object'}, + 'use_peer_verification': {'key': 'typeProperties.usePeerVerification', 'type': 'object'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, *, host, username, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, port=None, password=None, use_encrypted_endpoints=None, use_host_verification=None, use_peer_verification=None, encrypted_credential=None, **kwargs) -> None: + super(JiraLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.host = host + self.port = port + self.username = username + self.password = password + self.use_encrypted_endpoints = use_encrypted_endpoints + self.use_host_verification = use_host_verification + self.use_peer_verification = use_peer_verification + self.encrypted_credential = encrypted_credential + self.type = 'Jira' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_object_dataset.py index dcdc5b4f0772..d777d905f3b0 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_object_dataset.py @@ -15,6 +15,8 @@ class JiraObjectDataset(Dataset): """Jira Serivce dataset. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -23,7 +25,7 @@ class JiraObjectDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object - :param linked_service_name: Linked service reference. + :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference :param parameters: Parameters for dataset. @@ -32,7 +34,7 @@ class JiraObjectDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str """ @@ -41,6 +43,16 @@ class JiraObjectDataset(Dataset): 'type': {'required': True}, } - def __init__(self, linked_service_name, additional_properties=None, description=None, structure=None, parameters=None, annotations=None): - super(JiraObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations) + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, **kwargs): + super(JiraObjectDataset, self).__init__(**kwargs) self.type = 'JiraObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_object_dataset_py3.py new file mode 100644 index 000000000000..d1f5a34f516d --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_object_dataset_py3.py @@ -0,0 +1,58 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class JiraObjectDataset(Dataset): + """Jira Serivce dataset. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Required. Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: + super(JiraObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + self.type = 'JiraObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_source.py index d740e27f3b5b..993d5ecb1d11 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_source.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_source.py @@ -15,6 +15,8 @@ class JiraSource(CopySource): """A copy activity Jira Serivce source. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -25,7 +27,7 @@ class JiraSource(CopySource): with resultType string), pattern: ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type source_retry_wait: object - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). @@ -44,7 +46,7 @@ class JiraSource(CopySource): 'query': {'key': 'query', 'type': 'object'}, } - def __init__(self, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None): - super(JiraSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait) - self.query = query + def __init__(self, **kwargs): + super(JiraSource, self).__init__(**kwargs) + self.query = kwargs.get('query', None) self.type = 'JiraSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_source_py3.py new file mode 100644 index 000000000000..80e3e50a062d --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_source_py3.py @@ -0,0 +1,52 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class JiraSource(CopySource): + """A copy activity Jira Serivce source. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Required. Constant filled by server. + :type type: str + :param query: A query to retrieve data from source. Type: string (or + Expression with resultType string). + :type query: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'query': {'key': 'query', 'type': 'object'}, + } + + def __init__(self, *, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None, **kwargs) -> None: + super(JiraSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, **kwargs) + self.query = query + self.type = 'JiraSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/json_format.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/json_format.py index d300c1b200e9..736f9500018f 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/json_format.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/json_format.py @@ -15,6 +15,8 @@ class JsonFormat(DatasetStorageFormat): """The data stored in JSON format. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -24,7 +26,7 @@ class JsonFormat(DatasetStorageFormat): :param deserializer: Deserializer. Type: string (or Expression with resultType string). :type deserializer: object - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param file_pattern: File pattern of JSON. To be more specific, the way of separating a collection of JSON objects. The default value is @@ -72,11 +74,11 @@ class JsonFormat(DatasetStorageFormat): 'json_path_definition': {'key': 'jsonPathDefinition', 'type': 'object'}, } - def __init__(self, additional_properties=None, serializer=None, deserializer=None, file_pattern=None, nesting_separator=None, encoding_name=None, json_node_reference=None, json_path_definition=None): - super(JsonFormat, self).__init__(additional_properties=additional_properties, serializer=serializer, deserializer=deserializer) - self.file_pattern = file_pattern - self.nesting_separator = nesting_separator - self.encoding_name = encoding_name - self.json_node_reference = json_node_reference - self.json_path_definition = json_path_definition + def __init__(self, **kwargs): + super(JsonFormat, self).__init__(**kwargs) + self.file_pattern = kwargs.get('file_pattern', None) + self.nesting_separator = kwargs.get('nesting_separator', None) + self.encoding_name = kwargs.get('encoding_name', None) + self.json_node_reference = kwargs.get('json_node_reference', None) + self.json_path_definition = kwargs.get('json_path_definition', None) self.type = 'JsonFormat' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/json_format_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/json_format_py3.py new file mode 100644 index 000000000000..5213a17d8e4c --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/json_format_py3.py @@ -0,0 +1,84 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset_storage_format import DatasetStorageFormat + + +class JsonFormat(DatasetStorageFormat): + """The data stored in JSON format. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param serializer: Serializer. Type: string (or Expression with resultType + string). + :type serializer: object + :param deserializer: Deserializer. Type: string (or Expression with + resultType string). + :type deserializer: object + :param type: Required. Constant filled by server. + :type type: str + :param file_pattern: File pattern of JSON. To be more specific, the way of + separating a collection of JSON objects. The default value is + 'setOfObjects'. It is case-sensitive. Possible values include: + 'setOfObjects', 'arrayOfObjects' + :type file_pattern: str or + ~azure.mgmt.datafactory.models.JsonFormatFilePattern + :param nesting_separator: The character used to separate nesting levels. + Default value is '.' (dot). Type: string (or Expression with resultType + string). + :type nesting_separator: object + :param encoding_name: The code page name of the preferred encoding. If not + provided, the default value is 'utf-8', unless the byte order mark (BOM) + denotes another Unicode encoding. The full list of supported values can be + found in the 'Name' column of the table of encodings in the following + reference: https://go.microsoft.com/fwlink/?linkid=861078. Type: string + (or Expression with resultType string). + :type encoding_name: object + :param json_node_reference: The JSONPath of the JSON array element to be + flattened. Example: "$.ArrayPath". Type: string (or Expression with + resultType string). + :type json_node_reference: object + :param json_path_definition: The JSONPath definition for each column + mapping with a customized column name to extract data from JSON file. For + fields under root object, start with "$"; for fields inside the array + chosen by jsonNodeReference property, start from the array element. + Example: {"Column1": "$.Column1Path", "Column2": "Column2PathInArray"}. + Type: object (or Expression with resultType object). + :type json_path_definition: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'serializer': {'key': 'serializer', 'type': 'object'}, + 'deserializer': {'key': 'deserializer', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'file_pattern': {'key': 'filePattern', 'type': 'str'}, + 'nesting_separator': {'key': 'nestingSeparator', 'type': 'object'}, + 'encoding_name': {'key': 'encodingName', 'type': 'object'}, + 'json_node_reference': {'key': 'jsonNodeReference', 'type': 'object'}, + 'json_path_definition': {'key': 'jsonPathDefinition', 'type': 'object'}, + } + + def __init__(self, *, additional_properties=None, serializer=None, deserializer=None, file_pattern=None, nesting_separator=None, encoding_name=None, json_node_reference=None, json_path_definition=None, **kwargs) -> None: + super(JsonFormat, self).__init__(additional_properties=additional_properties, serializer=serializer, deserializer=deserializer, **kwargs) + self.file_pattern = file_pattern + self.nesting_separator = nesting_separator + self.encoding_name = encoding_name + self.json_node_reference = json_node_reference + self.json_path_definition = json_path_definition + self.type = 'JsonFormat' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime.py index 1a733032b07e..f4a4e7eb8bf0 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime.py @@ -49,8 +49,8 @@ class LinkedIntegrationRuntime(Model): 'create_time': {'key': 'createTime', 'type': 'iso-8601'}, } - def __init__(self): - super(LinkedIntegrationRuntime, self).__init__() + def __init__(self, **kwargs): + super(LinkedIntegrationRuntime, self).__init__(**kwargs) self.name = None self.subscription_id = None self.data_factory_name = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_key.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_key.py index 6c2f9dd80428..cbecfb9afae4 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_key.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_key.py @@ -15,9 +15,11 @@ class LinkedIntegrationRuntimeKey(LinkedIntegrationRuntimeProperties): """The base definition of a secret type. - :param authorization_type: Constant filled by server. + All required parameters must be populated in order to send to Azure. + + :param authorization_type: Required. Constant filled by server. :type authorization_type: str - :param key: Type of the secret. + :param key: Required. Type of the secret. :type key: ~azure.mgmt.datafactory.models.SecureString """ @@ -31,7 +33,7 @@ class LinkedIntegrationRuntimeKey(LinkedIntegrationRuntimeProperties): 'key': {'key': 'key', 'type': 'SecureString'}, } - def __init__(self, key): - super(LinkedIntegrationRuntimeKey, self).__init__() - self.key = key + def __init__(self, **kwargs): + super(LinkedIntegrationRuntimeKey, self).__init__(**kwargs) + self.key = kwargs.get('key', None) self.authorization_type = 'Key' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_key_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_key_py3.py new file mode 100644 index 000000000000..3e8f4dca2051 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_key_py3.py @@ -0,0 +1,39 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_integration_runtime_properties import LinkedIntegrationRuntimeProperties + + +class LinkedIntegrationRuntimeKey(LinkedIntegrationRuntimeProperties): + """The base definition of a secret type. + + All required parameters must be populated in order to send to Azure. + + :param authorization_type: Required. Constant filled by server. + :type authorization_type: str + :param key: Required. Type of the secret. + :type key: ~azure.mgmt.datafactory.models.SecureString + """ + + _validation = { + 'authorization_type': {'required': True}, + 'key': {'required': True}, + } + + _attribute_map = { + 'authorization_type': {'key': 'authorizationType', 'type': 'str'}, + 'key': {'key': 'key', 'type': 'SecureString'}, + } + + def __init__(self, *, key, **kwargs) -> None: + super(LinkedIntegrationRuntimeKey, self).__init__(, **kwargs) + self.key = key + self.authorization_type = 'Key' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_properties.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_properties.py index 4e91a5e9f45a..20a80b23277c 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_properties.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_properties.py @@ -18,7 +18,9 @@ class LinkedIntegrationRuntimeProperties(Model): You probably want to use the sub-classes and not this class directly. Known sub-classes are: LinkedIntegrationRuntimeRbac, LinkedIntegrationRuntimeKey - :param authorization_type: Constant filled by server. + All required parameters must be populated in order to send to Azure. + + :param authorization_type: Required. Constant filled by server. :type authorization_type: str """ @@ -34,6 +36,6 @@ class LinkedIntegrationRuntimeProperties(Model): 'authorization_type': {'RBAC': 'LinkedIntegrationRuntimeRbac', 'Key': 'LinkedIntegrationRuntimeKey'} } - def __init__(self): - super(LinkedIntegrationRuntimeProperties, self).__init__() + def __init__(self, **kwargs): + super(LinkedIntegrationRuntimeProperties, self).__init__(**kwargs) self.authorization_type = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_properties_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_properties_py3.py new file mode 100644 index 000000000000..da73c9b6b95a --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_properties_py3.py @@ -0,0 +1,41 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class LinkedIntegrationRuntimeProperties(Model): + """The base definition of a secret type. + + You probably want to use the sub-classes and not this class directly. Known + sub-classes are: LinkedIntegrationRuntimeRbac, LinkedIntegrationRuntimeKey + + All required parameters must be populated in order to send to Azure. + + :param authorization_type: Required. Constant filled by server. + :type authorization_type: str + """ + + _validation = { + 'authorization_type': {'required': True}, + } + + _attribute_map = { + 'authorization_type': {'key': 'authorizationType', 'type': 'str'}, + } + + _subtype_map = { + 'authorization_type': {'RBAC': 'LinkedIntegrationRuntimeRbac', 'Key': 'LinkedIntegrationRuntimeKey'} + } + + def __init__(self, **kwargs) -> None: + super(LinkedIntegrationRuntimeProperties, self).__init__(**kwargs) + self.authorization_type = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_py3.py new file mode 100644 index 000000000000..6c831ab5f511 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_py3.py @@ -0,0 +1,58 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class LinkedIntegrationRuntime(Model): + """The linked integration runtime information. + + Variables are only populated by the server, and will be ignored when + sending a request. + + :ivar name: The name of the linked integration runtime. + :vartype name: str + :ivar subscription_id: The subscription ID for which the linked + integration runtime belong to. + :vartype subscription_id: str + :ivar data_factory_name: The name of the data factory for which the linked + integration runtime belong to. + :vartype data_factory_name: str + :ivar data_factory_location: The location of the data factory for which + the linked integration runtime belong to. + :vartype data_factory_location: str + :ivar create_time: The creating time of the linked integration runtime. + :vartype create_time: datetime + """ + + _validation = { + 'name': {'readonly': True}, + 'subscription_id': {'readonly': True}, + 'data_factory_name': {'readonly': True}, + 'data_factory_location': {'readonly': True}, + 'create_time': {'readonly': True}, + } + + _attribute_map = { + 'name': {'key': 'name', 'type': 'str'}, + 'subscription_id': {'key': 'subscriptionId', 'type': 'str'}, + 'data_factory_name': {'key': 'dataFactoryName', 'type': 'str'}, + 'data_factory_location': {'key': 'dataFactoryLocation', 'type': 'str'}, + 'create_time': {'key': 'createTime', 'type': 'iso-8601'}, + } + + def __init__(self, **kwargs) -> None: + super(LinkedIntegrationRuntime, self).__init__(**kwargs) + self.name = None + self.subscription_id = None + self.data_factory_name = None + self.data_factory_location = None + self.create_time = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_rbac.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_rbac.py index c38d91d94918..3c7235ecc279 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_rbac.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_rbac.py @@ -15,10 +15,12 @@ class LinkedIntegrationRuntimeRbac(LinkedIntegrationRuntimeProperties): """The base definition of a secret type. - :param authorization_type: Constant filled by server. + All required parameters must be populated in order to send to Azure. + + :param authorization_type: Required. Constant filled by server. :type authorization_type: str - :param resource_id: The resource ID of the integration runtime to be - shared. + :param resource_id: Required. The resource ID of the integration runtime + to be shared. :type resource_id: str """ @@ -32,7 +34,7 @@ class LinkedIntegrationRuntimeRbac(LinkedIntegrationRuntimeProperties): 'resource_id': {'key': 'resourceId', 'type': 'str'}, } - def __init__(self, resource_id): - super(LinkedIntegrationRuntimeRbac, self).__init__() - self.resource_id = resource_id + def __init__(self, **kwargs): + super(LinkedIntegrationRuntimeRbac, self).__init__(**kwargs) + self.resource_id = kwargs.get('resource_id', None) self.authorization_type = 'RBAC' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_rbac_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_rbac_py3.py new file mode 100644 index 000000000000..16c070d71c47 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_rbac_py3.py @@ -0,0 +1,40 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_integration_runtime_properties import LinkedIntegrationRuntimeProperties + + +class LinkedIntegrationRuntimeRbac(LinkedIntegrationRuntimeProperties): + """The base definition of a secret type. + + All required parameters must be populated in order to send to Azure. + + :param authorization_type: Required. Constant filled by server. + :type authorization_type: str + :param resource_id: Required. The resource ID of the integration runtime + to be shared. + :type resource_id: str + """ + + _validation = { + 'authorization_type': {'required': True}, + 'resource_id': {'required': True}, + } + + _attribute_map = { + 'authorization_type': {'key': 'authorizationType', 'type': 'str'}, + 'resource_id': {'key': 'resourceId', 'type': 'str'}, + } + + def __init__(self, *, resource_id: str, **kwargs) -> None: + super(LinkedIntegrationRuntimeRbac, self).__init__(, **kwargs) + self.resource_id = resource_id + self.authorization_type = 'RBAC' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service.py index 308313501773..ecb6b42e145a 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service.py @@ -18,7 +18,7 @@ class LinkedService(Model): resource. You probably want to use the sub-classes and not this class directly. Known - sub-classes are: AzureDatabricksLinkedService, + sub-classes are: ResponsysLinkedService, AzureDatabricksLinkedService, AzureDataLakeAnalyticsLinkedService, HDInsightOnDemandLinkedService, SalesforceMarketingCloudLinkedService, NetezzaLinkedService, VerticaLinkedService, ZohoLinkedService, XeroLinkedService, @@ -45,6 +45,8 @@ class LinkedService(Model): AzureBatchLinkedService, AzureSqlDatabaseLinkedService, SqlServerLinkedService, AzureSqlDWLinkedService, AzureStorageLinkedService + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -59,7 +61,7 @@ class LinkedService(Model): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str """ @@ -77,14 +79,14 @@ class LinkedService(Model): } _subtype_map = { - 'type': {'AzureDatabricks': 'AzureDatabricksLinkedService', 'AzureDataLakeAnalytics': 'AzureDataLakeAnalyticsLinkedService', 'HDInsightOnDemand': 'HDInsightOnDemandLinkedService', 'SalesforceMarketingCloud': 'SalesforceMarketingCloudLinkedService', 'Netezza': 'NetezzaLinkedService', 'Vertica': 'VerticaLinkedService', 'Zoho': 'ZohoLinkedService', 'Xero': 'XeroLinkedService', 'Square': 'SquareLinkedService', 'Spark': 'SparkLinkedService', 'Shopify': 'ShopifyLinkedService', 'ServiceNow': 'ServiceNowLinkedService', 'QuickBooks': 'QuickBooksLinkedService', 'Presto': 'PrestoLinkedService', 'Phoenix': 'PhoenixLinkedService', 'Paypal': 'PaypalLinkedService', 'Marketo': 'MarketoLinkedService', 'MariaDB': 'MariaDBLinkedService', 'Magento': 'MagentoLinkedService', 'Jira': 'JiraLinkedService', 'Impala': 'ImpalaLinkedService', 'Hubspot': 'HubspotLinkedService', 'Hive': 'HiveLinkedService', 'HBase': 'HBaseLinkedService', 'Greenplum': 'GreenplumLinkedService', 'GoogleBigQuery': 'GoogleBigQueryLinkedService', 'Eloqua': 'EloquaLinkedService', 'Drill': 'DrillLinkedService', 'Couchbase': 'CouchbaseLinkedService', 'Concur': 'ConcurLinkedService', 'AzurePostgreSql': 'AzurePostgreSqlLinkedService', 'AmazonMWS': 'AmazonMWSLinkedService', 'SapHana': 'SapHanaLinkedService', 'SapBW': 'SapBWLinkedService', 'Sftp': 'SftpServerLinkedService', 'FtpServer': 'FtpServerLinkedService', 'HttpServer': 'HttpLinkedService', 'AzureSearch': 'AzureSearchLinkedService', 'CustomDataSource': 'CustomDataSourceLinkedService', 'AmazonRedshift': 'AmazonRedshiftLinkedService', 'AmazonS3': 'AmazonS3LinkedService', 'SapEcc': 'SapEccLinkedService', 'SapCloudForCustomer': 'SapCloudForCustomerLinkedService', 'Salesforce': 'SalesforceLinkedService', 'AzureDataLakeStore': 'AzureDataLakeStoreLinkedService', 'MongoDb': 'MongoDbLinkedService', 'Cassandra': 'CassandraLinkedService', 'Web': 'WebLinkedService', 'OData': 'ODataLinkedService', 'Hdfs': 'HdfsLinkedService', 'Odbc': 'OdbcLinkedService', 'AzureML': 'AzureMLLinkedService', 'Teradata': 'TeradataLinkedService', 'Db2': 'Db2LinkedService', 'Sybase': 'SybaseLinkedService', 'PostgreSql': 'PostgreSqlLinkedService', 'MySql': 'MySqlLinkedService', 'AzureMySql': 'AzureMySqlLinkedService', 'Oracle': 'OracleLinkedService', 'FileServer': 'FileServerLinkedService', 'HDInsight': 'HDInsightLinkedService', 'Dynamics': 'DynamicsLinkedService', 'CosmosDb': 'CosmosDbLinkedService', 'AzureKeyVault': 'AzureKeyVaultLinkedService', 'AzureBatch': 'AzureBatchLinkedService', 'AzureSqlDatabase': 'AzureSqlDatabaseLinkedService', 'SqlServer': 'SqlServerLinkedService', 'AzureSqlDW': 'AzureSqlDWLinkedService', 'AzureStorage': 'AzureStorageLinkedService'} + 'type': {'Responsys': 'ResponsysLinkedService', 'AzureDatabricks': 'AzureDatabricksLinkedService', 'AzureDataLakeAnalytics': 'AzureDataLakeAnalyticsLinkedService', 'HDInsightOnDemand': 'HDInsightOnDemandLinkedService', 'SalesforceMarketingCloud': 'SalesforceMarketingCloudLinkedService', 'Netezza': 'NetezzaLinkedService', 'Vertica': 'VerticaLinkedService', 'Zoho': 'ZohoLinkedService', 'Xero': 'XeroLinkedService', 'Square': 'SquareLinkedService', 'Spark': 'SparkLinkedService', 'Shopify': 'ShopifyLinkedService', 'ServiceNow': 'ServiceNowLinkedService', 'QuickBooks': 'QuickBooksLinkedService', 'Presto': 'PrestoLinkedService', 'Phoenix': 'PhoenixLinkedService', 'Paypal': 'PaypalLinkedService', 'Marketo': 'MarketoLinkedService', 'MariaDB': 'MariaDBLinkedService', 'Magento': 'MagentoLinkedService', 'Jira': 'JiraLinkedService', 'Impala': 'ImpalaLinkedService', 'Hubspot': 'HubspotLinkedService', 'Hive': 'HiveLinkedService', 'HBase': 'HBaseLinkedService', 'Greenplum': 'GreenplumLinkedService', 'GoogleBigQuery': 'GoogleBigQueryLinkedService', 'Eloqua': 'EloquaLinkedService', 'Drill': 'DrillLinkedService', 'Couchbase': 'CouchbaseLinkedService', 'Concur': 'ConcurLinkedService', 'AzurePostgreSql': 'AzurePostgreSqlLinkedService', 'AmazonMWS': 'AmazonMWSLinkedService', 'SapHana': 'SapHanaLinkedService', 'SapBW': 'SapBWLinkedService', 'Sftp': 'SftpServerLinkedService', 'FtpServer': 'FtpServerLinkedService', 'HttpServer': 'HttpLinkedService', 'AzureSearch': 'AzureSearchLinkedService', 'CustomDataSource': 'CustomDataSourceLinkedService', 'AmazonRedshift': 'AmazonRedshiftLinkedService', 'AmazonS3': 'AmazonS3LinkedService', 'SapEcc': 'SapEccLinkedService', 'SapCloudForCustomer': 'SapCloudForCustomerLinkedService', 'Salesforce': 'SalesforceLinkedService', 'AzureDataLakeStore': 'AzureDataLakeStoreLinkedService', 'MongoDb': 'MongoDbLinkedService', 'Cassandra': 'CassandraLinkedService', 'Web': 'WebLinkedService', 'OData': 'ODataLinkedService', 'Hdfs': 'HdfsLinkedService', 'Odbc': 'OdbcLinkedService', 'AzureML': 'AzureMLLinkedService', 'Teradata': 'TeradataLinkedService', 'Db2': 'Db2LinkedService', 'Sybase': 'SybaseLinkedService', 'PostgreSql': 'PostgreSqlLinkedService', 'MySql': 'MySqlLinkedService', 'AzureMySql': 'AzureMySqlLinkedService', 'Oracle': 'OracleLinkedService', 'FileServer': 'FileServerLinkedService', 'HDInsight': 'HDInsightLinkedService', 'Dynamics': 'DynamicsLinkedService', 'CosmosDb': 'CosmosDbLinkedService', 'AzureKeyVault': 'AzureKeyVaultLinkedService', 'AzureBatch': 'AzureBatchLinkedService', 'AzureSqlDatabase': 'AzureSqlDatabaseLinkedService', 'SqlServer': 'SqlServerLinkedService', 'AzureSqlDW': 'AzureSqlDWLinkedService', 'AzureStorage': 'AzureStorageLinkedService'} } - def __init__(self, additional_properties=None, connect_via=None, description=None, parameters=None, annotations=None): - super(LinkedService, self).__init__() - self.additional_properties = additional_properties - self.connect_via = connect_via - self.description = description - self.parameters = parameters - self.annotations = annotations + def __init__(self, **kwargs): + super(LinkedService, self).__init__(**kwargs) + self.additional_properties = kwargs.get('additional_properties', None) + self.connect_via = kwargs.get('connect_via', None) + self.description = kwargs.get('description', None) + self.parameters = kwargs.get('parameters', None) + self.annotations = kwargs.get('annotations', None) self.type = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service_py3.py new file mode 100644 index 000000000000..39119635cba5 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service_py3.py @@ -0,0 +1,92 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class LinkedService(Model): + """The Azure Data Factory nested object which contains the information and + credential which can be used to connect with related store or compute + resource. + + You probably want to use the sub-classes and not this class directly. Known + sub-classes are: ResponsysLinkedService, AzureDatabricksLinkedService, + AzureDataLakeAnalyticsLinkedService, HDInsightOnDemandLinkedService, + SalesforceMarketingCloudLinkedService, NetezzaLinkedService, + VerticaLinkedService, ZohoLinkedService, XeroLinkedService, + SquareLinkedService, SparkLinkedService, ShopifyLinkedService, + ServiceNowLinkedService, QuickBooksLinkedService, PrestoLinkedService, + PhoenixLinkedService, PaypalLinkedService, MarketoLinkedService, + MariaDBLinkedService, MagentoLinkedService, JiraLinkedService, + ImpalaLinkedService, HubspotLinkedService, HiveLinkedService, + HBaseLinkedService, GreenplumLinkedService, GoogleBigQueryLinkedService, + EloquaLinkedService, DrillLinkedService, CouchbaseLinkedService, + ConcurLinkedService, AzurePostgreSqlLinkedService, AmazonMWSLinkedService, + SapHanaLinkedService, SapBWLinkedService, SftpServerLinkedService, + FtpServerLinkedService, HttpLinkedService, AzureSearchLinkedService, + CustomDataSourceLinkedService, AmazonRedshiftLinkedService, + AmazonS3LinkedService, SapEccLinkedService, + SapCloudForCustomerLinkedService, SalesforceLinkedService, + AzureDataLakeStoreLinkedService, MongoDbLinkedService, + CassandraLinkedService, WebLinkedService, ODataLinkedService, + HdfsLinkedService, OdbcLinkedService, AzureMLLinkedService, + TeradataLinkedService, Db2LinkedService, SybaseLinkedService, + PostgreSqlLinkedService, MySqlLinkedService, AzureMySqlLinkedService, + OracleLinkedService, FileServerLinkedService, HDInsightLinkedService, + DynamicsLinkedService, CosmosDbLinkedService, AzureKeyVaultLinkedService, + AzureBatchLinkedService, AzureSqlDatabaseLinkedService, + SqlServerLinkedService, AzureSqlDWLinkedService, AzureStorageLinkedService + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + _subtype_map = { + 'type': {'Responsys': 'ResponsysLinkedService', 'AzureDatabricks': 'AzureDatabricksLinkedService', 'AzureDataLakeAnalytics': 'AzureDataLakeAnalyticsLinkedService', 'HDInsightOnDemand': 'HDInsightOnDemandLinkedService', 'SalesforceMarketingCloud': 'SalesforceMarketingCloudLinkedService', 'Netezza': 'NetezzaLinkedService', 'Vertica': 'VerticaLinkedService', 'Zoho': 'ZohoLinkedService', 'Xero': 'XeroLinkedService', 'Square': 'SquareLinkedService', 'Spark': 'SparkLinkedService', 'Shopify': 'ShopifyLinkedService', 'ServiceNow': 'ServiceNowLinkedService', 'QuickBooks': 'QuickBooksLinkedService', 'Presto': 'PrestoLinkedService', 'Phoenix': 'PhoenixLinkedService', 'Paypal': 'PaypalLinkedService', 'Marketo': 'MarketoLinkedService', 'MariaDB': 'MariaDBLinkedService', 'Magento': 'MagentoLinkedService', 'Jira': 'JiraLinkedService', 'Impala': 'ImpalaLinkedService', 'Hubspot': 'HubspotLinkedService', 'Hive': 'HiveLinkedService', 'HBase': 'HBaseLinkedService', 'Greenplum': 'GreenplumLinkedService', 'GoogleBigQuery': 'GoogleBigQueryLinkedService', 'Eloqua': 'EloquaLinkedService', 'Drill': 'DrillLinkedService', 'Couchbase': 'CouchbaseLinkedService', 'Concur': 'ConcurLinkedService', 'AzurePostgreSql': 'AzurePostgreSqlLinkedService', 'AmazonMWS': 'AmazonMWSLinkedService', 'SapHana': 'SapHanaLinkedService', 'SapBW': 'SapBWLinkedService', 'Sftp': 'SftpServerLinkedService', 'FtpServer': 'FtpServerLinkedService', 'HttpServer': 'HttpLinkedService', 'AzureSearch': 'AzureSearchLinkedService', 'CustomDataSource': 'CustomDataSourceLinkedService', 'AmazonRedshift': 'AmazonRedshiftLinkedService', 'AmazonS3': 'AmazonS3LinkedService', 'SapEcc': 'SapEccLinkedService', 'SapCloudForCustomer': 'SapCloudForCustomerLinkedService', 'Salesforce': 'SalesforceLinkedService', 'AzureDataLakeStore': 'AzureDataLakeStoreLinkedService', 'MongoDb': 'MongoDbLinkedService', 'Cassandra': 'CassandraLinkedService', 'Web': 'WebLinkedService', 'OData': 'ODataLinkedService', 'Hdfs': 'HdfsLinkedService', 'Odbc': 'OdbcLinkedService', 'AzureML': 'AzureMLLinkedService', 'Teradata': 'TeradataLinkedService', 'Db2': 'Db2LinkedService', 'Sybase': 'SybaseLinkedService', 'PostgreSql': 'PostgreSqlLinkedService', 'MySql': 'MySqlLinkedService', 'AzureMySql': 'AzureMySqlLinkedService', 'Oracle': 'OracleLinkedService', 'FileServer': 'FileServerLinkedService', 'HDInsight': 'HDInsightLinkedService', 'Dynamics': 'DynamicsLinkedService', 'CosmosDb': 'CosmosDbLinkedService', 'AzureKeyVault': 'AzureKeyVaultLinkedService', 'AzureBatch': 'AzureBatchLinkedService', 'AzureSqlDatabase': 'AzureSqlDatabaseLinkedService', 'SqlServer': 'SqlServerLinkedService', 'AzureSqlDW': 'AzureSqlDWLinkedService', 'AzureStorage': 'AzureStorageLinkedService'} + } + + def __init__(self, *, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, **kwargs) -> None: + super(LinkedService, self).__init__(**kwargs) + self.additional_properties = additional_properties + self.connect_via = connect_via + self.description = description + self.parameters = parameters + self.annotations = annotations + self.type = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service_reference.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service_reference.py index bedc0f2d8fe1..28ffeda7d01a 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service_reference.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service_reference.py @@ -18,10 +18,12 @@ class LinkedServiceReference(Model): Variables are only populated by the server, and will be ignored when sending a request. - :ivar type: Linked service reference type. Default value: + All required parameters must be populated in order to send to Azure. + + :ivar type: Required. Linked service reference type. Default value: "LinkedServiceReference" . :vartype type: str - :param reference_name: Reference LinkedService name. + :param reference_name: Required. Reference LinkedService name. :type reference_name: str :param parameters: Arguments for LinkedService. :type parameters: dict[str, object] @@ -40,7 +42,7 @@ class LinkedServiceReference(Model): type = "LinkedServiceReference" - def __init__(self, reference_name, parameters=None): - super(LinkedServiceReference, self).__init__() - self.reference_name = reference_name - self.parameters = parameters + def __init__(self, **kwargs): + super(LinkedServiceReference, self).__init__(**kwargs) + self.reference_name = kwargs.get('reference_name', None) + self.parameters = kwargs.get('parameters', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service_reference_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service_reference_py3.py new file mode 100644 index 000000000000..b6238130bdb6 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service_reference_py3.py @@ -0,0 +1,48 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class LinkedServiceReference(Model): + """Linked service reference type. + + Variables are only populated by the server, and will be ignored when + sending a request. + + All required parameters must be populated in order to send to Azure. + + :ivar type: Required. Linked service reference type. Default value: + "LinkedServiceReference" . + :vartype type: str + :param reference_name: Required. Reference LinkedService name. + :type reference_name: str + :param parameters: Arguments for LinkedService. + :type parameters: dict[str, object] + """ + + _validation = { + 'type': {'required': True, 'constant': True}, + 'reference_name': {'required': True}, + } + + _attribute_map = { + 'type': {'key': 'type', 'type': 'str'}, + 'reference_name': {'key': 'referenceName', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{object}'}, + } + + type = "LinkedServiceReference" + + def __init__(self, *, reference_name: str, parameters=None, **kwargs) -> None: + super(LinkedServiceReference, self).__init__(**kwargs) + self.reference_name = reference_name + self.parameters = parameters diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service_resource.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service_resource.py index cbbc70cbc1ce..75828718f589 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service_resource.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service_resource.py @@ -18,6 +18,8 @@ class LinkedServiceResource(SubResource): Variables are only populated by the server, and will be ignored when sending a request. + All required parameters must be populated in order to send to Azure. + :ivar id: The resource identifier. :vartype id: str :ivar name: The resource name. @@ -26,7 +28,7 @@ class LinkedServiceResource(SubResource): :vartype type: str :ivar etag: Etag identifies change in the resource. :vartype etag: str - :param properties: Properties of linked service. + :param properties: Required. Properties of linked service. :type properties: ~azure.mgmt.datafactory.models.LinkedService """ @@ -46,6 +48,6 @@ class LinkedServiceResource(SubResource): 'properties': {'key': 'properties', 'type': 'LinkedService'}, } - def __init__(self, properties): - super(LinkedServiceResource, self).__init__() - self.properties = properties + def __init__(self, **kwargs): + super(LinkedServiceResource, self).__init__(**kwargs) + self.properties = kwargs.get('properties', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service_resource_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service_resource_py3.py new file mode 100644 index 000000000000..c3a20944a87d --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service_resource_py3.py @@ -0,0 +1,53 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .sub_resource import SubResource + + +class LinkedServiceResource(SubResource): + """Linked service resource type. + + Variables are only populated by the server, and will be ignored when + sending a request. + + All required parameters must be populated in order to send to Azure. + + :ivar id: The resource identifier. + :vartype id: str + :ivar name: The resource name. + :vartype name: str + :ivar type: The resource type. + :vartype type: str + :ivar etag: Etag identifies change in the resource. + :vartype etag: str + :param properties: Required. Properties of linked service. + :type properties: ~azure.mgmt.datafactory.models.LinkedService + """ + + _validation = { + 'id': {'readonly': True}, + 'name': {'readonly': True}, + 'type': {'readonly': True}, + 'etag': {'readonly': True}, + 'properties': {'required': True}, + } + + _attribute_map = { + 'id': {'key': 'id', 'type': 'str'}, + 'name': {'key': 'name', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'str'}, + 'etag': {'key': 'etag', 'type': 'str'}, + 'properties': {'key': 'properties', 'type': 'LinkedService'}, + } + + def __init__(self, *, properties, **kwargs) -> None: + super(LinkedServiceResource, self).__init__(, **kwargs) + self.properties = properties diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/lookup_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/lookup_activity.py index 19f65f13047f..d5bfff330ee2 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/lookup_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/lookup_activity.py @@ -15,26 +15,28 @@ class LookupActivity(ExecutionActivity): """Lookup activity. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] - :param name: Activity name. + :param name: Required. Activity name. :type name: str :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference :param policy: Activity policy. :type policy: ~azure.mgmt.datafactory.models.ActivityPolicy - :param source: Dataset-specific source properties, same as copy activity - source. + :param source: Required. Dataset-specific source properties, same as copy + activity source. :type source: ~azure.mgmt.datafactory.models.CopySource - :param dataset: Lookup activity dataset reference. + :param dataset: Required. Lookup activity dataset reference. :type dataset: ~azure.mgmt.datafactory.models.DatasetReference :param first_row_only: Whether to return first row or all rows. Default value is true. Type: boolean (or Expression with resultType boolean). @@ -61,9 +63,9 @@ class LookupActivity(ExecutionActivity): 'first_row_only': {'key': 'typeProperties.firstRowOnly', 'type': 'object'}, } - def __init__(self, name, source, dataset, additional_properties=None, description=None, depends_on=None, linked_service_name=None, policy=None, first_row_only=None): - super(LookupActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, linked_service_name=linked_service_name, policy=policy) - self.source = source - self.dataset = dataset - self.first_row_only = first_row_only + def __init__(self, **kwargs): + super(LookupActivity, self).__init__(**kwargs) + self.source = kwargs.get('source', None) + self.dataset = kwargs.get('dataset', None) + self.first_row_only = kwargs.get('first_row_only', None) self.type = 'Lookup' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/lookup_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/lookup_activity_py3.py new file mode 100644 index 000000000000..d0ab996d6f3b --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/lookup_activity_py3.py @@ -0,0 +1,71 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .execution_activity import ExecutionActivity + + +class LookupActivity(ExecutionActivity): + """Lookup activity. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param name: Required. Activity name. + :type name: str + :param description: Activity description. + :type description: str + :param depends_on: Activity depends on condition. + :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param type: Required. Constant filled by server. + :type type: str + :param linked_service_name: Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param policy: Activity policy. + :type policy: ~azure.mgmt.datafactory.models.ActivityPolicy + :param source: Required. Dataset-specific source properties, same as copy + activity source. + :type source: ~azure.mgmt.datafactory.models.CopySource + :param dataset: Required. Lookup activity dataset reference. + :type dataset: ~azure.mgmt.datafactory.models.DatasetReference + :param first_row_only: Whether to return first row or all rows. Default + value is true. Type: boolean (or Expression with resultType boolean). + :type first_row_only: object + """ + + _validation = { + 'name': {'required': True}, + 'type': {'required': True}, + 'source': {'required': True}, + 'dataset': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'name': {'key': 'name', 'type': 'str'}, + 'description': {'key': 'description', 'type': 'str'}, + 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, + 'source': {'key': 'typeProperties.source', 'type': 'CopySource'}, + 'dataset': {'key': 'typeProperties.dataset', 'type': 'DatasetReference'}, + 'first_row_only': {'key': 'typeProperties.firstRowOnly', 'type': 'object'}, + } + + def __init__(self, *, name: str, source, dataset, additional_properties=None, description: str=None, depends_on=None, linked_service_name=None, policy=None, first_row_only=None, **kwargs) -> None: + super(LookupActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, linked_service_name=linked_service_name, policy=policy, **kwargs) + self.source = source + self.dataset = dataset + self.first_row_only = first_row_only + self.type = 'Lookup' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/magento_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/magento_linked_service.py index c908668f828f..5fb8974f28db 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/magento_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/magento_linked_service.py @@ -15,6 +15,8 @@ class MagentoLinkedService(LinkedService): """Magento server linked service. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -29,9 +31,9 @@ class MagentoLinkedService(LinkedService): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param host: The URL of the Magento instance. (i.e. + :param host: Required. The URL of the Magento instance. (i.e. 192.168.222.110/magento3) :type host: object :param access_token: The access token from Magento. @@ -72,12 +74,12 @@ class MagentoLinkedService(LinkedService): 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, host, additional_properties=None, connect_via=None, description=None, parameters=None, annotations=None, access_token=None, use_encrypted_endpoints=None, use_host_verification=None, use_peer_verification=None, encrypted_credential=None): - super(MagentoLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations) - self.host = host - self.access_token = access_token - self.use_encrypted_endpoints = use_encrypted_endpoints - self.use_host_verification = use_host_verification - self.use_peer_verification = use_peer_verification - self.encrypted_credential = encrypted_credential + def __init__(self, **kwargs): + super(MagentoLinkedService, self).__init__(**kwargs) + self.host = kwargs.get('host', None) + self.access_token = kwargs.get('access_token', None) + self.use_encrypted_endpoints = kwargs.get('use_encrypted_endpoints', None) + self.use_host_verification = kwargs.get('use_host_verification', None) + self.use_peer_verification = kwargs.get('use_peer_verification', None) + self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'Magento' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/magento_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/magento_linked_service_py3.py new file mode 100644 index 000000000000..ef85812fa1e1 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/magento_linked_service_py3.py @@ -0,0 +1,85 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class MagentoLinkedService(LinkedService): + """Magento server linked service. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param host: Required. The URL of the Magento instance. (i.e. + 192.168.222.110/magento3) + :type host: object + :param access_token: The access token from Magento. + :type access_token: ~azure.mgmt.datafactory.models.SecretBase + :param use_encrypted_endpoints: Specifies whether the data source + endpoints are encrypted using HTTPS. The default value is true. + :type use_encrypted_endpoints: object + :param use_host_verification: Specifies whether to require the host name + in the server's certificate to match the host name of the server when + connecting over SSL. The default value is true. + :type use_host_verification: object + :param use_peer_verification: Specifies whether to verify the identity of + the server when connecting over SSL. The default value is true. + :type use_peer_verification: object + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'host': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'host': {'key': 'typeProperties.host', 'type': 'object'}, + 'access_token': {'key': 'typeProperties.accessToken', 'type': 'SecretBase'}, + 'use_encrypted_endpoints': {'key': 'typeProperties.useEncryptedEndpoints', 'type': 'object'}, + 'use_host_verification': {'key': 'typeProperties.useHostVerification', 'type': 'object'}, + 'use_peer_verification': {'key': 'typeProperties.usePeerVerification', 'type': 'object'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, *, host, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, access_token=None, use_encrypted_endpoints=None, use_host_verification=None, use_peer_verification=None, encrypted_credential=None, **kwargs) -> None: + super(MagentoLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.host = host + self.access_token = access_token + self.use_encrypted_endpoints = use_encrypted_endpoints + self.use_host_verification = use_host_verification + self.use_peer_verification = use_peer_verification + self.encrypted_credential = encrypted_credential + self.type = 'Magento' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/magento_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/magento_object_dataset.py index 531ee5cea94f..2291b8c47796 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/magento_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/magento_object_dataset.py @@ -15,6 +15,8 @@ class MagentoObjectDataset(Dataset): """Magento server dataset. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -23,7 +25,7 @@ class MagentoObjectDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object - :param linked_service_name: Linked service reference. + :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference :param parameters: Parameters for dataset. @@ -32,7 +34,7 @@ class MagentoObjectDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str """ @@ -41,6 +43,16 @@ class MagentoObjectDataset(Dataset): 'type': {'required': True}, } - def __init__(self, linked_service_name, additional_properties=None, description=None, structure=None, parameters=None, annotations=None): - super(MagentoObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations) + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, **kwargs): + super(MagentoObjectDataset, self).__init__(**kwargs) self.type = 'MagentoObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/magento_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/magento_object_dataset_py3.py new file mode 100644 index 000000000000..4386f3704d13 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/magento_object_dataset_py3.py @@ -0,0 +1,58 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class MagentoObjectDataset(Dataset): + """Magento server dataset. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Required. Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: + super(MagentoObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + self.type = 'MagentoObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/magento_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/magento_source.py index 955ab68011b3..679ba2a0669e 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/magento_source.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/magento_source.py @@ -15,6 +15,8 @@ class MagentoSource(CopySource): """A copy activity Magento server source. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -25,7 +27,7 @@ class MagentoSource(CopySource): with resultType string), pattern: ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type source_retry_wait: object - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). @@ -44,7 +46,7 @@ class MagentoSource(CopySource): 'query': {'key': 'query', 'type': 'object'}, } - def __init__(self, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None): - super(MagentoSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait) - self.query = query + def __init__(self, **kwargs): + super(MagentoSource, self).__init__(**kwargs) + self.query = kwargs.get('query', None) self.type = 'MagentoSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/magento_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/magento_source_py3.py new file mode 100644 index 000000000000..3ce50a3ed452 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/magento_source_py3.py @@ -0,0 +1,52 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class MagentoSource(CopySource): + """A copy activity Magento server source. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Required. Constant filled by server. + :type type: str + :param query: A query to retrieve data from source. Type: string (or + Expression with resultType string). + :type query: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'query': {'key': 'query', 'type': 'object'}, + } + + def __init__(self, *, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None, **kwargs) -> None: + super(MagentoSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, **kwargs) + self.query = query + self.type = 'MagentoSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/managed_integration_runtime.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/managed_integration_runtime.py index 23e29e16a37c..6d2afaffd90a 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/managed_integration_runtime.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/managed_integration_runtime.py @@ -19,12 +19,14 @@ class ManagedIntegrationRuntime(IntegrationRuntime): Variables are only populated by the server, and will be ignored when sending a request. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] :param description: Integration runtime description. :type description: str - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :ivar state: Integration runtime state, only valid for managed dedicated integration runtime. Possible values include: 'Initial', 'Stopped', @@ -55,9 +57,9 @@ class ManagedIntegrationRuntime(IntegrationRuntime): 'ssis_properties': {'key': 'typeProperties.ssisProperties', 'type': 'IntegrationRuntimeSsisProperties'}, } - def __init__(self, additional_properties=None, description=None, compute_properties=None, ssis_properties=None): - super(ManagedIntegrationRuntime, self).__init__(additional_properties=additional_properties, description=description) + def __init__(self, **kwargs): + super(ManagedIntegrationRuntime, self).__init__(**kwargs) self.state = None - self.compute_properties = compute_properties - self.ssis_properties = ssis_properties + self.compute_properties = kwargs.get('compute_properties', None) + self.ssis_properties = kwargs.get('ssis_properties', None) self.type = 'Managed' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/managed_integration_runtime_error.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/managed_integration_runtime_error.py index 642975fcf5ef..c70323697fdf 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/managed_integration_runtime_error.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/managed_integration_runtime_error.py @@ -46,9 +46,9 @@ class ManagedIntegrationRuntimeError(Model): 'message': {'key': 'message', 'type': 'str'}, } - def __init__(self, additional_properties=None): - super(ManagedIntegrationRuntimeError, self).__init__() - self.additional_properties = additional_properties + def __init__(self, **kwargs): + super(ManagedIntegrationRuntimeError, self).__init__(**kwargs) + self.additional_properties = kwargs.get('additional_properties', None) self.time = None self.code = None self.parameters = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/managed_integration_runtime_error_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/managed_integration_runtime_error_py3.py new file mode 100644 index 000000000000..1668c5196537 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/managed_integration_runtime_error_py3.py @@ -0,0 +1,55 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class ManagedIntegrationRuntimeError(Model): + """Error definition for managed integration runtime. + + Variables are only populated by the server, and will be ignored when + sending a request. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :ivar time: The time when the error occurred. + :vartype time: datetime + :ivar code: Error code. + :vartype code: str + :ivar parameters: Managed integration runtime error parameters. + :vartype parameters: list[str] + :ivar message: Error message. + :vartype message: str + """ + + _validation = { + 'time': {'readonly': True}, + 'code': {'readonly': True}, + 'parameters': {'readonly': True}, + 'message': {'readonly': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'time': {'key': 'time', 'type': 'iso-8601'}, + 'code': {'key': 'code', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '[str]'}, + 'message': {'key': 'message', 'type': 'str'}, + } + + def __init__(self, *, additional_properties=None, **kwargs) -> None: + super(ManagedIntegrationRuntimeError, self).__init__(**kwargs) + self.additional_properties = additional_properties + self.time = None + self.code = None + self.parameters = None + self.message = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/managed_integration_runtime_node.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/managed_integration_runtime_node.py index 306b51ec9e45..e9c0169cf6c5 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/managed_integration_runtime_node.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/managed_integration_runtime_node.py @@ -44,9 +44,9 @@ class ManagedIntegrationRuntimeNode(Model): 'errors': {'key': 'errors', 'type': '[ManagedIntegrationRuntimeError]'}, } - def __init__(self, additional_properties=None, errors=None): - super(ManagedIntegrationRuntimeNode, self).__init__() - self.additional_properties = additional_properties + def __init__(self, **kwargs): + super(ManagedIntegrationRuntimeNode, self).__init__(**kwargs) + self.additional_properties = kwargs.get('additional_properties', None) self.node_id = None self.status = None - self.errors = errors + self.errors = kwargs.get('errors', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/managed_integration_runtime_node_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/managed_integration_runtime_node_py3.py new file mode 100644 index 000000000000..0e8104d0de05 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/managed_integration_runtime_node_py3.py @@ -0,0 +1,52 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class ManagedIntegrationRuntimeNode(Model): + """Properties of integration runtime node. + + Variables are only populated by the server, and will be ignored when + sending a request. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :ivar node_id: The managed integration runtime node id. + :vartype node_id: str + :ivar status: The managed integration runtime node status. Possible values + include: 'Starting', 'Available', 'Recycling', 'Unavailable' + :vartype status: str or + ~azure.mgmt.datafactory.models.ManagedIntegrationRuntimeNodeStatus + :param errors: The errors that occurred on this integration runtime node. + :type errors: + list[~azure.mgmt.datafactory.models.ManagedIntegrationRuntimeError] + """ + + _validation = { + 'node_id': {'readonly': True}, + 'status': {'readonly': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'node_id': {'key': 'nodeId', 'type': 'str'}, + 'status': {'key': 'status', 'type': 'str'}, + 'errors': {'key': 'errors', 'type': '[ManagedIntegrationRuntimeError]'}, + } + + def __init__(self, *, additional_properties=None, errors=None, **kwargs) -> None: + super(ManagedIntegrationRuntimeNode, self).__init__(**kwargs) + self.additional_properties = additional_properties + self.node_id = None + self.status = None + self.errors = errors diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/managed_integration_runtime_operation_result.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/managed_integration_runtime_operation_result.py index 83dc66fbb496..2329f7a2ba36 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/managed_integration_runtime_operation_result.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/managed_integration_runtime_operation_result.py @@ -54,9 +54,9 @@ class ManagedIntegrationRuntimeOperationResult(Model): 'activity_id': {'key': 'activityId', 'type': 'str'}, } - def __init__(self, additional_properties=None): - super(ManagedIntegrationRuntimeOperationResult, self).__init__() - self.additional_properties = additional_properties + def __init__(self, **kwargs): + super(ManagedIntegrationRuntimeOperationResult, self).__init__(**kwargs) + self.additional_properties = kwargs.get('additional_properties', None) self.type = None self.start_time = None self.result = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/managed_integration_runtime_operation_result_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/managed_integration_runtime_operation_result_py3.py new file mode 100644 index 000000000000..58a80c0e600e --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/managed_integration_runtime_operation_result_py3.py @@ -0,0 +1,65 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class ManagedIntegrationRuntimeOperationResult(Model): + """Properties of managed integration runtime operation result. + + Variables are only populated by the server, and will be ignored when + sending a request. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :ivar type: The operation type. Could be start or stop. + :vartype type: str + :ivar start_time: The start time of the operation. + :vartype start_time: datetime + :ivar result: The operation result. + :vartype result: str + :ivar error_code: The error code. + :vartype error_code: str + :ivar parameters: Managed integration runtime error parameters. + :vartype parameters: list[str] + :ivar activity_id: The activity id for the operation request. + :vartype activity_id: str + """ + + _validation = { + 'type': {'readonly': True}, + 'start_time': {'readonly': True}, + 'result': {'readonly': True}, + 'error_code': {'readonly': True}, + 'parameters': {'readonly': True}, + 'activity_id': {'readonly': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'type': {'key': 'type', 'type': 'str'}, + 'start_time': {'key': 'startTime', 'type': 'iso-8601'}, + 'result': {'key': 'result', 'type': 'str'}, + 'error_code': {'key': 'errorCode', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '[str]'}, + 'activity_id': {'key': 'activityId', 'type': 'str'}, + } + + def __init__(self, *, additional_properties=None, **kwargs) -> None: + super(ManagedIntegrationRuntimeOperationResult, self).__init__(**kwargs) + self.additional_properties = additional_properties + self.type = None + self.start_time = None + self.result = None + self.error_code = None + self.parameters = None + self.activity_id = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/managed_integration_runtime_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/managed_integration_runtime_py3.py new file mode 100644 index 000000000000..d2e2468aa741 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/managed_integration_runtime_py3.py @@ -0,0 +1,65 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .integration_runtime import IntegrationRuntime + + +class ManagedIntegrationRuntime(IntegrationRuntime): + """Managed integration runtime, including managed elastic and managed + dedicated integration runtimes. + + Variables are only populated by the server, and will be ignored when + sending a request. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param description: Integration runtime description. + :type description: str + :param type: Required. Constant filled by server. + :type type: str + :ivar state: Integration runtime state, only valid for managed dedicated + integration runtime. Possible values include: 'Initial', 'Stopped', + 'Started', 'Starting', 'Stopping', 'NeedRegistration', 'Online', + 'Limited', 'Offline' + :vartype state: str or + ~azure.mgmt.datafactory.models.IntegrationRuntimeState + :param compute_properties: The compute resource for managed integration + runtime. + :type compute_properties: + ~azure.mgmt.datafactory.models.IntegrationRuntimeComputeProperties + :param ssis_properties: SSIS properties for managed integration runtime. + :type ssis_properties: + ~azure.mgmt.datafactory.models.IntegrationRuntimeSsisProperties + """ + + _validation = { + 'type': {'required': True}, + 'state': {'readonly': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'str'}, + 'state': {'key': 'state', 'type': 'str'}, + 'compute_properties': {'key': 'typeProperties.computeProperties', 'type': 'IntegrationRuntimeComputeProperties'}, + 'ssis_properties': {'key': 'typeProperties.ssisProperties', 'type': 'IntegrationRuntimeSsisProperties'}, + } + + def __init__(self, *, additional_properties=None, description: str=None, compute_properties=None, ssis_properties=None, **kwargs) -> None: + super(ManagedIntegrationRuntime, self).__init__(additional_properties=additional_properties, description=description, **kwargs) + self.state = None + self.compute_properties = compute_properties + self.ssis_properties = ssis_properties + self.type = 'Managed' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/managed_integration_runtime_status.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/managed_integration_runtime_status.py index 0013f6a7dc54..9ebe40deae74 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/managed_integration_runtime_status.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/managed_integration_runtime_status.py @@ -18,6 +18,8 @@ class ManagedIntegrationRuntimeStatus(IntegrationRuntimeStatus): Variables are only populated by the server, and will be ignored when sending a request. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -29,7 +31,7 @@ class ManagedIntegrationRuntimeStatus(IntegrationRuntimeStatus): 'NeedRegistration', 'Online', 'Limited', 'Offline' :vartype state: str or ~azure.mgmt.datafactory.models.IntegrationRuntimeState - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :ivar create_time: The time at which the integration runtime was created, in ISO8601 format. @@ -67,8 +69,8 @@ class ManagedIntegrationRuntimeStatus(IntegrationRuntimeStatus): 'last_operation': {'key': 'typeProperties.lastOperation', 'type': 'ManagedIntegrationRuntimeOperationResult'}, } - def __init__(self, additional_properties=None): - super(ManagedIntegrationRuntimeStatus, self).__init__(additional_properties=additional_properties) + def __init__(self, **kwargs): + super(ManagedIntegrationRuntimeStatus, self).__init__(**kwargs) self.create_time = None self.nodes = None self.other_errors = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/managed_integration_runtime_status_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/managed_integration_runtime_status_py3.py new file mode 100644 index 000000000000..9944628a9997 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/managed_integration_runtime_status_py3.py @@ -0,0 +1,78 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .integration_runtime_status import IntegrationRuntimeStatus + + +class ManagedIntegrationRuntimeStatus(IntegrationRuntimeStatus): + """Managed integration runtime status. + + Variables are only populated by the server, and will be ignored when + sending a request. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :ivar data_factory_name: The data factory name which the integration + runtime belong to. + :vartype data_factory_name: str + :ivar state: The state of integration runtime. Possible values include: + 'Initial', 'Stopped', 'Started', 'Starting', 'Stopping', + 'NeedRegistration', 'Online', 'Limited', 'Offline' + :vartype state: str or + ~azure.mgmt.datafactory.models.IntegrationRuntimeState + :param type: Required. Constant filled by server. + :type type: str + :ivar create_time: The time at which the integration runtime was created, + in ISO8601 format. + :vartype create_time: datetime + :ivar nodes: The list of nodes for managed integration runtime. + :vartype nodes: + list[~azure.mgmt.datafactory.models.ManagedIntegrationRuntimeNode] + :ivar other_errors: The errors that occurred on this integration runtime. + :vartype other_errors: + list[~azure.mgmt.datafactory.models.ManagedIntegrationRuntimeError] + :ivar last_operation: The last operation result that occurred on this + integration runtime. + :vartype last_operation: + ~azure.mgmt.datafactory.models.ManagedIntegrationRuntimeOperationResult + """ + + _validation = { + 'data_factory_name': {'readonly': True}, + 'state': {'readonly': True}, + 'type': {'required': True}, + 'create_time': {'readonly': True}, + 'nodes': {'readonly': True}, + 'other_errors': {'readonly': True}, + 'last_operation': {'readonly': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'data_factory_name': {'key': 'dataFactoryName', 'type': 'str'}, + 'state': {'key': 'state', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'str'}, + 'create_time': {'key': 'typeProperties.createTime', 'type': 'iso-8601'}, + 'nodes': {'key': 'typeProperties.nodes', 'type': '[ManagedIntegrationRuntimeNode]'}, + 'other_errors': {'key': 'typeProperties.otherErrors', 'type': '[ManagedIntegrationRuntimeError]'}, + 'last_operation': {'key': 'typeProperties.lastOperation', 'type': 'ManagedIntegrationRuntimeOperationResult'}, + } + + def __init__(self, *, additional_properties=None, **kwargs) -> None: + super(ManagedIntegrationRuntimeStatus, self).__init__(additional_properties=additional_properties, **kwargs) + self.create_time = None + self.nodes = None + self.other_errors = None + self.last_operation = None + self.type = 'Managed' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_linked_service.py index 2e9a61f56ddc..24c0b0eef87d 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_linked_service.py @@ -15,6 +15,8 @@ class MariaDBLinkedService(LinkedService): """MariaDB server linked service. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -29,7 +31,7 @@ class MariaDBLinkedService(LinkedService): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param connection_string: An ODBC connection string. :type connection_string: ~azure.mgmt.datafactory.models.SecretBase @@ -54,8 +56,8 @@ class MariaDBLinkedService(LinkedService): 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, additional_properties=None, connect_via=None, description=None, parameters=None, annotations=None, connection_string=None, encrypted_credential=None): - super(MariaDBLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations) - self.connection_string = connection_string - self.encrypted_credential = encrypted_credential + def __init__(self, **kwargs): + super(MariaDBLinkedService, self).__init__(**kwargs) + self.connection_string = kwargs.get('connection_string', None) + self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'MariaDB' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_linked_service_py3.py new file mode 100644 index 000000000000..7a428d54120b --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_linked_service_py3.py @@ -0,0 +1,63 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class MariaDBLinkedService(LinkedService): + """MariaDB server linked service. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param connection_string: An ODBC connection string. + :type connection_string: ~azure.mgmt.datafactory.models.SecretBase + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'SecretBase'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, *, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, connection_string=None, encrypted_credential=None, **kwargs) -> None: + super(MariaDBLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.connection_string = connection_string + self.encrypted_credential = encrypted_credential + self.type = 'MariaDB' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_source.py index 0a0f978ac652..96b7116cd3ac 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_source.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_source.py @@ -15,6 +15,8 @@ class MariaDBSource(CopySource): """A copy activity MariaDB server source. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -25,7 +27,7 @@ class MariaDBSource(CopySource): with resultType string), pattern: ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type source_retry_wait: object - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). @@ -44,7 +46,7 @@ class MariaDBSource(CopySource): 'query': {'key': 'query', 'type': 'object'}, } - def __init__(self, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None): - super(MariaDBSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait) - self.query = query + def __init__(self, **kwargs): + super(MariaDBSource, self).__init__(**kwargs) + self.query = kwargs.get('query', None) self.type = 'MariaDBSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_source_py3.py new file mode 100644 index 000000000000..99fd864e7e62 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_source_py3.py @@ -0,0 +1,52 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class MariaDBSource(CopySource): + """A copy activity MariaDB server source. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Required. Constant filled by server. + :type type: str + :param query: A query to retrieve data from source. Type: string (or + Expression with resultType string). + :type query: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'query': {'key': 'query', 'type': 'object'}, + } + + def __init__(self, *, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None, **kwargs) -> None: + super(MariaDBSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, **kwargs) + self.query = query + self.type = 'MariaDBSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_table_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_table_dataset.py index a0839e26cb50..3c4d6d2874a3 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_table_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_table_dataset.py @@ -15,6 +15,8 @@ class MariaDBTableDataset(Dataset): """MariaDB server dataset. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -23,7 +25,7 @@ class MariaDBTableDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object - :param linked_service_name: Linked service reference. + :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference :param parameters: Parameters for dataset. @@ -32,7 +34,7 @@ class MariaDBTableDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str """ @@ -41,6 +43,16 @@ class MariaDBTableDataset(Dataset): 'type': {'required': True}, } - def __init__(self, linked_service_name, additional_properties=None, description=None, structure=None, parameters=None, annotations=None): - super(MariaDBTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations) + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, **kwargs): + super(MariaDBTableDataset, self).__init__(**kwargs) self.type = 'MariaDBTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_table_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_table_dataset_py3.py new file mode 100644 index 000000000000..d9a30f2deccc --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_table_dataset_py3.py @@ -0,0 +1,58 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class MariaDBTableDataset(Dataset): + """MariaDB server dataset. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Required. Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: + super(MariaDBTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + self.type = 'MariaDBTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/marketo_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/marketo_linked_service.py index 6a5c9796c590..432676824a75 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/marketo_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/marketo_linked_service.py @@ -15,6 +15,8 @@ class MarketoLinkedService(LinkedService): """Marketo server linked service. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -29,12 +31,12 @@ class MarketoLinkedService(LinkedService): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param endpoint: The endpoint of the Marketo server. (i.e. + :param endpoint: Required. The endpoint of the Marketo server. (i.e. 123-ABC-321.mktorest.com) :type endpoint: object - :param client_id: The client Id of your Marketo service. + :param client_id: Required. The client Id of your Marketo service. :type client_id: object :param client_secret: The client secret of your Marketo service. :type client_secret: ~azure.mgmt.datafactory.models.SecretBase @@ -76,13 +78,13 @@ class MarketoLinkedService(LinkedService): 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, endpoint, client_id, additional_properties=None, connect_via=None, description=None, parameters=None, annotations=None, client_secret=None, use_encrypted_endpoints=None, use_host_verification=None, use_peer_verification=None, encrypted_credential=None): - super(MarketoLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations) - self.endpoint = endpoint - self.client_id = client_id - self.client_secret = client_secret - self.use_encrypted_endpoints = use_encrypted_endpoints - self.use_host_verification = use_host_verification - self.use_peer_verification = use_peer_verification - self.encrypted_credential = encrypted_credential + def __init__(self, **kwargs): + super(MarketoLinkedService, self).__init__(**kwargs) + self.endpoint = kwargs.get('endpoint', None) + self.client_id = kwargs.get('client_id', None) + self.client_secret = kwargs.get('client_secret', None) + self.use_encrypted_endpoints = kwargs.get('use_encrypted_endpoints', None) + self.use_host_verification = kwargs.get('use_host_verification', None) + self.use_peer_verification = kwargs.get('use_peer_verification', None) + self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'Marketo' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/marketo_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/marketo_linked_service_py3.py new file mode 100644 index 000000000000..6ee722e1577c --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/marketo_linked_service_py3.py @@ -0,0 +1,90 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class MarketoLinkedService(LinkedService): + """Marketo server linked service. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param endpoint: Required. The endpoint of the Marketo server. (i.e. + 123-ABC-321.mktorest.com) + :type endpoint: object + :param client_id: Required. The client Id of your Marketo service. + :type client_id: object + :param client_secret: The client secret of your Marketo service. + :type client_secret: ~azure.mgmt.datafactory.models.SecretBase + :param use_encrypted_endpoints: Specifies whether the data source + endpoints are encrypted using HTTPS. The default value is true. + :type use_encrypted_endpoints: object + :param use_host_verification: Specifies whether to require the host name + in the server's certificate to match the host name of the server when + connecting over SSL. The default value is true. + :type use_host_verification: object + :param use_peer_verification: Specifies whether to verify the identity of + the server when connecting over SSL. The default value is true. + :type use_peer_verification: object + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'endpoint': {'required': True}, + 'client_id': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'endpoint': {'key': 'typeProperties.endpoint', 'type': 'object'}, + 'client_id': {'key': 'typeProperties.clientId', 'type': 'object'}, + 'client_secret': {'key': 'typeProperties.clientSecret', 'type': 'SecretBase'}, + 'use_encrypted_endpoints': {'key': 'typeProperties.useEncryptedEndpoints', 'type': 'object'}, + 'use_host_verification': {'key': 'typeProperties.useHostVerification', 'type': 'object'}, + 'use_peer_verification': {'key': 'typeProperties.usePeerVerification', 'type': 'object'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, *, endpoint, client_id, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, client_secret=None, use_encrypted_endpoints=None, use_host_verification=None, use_peer_verification=None, encrypted_credential=None, **kwargs) -> None: + super(MarketoLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.endpoint = endpoint + self.client_id = client_id + self.client_secret = client_secret + self.use_encrypted_endpoints = use_encrypted_endpoints + self.use_host_verification = use_host_verification + self.use_peer_verification = use_peer_verification + self.encrypted_credential = encrypted_credential + self.type = 'Marketo' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/marketo_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/marketo_object_dataset.py index d83eca684f22..ad553cde51b5 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/marketo_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/marketo_object_dataset.py @@ -15,6 +15,8 @@ class MarketoObjectDataset(Dataset): """Marketo server dataset. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -23,7 +25,7 @@ class MarketoObjectDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object - :param linked_service_name: Linked service reference. + :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference :param parameters: Parameters for dataset. @@ -32,7 +34,7 @@ class MarketoObjectDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str """ @@ -41,6 +43,16 @@ class MarketoObjectDataset(Dataset): 'type': {'required': True}, } - def __init__(self, linked_service_name, additional_properties=None, description=None, structure=None, parameters=None, annotations=None): - super(MarketoObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations) + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, **kwargs): + super(MarketoObjectDataset, self).__init__(**kwargs) self.type = 'MarketoObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/marketo_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/marketo_object_dataset_py3.py new file mode 100644 index 000000000000..3add966719c6 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/marketo_object_dataset_py3.py @@ -0,0 +1,58 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class MarketoObjectDataset(Dataset): + """Marketo server dataset. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Required. Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: + super(MarketoObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + self.type = 'MarketoObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/marketo_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/marketo_source.py index 6b6456d7b98a..4867951baae7 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/marketo_source.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/marketo_source.py @@ -15,6 +15,8 @@ class MarketoSource(CopySource): """A copy activity Marketo server source. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -25,7 +27,7 @@ class MarketoSource(CopySource): with resultType string), pattern: ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type source_retry_wait: object - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). @@ -44,7 +46,7 @@ class MarketoSource(CopySource): 'query': {'key': 'query', 'type': 'object'}, } - def __init__(self, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None): - super(MarketoSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait) - self.query = query + def __init__(self, **kwargs): + super(MarketoSource, self).__init__(**kwargs) + self.query = kwargs.get('query', None) self.type = 'MarketoSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/marketo_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/marketo_source_py3.py new file mode 100644 index 000000000000..34a97998f96e --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/marketo_source_py3.py @@ -0,0 +1,52 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class MarketoSource(CopySource): + """A copy activity Marketo server source. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Required. Constant filled by server. + :type type: str + :param query: A query to retrieve data from source. Type: string (or + Expression with resultType string). + :type query: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'query': {'key': 'query', 'type': 'object'}, + } + + def __init__(self, *, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None, **kwargs) -> None: + super(MarketoSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, **kwargs) + self.query = query + self.type = 'MarketoSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/mongo_db_collection_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/mongo_db_collection_dataset.py index 2f8e88b06735..c76dc730105a 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/mongo_db_collection_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/mongo_db_collection_dataset.py @@ -15,6 +15,8 @@ class MongoDbCollectionDataset(Dataset): """The MongoDB database dataset. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -23,7 +25,7 @@ class MongoDbCollectionDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object - :param linked_service_name: Linked service reference. + :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference :param parameters: Parameters for dataset. @@ -32,10 +34,10 @@ class MongoDbCollectionDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param collection_name: The table name of the MongoDB database. Type: - string (or Expression with resultType string). + :param collection_name: Required. The table name of the MongoDB database. + Type: string (or Expression with resultType string). :type collection_name: object """ @@ -56,7 +58,7 @@ class MongoDbCollectionDataset(Dataset): 'collection_name': {'key': 'typeProperties.collectionName', 'type': 'object'}, } - def __init__(self, linked_service_name, collection_name, additional_properties=None, description=None, structure=None, parameters=None, annotations=None): - super(MongoDbCollectionDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations) - self.collection_name = collection_name + def __init__(self, **kwargs): + super(MongoDbCollectionDataset, self).__init__(**kwargs) + self.collection_name = kwargs.get('collection_name', None) self.type = 'MongoDbCollection' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/mongo_db_collection_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/mongo_db_collection_dataset_py3.py new file mode 100644 index 000000000000..db839b82af25 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/mongo_db_collection_dataset_py3.py @@ -0,0 +1,64 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class MongoDbCollectionDataset(Dataset): + """The MongoDB database dataset. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Required. Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param collection_name: Required. The table name of the MongoDB database. + Type: string (or Expression with resultType string). + :type collection_name: object + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + 'collection_name': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'collection_name': {'key': 'typeProperties.collectionName', 'type': 'object'}, + } + + def __init__(self, *, linked_service_name, collection_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: + super(MongoDbCollectionDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + self.collection_name = collection_name + self.type = 'MongoDbCollection' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/mongo_db_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/mongo_db_linked_service.py index d0978bab47b4..49d53510f7fd 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/mongo_db_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/mongo_db_linked_service.py @@ -15,6 +15,8 @@ class MongoDbLinkedService(LinkedService): """Linked service for MongoDb data source. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -29,17 +31,17 @@ class MongoDbLinkedService(LinkedService): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param server: The IP address or server name of the MongoDB server. Type: - string (or Expression with resultType string). + :param server: Required. The IP address or server name of the MongoDB + server. Type: string (or Expression with resultType string). :type server: object :param authentication_type: The authentication type to be used to connect to the MongoDB database. Possible values include: 'Basic', 'Anonymous' :type authentication_type: str or ~azure.mgmt.datafactory.models.MongoDbAuthenticationType - :param database_name: The name of the MongoDB database that you want to - access. Type: string (or Expression with resultType string). + :param database_name: Required. The name of the MongoDB database that you + want to access. Type: string (or Expression with resultType string). :type database_name: object :param username: Username for authentication. Type: string (or Expression with resultType string). @@ -92,16 +94,16 @@ class MongoDbLinkedService(LinkedService): 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, server, database_name, additional_properties=None, connect_via=None, description=None, parameters=None, annotations=None, authentication_type=None, username=None, password=None, auth_source=None, port=None, enable_ssl=None, allow_self_signed_server_cert=None, encrypted_credential=None): - super(MongoDbLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations) - self.server = server - self.authentication_type = authentication_type - self.database_name = database_name - self.username = username - self.password = password - self.auth_source = auth_source - self.port = port - self.enable_ssl = enable_ssl - self.allow_self_signed_server_cert = allow_self_signed_server_cert - self.encrypted_credential = encrypted_credential + def __init__(self, **kwargs): + super(MongoDbLinkedService, self).__init__(**kwargs) + self.server = kwargs.get('server', None) + self.authentication_type = kwargs.get('authentication_type', None) + self.database_name = kwargs.get('database_name', None) + self.username = kwargs.get('username', None) + self.password = kwargs.get('password', None) + self.auth_source = kwargs.get('auth_source', None) + self.port = kwargs.get('port', None) + self.enable_ssl = kwargs.get('enable_ssl', None) + self.allow_self_signed_server_cert = kwargs.get('allow_self_signed_server_cert', None) + self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'MongoDb' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/mongo_db_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/mongo_db_linked_service_py3.py new file mode 100644 index 000000000000..4f260bf37e90 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/mongo_db_linked_service_py3.py @@ -0,0 +1,109 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class MongoDbLinkedService(LinkedService): + """Linked service for MongoDb data source. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param server: Required. The IP address or server name of the MongoDB + server. Type: string (or Expression with resultType string). + :type server: object + :param authentication_type: The authentication type to be used to connect + to the MongoDB database. Possible values include: 'Basic', 'Anonymous' + :type authentication_type: str or + ~azure.mgmt.datafactory.models.MongoDbAuthenticationType + :param database_name: Required. The name of the MongoDB database that you + want to access. Type: string (or Expression with resultType string). + :type database_name: object + :param username: Username for authentication. Type: string (or Expression + with resultType string). + :type username: object + :param password: Password for authentication. + :type password: ~azure.mgmt.datafactory.models.SecretBase + :param auth_source: Database to verify the username and password. Type: + string (or Expression with resultType string). + :type auth_source: object + :param port: The TCP port number that the MongoDB server uses to listen + for client connections. The default value is 27017. Type: integer (or + Expression with resultType integer), minimum: 0. + :type port: object + :param enable_ssl: Specifies whether the connections to the server are + encrypted using SSL. The default value is false. Type: boolean (or + Expression with resultType boolean). + :type enable_ssl: object + :param allow_self_signed_server_cert: Specifies whether to allow + self-signed certificates from the server. The default value is false. + Type: boolean (or Expression with resultType boolean). + :type allow_self_signed_server_cert: object + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'server': {'required': True}, + 'database_name': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'server': {'key': 'typeProperties.server', 'type': 'object'}, + 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, + 'database_name': {'key': 'typeProperties.databaseName', 'type': 'object'}, + 'username': {'key': 'typeProperties.username', 'type': 'object'}, + 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, + 'auth_source': {'key': 'typeProperties.authSource', 'type': 'object'}, + 'port': {'key': 'typeProperties.port', 'type': 'object'}, + 'enable_ssl': {'key': 'typeProperties.enableSsl', 'type': 'object'}, + 'allow_self_signed_server_cert': {'key': 'typeProperties.allowSelfSignedServerCert', 'type': 'object'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, *, server, database_name, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, authentication_type=None, username=None, password=None, auth_source=None, port=None, enable_ssl=None, allow_self_signed_server_cert=None, encrypted_credential=None, **kwargs) -> None: + super(MongoDbLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.server = server + self.authentication_type = authentication_type + self.database_name = database_name + self.username = username + self.password = password + self.auth_source = auth_source + self.port = port + self.enable_ssl = enable_ssl + self.allow_self_signed_server_cert = allow_self_signed_server_cert + self.encrypted_credential = encrypted_credential + self.type = 'MongoDb' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/mongo_db_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/mongo_db_source.py index d7e058c9b480..b9f0be6b97d3 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/mongo_db_source.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/mongo_db_source.py @@ -15,6 +15,8 @@ class MongoDbSource(CopySource): """A copy activity source for a MongoDB database. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -25,7 +27,7 @@ class MongoDbSource(CopySource): with resultType string), pattern: ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type source_retry_wait: object - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param query: Database query. Should be a SQL-92 query expression. Type: string (or Expression with resultType string). @@ -44,7 +46,7 @@ class MongoDbSource(CopySource): 'query': {'key': 'query', 'type': 'object'}, } - def __init__(self, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None): - super(MongoDbSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait) - self.query = query + def __init__(self, **kwargs): + super(MongoDbSource, self).__init__(**kwargs) + self.query = kwargs.get('query', None) self.type = 'MongoDbSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/mongo_db_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/mongo_db_source_py3.py new file mode 100644 index 000000000000..4313cac8c921 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/mongo_db_source_py3.py @@ -0,0 +1,52 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class MongoDbSource(CopySource): + """A copy activity source for a MongoDB database. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Required. Constant filled by server. + :type type: str + :param query: Database query. Should be a SQL-92 query expression. Type: + string (or Expression with resultType string). + :type query: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'query': {'key': 'query', 'type': 'object'}, + } + + def __init__(self, *, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None, **kwargs) -> None: + super(MongoDbSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, **kwargs) + self.query = query + self.type = 'MongoDbSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/multiple_pipeline_trigger.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/multiple_pipeline_trigger.py index a49ed60681b0..ba053baee118 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/multiple_pipeline_trigger.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/multiple_pipeline_trigger.py @@ -22,6 +22,8 @@ class MultiplePipelineTrigger(Trigger): Variables are only populated by the server, and will be ignored when sending a request. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -32,7 +34,7 @@ class MultiplePipelineTrigger(Trigger): 'Started', 'Stopped', 'Disabled' :vartype runtime_state: str or ~azure.mgmt.datafactory.models.TriggerRuntimeState - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param pipelines: Pipelines that need to be started. :type pipelines: @@ -56,7 +58,7 @@ class MultiplePipelineTrigger(Trigger): 'type': {'BlobTrigger': 'BlobTrigger', 'ScheduleTrigger': 'ScheduleTrigger'} } - def __init__(self, additional_properties=None, description=None, pipelines=None): - super(MultiplePipelineTrigger, self).__init__(additional_properties=additional_properties, description=description) - self.pipelines = pipelines + def __init__(self, **kwargs): + super(MultiplePipelineTrigger, self).__init__(**kwargs) + self.pipelines = kwargs.get('pipelines', None) self.type = 'MultiplePipelineTrigger' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/multiple_pipeline_trigger_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/multiple_pipeline_trigger_py3.py new file mode 100644 index 000000000000..8724414f4e0b --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/multiple_pipeline_trigger_py3.py @@ -0,0 +1,64 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .trigger import Trigger + + +class MultiplePipelineTrigger(Trigger): + """Base class for all triggers that support one to many model for trigger to + pipeline. + + You probably want to use the sub-classes and not this class directly. Known + sub-classes are: BlobTrigger, ScheduleTrigger + + Variables are only populated by the server, and will be ignored when + sending a request. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param description: Trigger description. + :type description: str + :ivar runtime_state: Indicates if trigger is running or not. Updated when + Start/Stop APIs are called on the Trigger. Possible values include: + 'Started', 'Stopped', 'Disabled' + :vartype runtime_state: str or + ~azure.mgmt.datafactory.models.TriggerRuntimeState + :param type: Required. Constant filled by server. + :type type: str + :param pipelines: Pipelines that need to be started. + :type pipelines: + list[~azure.mgmt.datafactory.models.TriggerPipelineReference] + """ + + _validation = { + 'runtime_state': {'readonly': True}, + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'runtime_state': {'key': 'runtimeState', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'str'}, + 'pipelines': {'key': 'pipelines', 'type': '[TriggerPipelineReference]'}, + } + + _subtype_map = { + 'type': {'BlobTrigger': 'BlobTrigger', 'ScheduleTrigger': 'ScheduleTrigger'} + } + + def __init__(self, *, additional_properties=None, description: str=None, pipelines=None, **kwargs) -> None: + super(MultiplePipelineTrigger, self).__init__(additional_properties=additional_properties, description=description, **kwargs) + self.pipelines = pipelines + self.type = 'MultiplePipelineTrigger' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/my_sql_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/my_sql_linked_service.py index f6f16648d036..bc5d7e820333 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/my_sql_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/my_sql_linked_service.py @@ -15,6 +15,8 @@ class MySqlLinkedService(LinkedService): """Linked service for MySQL data source. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -29,13 +31,13 @@ class MySqlLinkedService(LinkedService): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param server: Server name for connection. Type: string (or Expression - with resultType string). + :param server: Required. Server name for connection. Type: string (or + Expression with resultType string). :type server: object - :param database: Database name for connection. Type: string (or Expression - with resultType string). + :param database: Required. Database name for connection. Type: string (or + Expression with resultType string). :type database: object :param schema: Schema name for connection. Type: string (or Expression with resultType string). @@ -72,12 +74,12 @@ class MySqlLinkedService(LinkedService): 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, server, database, additional_properties=None, connect_via=None, description=None, parameters=None, annotations=None, schema=None, username=None, password=None, encrypted_credential=None): - super(MySqlLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations) - self.server = server - self.database = database - self.schema = schema - self.username = username - self.password = password - self.encrypted_credential = encrypted_credential + def __init__(self, **kwargs): + super(MySqlLinkedService, self).__init__(**kwargs) + self.server = kwargs.get('server', None) + self.database = kwargs.get('database', None) + self.schema = kwargs.get('schema', None) + self.username = kwargs.get('username', None) + self.password = kwargs.get('password', None) + self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'MySql' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/my_sql_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/my_sql_linked_service_py3.py new file mode 100644 index 000000000000..44170220fdfe --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/my_sql_linked_service_py3.py @@ -0,0 +1,85 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class MySqlLinkedService(LinkedService): + """Linked service for MySQL data source. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param server: Required. Server name for connection. Type: string (or + Expression with resultType string). + :type server: object + :param database: Required. Database name for connection. Type: string (or + Expression with resultType string). + :type database: object + :param schema: Schema name for connection. Type: string (or Expression + with resultType string). + :type schema: object + :param username: Username for authentication. Type: string (or Expression + with resultType string). + :type username: object + :param password: Password for authentication. + :type password: ~azure.mgmt.datafactory.models.SecretBase + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'server': {'required': True}, + 'database': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'server': {'key': 'typeProperties.server', 'type': 'object'}, + 'database': {'key': 'typeProperties.database', 'type': 'object'}, + 'schema': {'key': 'typeProperties.schema', 'type': 'object'}, + 'username': {'key': 'typeProperties.username', 'type': 'object'}, + 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, *, server, database, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, schema=None, username=None, password=None, encrypted_credential=None, **kwargs) -> None: + super(MySqlLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.server = server + self.database = database + self.schema = schema + self.username = username + self.password = password + self.encrypted_credential = encrypted_credential + self.type = 'MySql' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/netezza_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/netezza_linked_service.py index 8ab0af664672..a848845b448e 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/netezza_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/netezza_linked_service.py @@ -15,6 +15,8 @@ class NetezzaLinkedService(LinkedService): """Netezza linked service. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -29,7 +31,7 @@ class NetezzaLinkedService(LinkedService): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param connection_string: An ODBC connection string. :type connection_string: ~azure.mgmt.datafactory.models.SecretBase @@ -54,8 +56,8 @@ class NetezzaLinkedService(LinkedService): 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, additional_properties=None, connect_via=None, description=None, parameters=None, annotations=None, connection_string=None, encrypted_credential=None): - super(NetezzaLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations) - self.connection_string = connection_string - self.encrypted_credential = encrypted_credential + def __init__(self, **kwargs): + super(NetezzaLinkedService, self).__init__(**kwargs) + self.connection_string = kwargs.get('connection_string', None) + self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'Netezza' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/netezza_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/netezza_linked_service_py3.py new file mode 100644 index 000000000000..52b3cc6932a0 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/netezza_linked_service_py3.py @@ -0,0 +1,63 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class NetezzaLinkedService(LinkedService): + """Netezza linked service. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param connection_string: An ODBC connection string. + :type connection_string: ~azure.mgmt.datafactory.models.SecretBase + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'SecretBase'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, *, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, connection_string=None, encrypted_credential=None, **kwargs) -> None: + super(NetezzaLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.connection_string = connection_string + self.encrypted_credential = encrypted_credential + self.type = 'Netezza' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/netezza_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/netezza_source.py index cf251c7e1174..0c08b1440614 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/netezza_source.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/netezza_source.py @@ -15,6 +15,8 @@ class NetezzaSource(CopySource): """A copy activity Netezza source. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -25,7 +27,7 @@ class NetezzaSource(CopySource): with resultType string), pattern: ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type source_retry_wait: object - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). @@ -44,7 +46,7 @@ class NetezzaSource(CopySource): 'query': {'key': 'query', 'type': 'object'}, } - def __init__(self, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None): - super(NetezzaSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait) - self.query = query + def __init__(self, **kwargs): + super(NetezzaSource, self).__init__(**kwargs) + self.query = kwargs.get('query', None) self.type = 'NetezzaSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/netezza_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/netezza_source_py3.py new file mode 100644 index 000000000000..393dedf4dc55 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/netezza_source_py3.py @@ -0,0 +1,52 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class NetezzaSource(CopySource): + """A copy activity Netezza source. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Required. Constant filled by server. + :type type: str + :param query: A query to retrieve data from source. Type: string (or + Expression with resultType string). + :type query: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'query': {'key': 'query', 'type': 'object'}, + } + + def __init__(self, *, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None, **kwargs) -> None: + super(NetezzaSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, **kwargs) + self.query = query + self.type = 'NetezzaSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/netezza_table_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/netezza_table_dataset.py index 0f1564aa8530..399eb3872869 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/netezza_table_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/netezza_table_dataset.py @@ -15,6 +15,8 @@ class NetezzaTableDataset(Dataset): """Netezza dataset. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -23,7 +25,7 @@ class NetezzaTableDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object - :param linked_service_name: Linked service reference. + :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference :param parameters: Parameters for dataset. @@ -32,7 +34,7 @@ class NetezzaTableDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str """ @@ -41,6 +43,16 @@ class NetezzaTableDataset(Dataset): 'type': {'required': True}, } - def __init__(self, linked_service_name, additional_properties=None, description=None, structure=None, parameters=None, annotations=None): - super(NetezzaTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations) + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, **kwargs): + super(NetezzaTableDataset, self).__init__(**kwargs) self.type = 'NetezzaTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/netezza_table_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/netezza_table_dataset_py3.py new file mode 100644 index 000000000000..4acbace2398e --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/netezza_table_dataset_py3.py @@ -0,0 +1,58 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class NetezzaTableDataset(Dataset): + """Netezza dataset. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Required. Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: + super(NetezzaTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + self.type = 'NetezzaTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odata_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odata_linked_service.py index 8b3723629d72..9a7edca9ddb1 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odata_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odata_linked_service.py @@ -15,6 +15,8 @@ class ODataLinkedService(LinkedService): """Open Data Protocol (OData) linked service. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -29,10 +31,10 @@ class ODataLinkedService(LinkedService): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param url: The URL of the OData service endpoint. Type: string (or - Expression with resultType string). + :param url: Required. The URL of the OData service endpoint. Type: string + (or Expression with resultType string). :type url: object :param authentication_type: Type of authentication used to connect to the OData service. Possible values include: 'Basic', 'Anonymous' @@ -68,11 +70,11 @@ class ODataLinkedService(LinkedService): 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, url, additional_properties=None, connect_via=None, description=None, parameters=None, annotations=None, authentication_type=None, user_name=None, password=None, encrypted_credential=None): - super(ODataLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations) - self.url = url - self.authentication_type = authentication_type - self.user_name = user_name - self.password = password - self.encrypted_credential = encrypted_credential + def __init__(self, **kwargs): + super(ODataLinkedService, self).__init__(**kwargs) + self.url = kwargs.get('url', None) + self.authentication_type = kwargs.get('authentication_type', None) + self.user_name = kwargs.get('user_name', None) + self.password = kwargs.get('password', None) + self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'OData' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odata_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odata_linked_service_py3.py new file mode 100644 index 000000000000..abb4fedf5adf --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odata_linked_service_py3.py @@ -0,0 +1,80 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class ODataLinkedService(LinkedService): + """Open Data Protocol (OData) linked service. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param url: Required. The URL of the OData service endpoint. Type: string + (or Expression with resultType string). + :type url: object + :param authentication_type: Type of authentication used to connect to the + OData service. Possible values include: 'Basic', 'Anonymous' + :type authentication_type: str or + ~azure.mgmt.datafactory.models.ODataAuthenticationType + :param user_name: User name of the OData service. Type: string (or + Expression with resultType string). + :type user_name: object + :param password: Password of the OData service. + :type password: ~azure.mgmt.datafactory.models.SecretBase + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'url': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'url': {'key': 'typeProperties.url', 'type': 'object'}, + 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, + 'user_name': {'key': 'typeProperties.userName', 'type': 'object'}, + 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, *, url, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, authentication_type=None, user_name=None, password=None, encrypted_credential=None, **kwargs) -> None: + super(ODataLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.url = url + self.authentication_type = authentication_type + self.user_name = user_name + self.password = password + self.encrypted_credential = encrypted_credential + self.type = 'OData' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odata_resource_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odata_resource_dataset.py index 18d6bf3cc301..9e23abc0dd0a 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odata_resource_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odata_resource_dataset.py @@ -15,6 +15,8 @@ class ODataResourceDataset(Dataset): """The Open Data Protocol (OData) resource dataset. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -23,7 +25,7 @@ class ODataResourceDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object - :param linked_service_name: Linked service reference. + :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference :param parameters: Parameters for dataset. @@ -32,7 +34,7 @@ class ODataResourceDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param path: The OData resource path. Type: string (or Expression with resultType string). @@ -55,7 +57,7 @@ class ODataResourceDataset(Dataset): 'path': {'key': 'typeProperties.path', 'type': 'object'}, } - def __init__(self, linked_service_name, additional_properties=None, description=None, structure=None, parameters=None, annotations=None, path=None): - super(ODataResourceDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations) - self.path = path + def __init__(self, **kwargs): + super(ODataResourceDataset, self).__init__(**kwargs) + self.path = kwargs.get('path', None) self.type = 'ODataResource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odata_resource_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odata_resource_dataset_py3.py new file mode 100644 index 000000000000..03227b214202 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odata_resource_dataset_py3.py @@ -0,0 +1,63 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class ODataResourceDataset(Dataset): + """The Open Data Protocol (OData) resource dataset. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Required. Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param path: The OData resource path. Type: string (or Expression with + resultType string). + :type path: object + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'path': {'key': 'typeProperties.path', 'type': 'object'}, + } + + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, path=None, **kwargs) -> None: + super(ODataResourceDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + self.path = path + self.type = 'ODataResource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odbc_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odbc_linked_service.py index 09b31d48542d..4f04d86f8f8f 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odbc_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odbc_linked_service.py @@ -15,6 +15,8 @@ class OdbcLinkedService(LinkedService): """Open Database Connectivity (ODBC) linked service. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -29,10 +31,10 @@ class OdbcLinkedService(LinkedService): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param connection_string: The non-access credential portion of the - connection string as well as an optional encrypted credential. + :param connection_string: Required. The non-access credential portion of + the connection string as well as an optional encrypted credential. :type connection_string: ~azure.mgmt.datafactory.models.SecretBase :param authentication_type: Type of authentication used to connect to the ODBC data store. Possible values are: Anonymous and Basic. Type: string @@ -72,12 +74,12 @@ class OdbcLinkedService(LinkedService): 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, connection_string, additional_properties=None, connect_via=None, description=None, parameters=None, annotations=None, authentication_type=None, credential=None, user_name=None, password=None, encrypted_credential=None): - super(OdbcLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations) - self.connection_string = connection_string - self.authentication_type = authentication_type - self.credential = credential - self.user_name = user_name - self.password = password - self.encrypted_credential = encrypted_credential + def __init__(self, **kwargs): + super(OdbcLinkedService, self).__init__(**kwargs) + self.connection_string = kwargs.get('connection_string', None) + self.authentication_type = kwargs.get('authentication_type', None) + self.credential = kwargs.get('credential', None) + self.user_name = kwargs.get('user_name', None) + self.password = kwargs.get('password', None) + self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'Odbc' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odbc_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odbc_linked_service_py3.py new file mode 100644 index 000000000000..3d30c01d7397 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odbc_linked_service_py3.py @@ -0,0 +1,85 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class OdbcLinkedService(LinkedService): + """Open Database Connectivity (ODBC) linked service. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param connection_string: Required. The non-access credential portion of + the connection string as well as an optional encrypted credential. + :type connection_string: ~azure.mgmt.datafactory.models.SecretBase + :param authentication_type: Type of authentication used to connect to the + ODBC data store. Possible values are: Anonymous and Basic. Type: string + (or Expression with resultType string). + :type authentication_type: object + :param credential: The access credential portion of the connection string + specified in driver-specific property-value format. + :type credential: ~azure.mgmt.datafactory.models.SecretBase + :param user_name: User name for Basic authentication. Type: string (or + Expression with resultType string). + :type user_name: object + :param password: Password for Basic authentication. + :type password: ~azure.mgmt.datafactory.models.SecretBase + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'connection_string': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'SecretBase'}, + 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'object'}, + 'credential': {'key': 'typeProperties.credential', 'type': 'SecretBase'}, + 'user_name': {'key': 'typeProperties.userName', 'type': 'object'}, + 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, *, connection_string, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, authentication_type=None, credential=None, user_name=None, password=None, encrypted_credential=None, **kwargs) -> None: + super(OdbcLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.connection_string = connection_string + self.authentication_type = authentication_type + self.credential = credential + self.user_name = user_name + self.password = password + self.encrypted_credential = encrypted_credential + self.type = 'Odbc' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odbc_sink.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odbc_sink.py index a91343f20d30..4598952cb21b 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odbc_sink.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odbc_sink.py @@ -15,6 +15,8 @@ class OdbcSink(CopySink): """A copy activity ODBC sink. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -32,7 +34,7 @@ class OdbcSink(CopySink): resultType string), pattern: ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type sink_retry_wait: object - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param pre_copy_script: A query to execute before starting the copy. Type: string (or Expression with resultType string). @@ -53,7 +55,7 @@ class OdbcSink(CopySink): 'pre_copy_script': {'key': 'preCopyScript', 'type': 'object'}, } - def __init__(self, additional_properties=None, write_batch_size=None, write_batch_timeout=None, sink_retry_count=None, sink_retry_wait=None, pre_copy_script=None): - super(OdbcSink, self).__init__(additional_properties=additional_properties, write_batch_size=write_batch_size, write_batch_timeout=write_batch_timeout, sink_retry_count=sink_retry_count, sink_retry_wait=sink_retry_wait) - self.pre_copy_script = pre_copy_script + def __init__(self, **kwargs): + super(OdbcSink, self).__init__(**kwargs) + self.pre_copy_script = kwargs.get('pre_copy_script', None) self.type = 'OdbcSink' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odbc_sink_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odbc_sink_py3.py new file mode 100644 index 000000000000..8d098f6a96a8 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odbc_sink_py3.py @@ -0,0 +1,61 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_sink import CopySink + + +class OdbcSink(CopySink): + """A copy activity ODBC sink. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param write_batch_size: Write batch size. Type: integer (or Expression + with resultType integer), minimum: 0. + :type write_batch_size: object + :param write_batch_timeout: Write batch timeout. Type: string (or + Expression with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type write_batch_timeout: object + :param sink_retry_count: Sink retry count. Type: integer (or Expression + with resultType integer). + :type sink_retry_count: object + :param sink_retry_wait: Sink retry wait. Type: string (or Expression with + resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type sink_retry_wait: object + :param type: Required. Constant filled by server. + :type type: str + :param pre_copy_script: A query to execute before starting the copy. Type: + string (or Expression with resultType string). + :type pre_copy_script: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, + 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, + 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, + 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'pre_copy_script': {'key': 'preCopyScript', 'type': 'object'}, + } + + def __init__(self, *, additional_properties=None, write_batch_size=None, write_batch_timeout=None, sink_retry_count=None, sink_retry_wait=None, pre_copy_script=None, **kwargs) -> None: + super(OdbcSink, self).__init__(additional_properties=additional_properties, write_batch_size=write_batch_size, write_batch_timeout=write_batch_timeout, sink_retry_count=sink_retry_count, sink_retry_wait=sink_retry_wait, **kwargs) + self.pre_copy_script = pre_copy_script + self.type = 'OdbcSink' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation.py index dfaf8d979082..db8cde8db784 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation.py @@ -33,9 +33,9 @@ class Operation(Model): 'service_specification': {'key': 'properties.serviceSpecification', 'type': 'OperationServiceSpecification'}, } - def __init__(self, name=None, origin=None, display=None, service_specification=None): - super(Operation, self).__init__() - self.name = name - self.origin = origin - self.display = display - self.service_specification = service_specification + def __init__(self, **kwargs): + super(Operation, self).__init__(**kwargs) + self.name = kwargs.get('name', None) + self.origin = kwargs.get('origin', None) + self.display = kwargs.get('display', None) + self.service_specification = kwargs.get('service_specification', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_display.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_display.py index 44a481206fb6..1d96541c0581 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_display.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_display.py @@ -33,9 +33,9 @@ class OperationDisplay(Model): 'operation': {'key': 'operation', 'type': 'str'}, } - def __init__(self, description=None, provider=None, resource=None, operation=None): - super(OperationDisplay, self).__init__() - self.description = description - self.provider = provider - self.resource = resource - self.operation = operation + def __init__(self, **kwargs): + super(OperationDisplay, self).__init__(**kwargs) + self.description = kwargs.get('description', None) + self.provider = kwargs.get('provider', None) + self.resource = kwargs.get('resource', None) + self.operation = kwargs.get('operation', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_display_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_display_py3.py new file mode 100644 index 000000000000..dfbb782627f4 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_display_py3.py @@ -0,0 +1,41 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class OperationDisplay(Model): + """Metadata associated with the operation. + + :param description: The description of the operation. + :type description: str + :param provider: The name of the provider. + :type provider: str + :param resource: The name of the resource type on which the operation is + performed. + :type resource: str + :param operation: The type of operation: get, read, delete, etc. + :type operation: str + """ + + _attribute_map = { + 'description': {'key': 'description', 'type': 'str'}, + 'provider': {'key': 'provider', 'type': 'str'}, + 'resource': {'key': 'resource', 'type': 'str'}, + 'operation': {'key': 'operation', 'type': 'str'}, + } + + def __init__(self, *, description: str=None, provider: str=None, resource: str=None, operation: str=None, **kwargs) -> None: + super(OperationDisplay, self).__init__(**kwargs) + self.description = description + self.provider = provider + self.resource = resource + self.operation = operation diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_list_response.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_list_response.py index c56c649a09c9..5b2c7f8cb037 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_list_response.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_list_response.py @@ -28,7 +28,7 @@ class OperationListResponse(Model): 'next_link': {'key': 'nextLink', 'type': 'str'}, } - def __init__(self, value=None, next_link=None): - super(OperationListResponse, self).__init__() - self.value = value - self.next_link = next_link + def __init__(self, **kwargs): + super(OperationListResponse, self).__init__(**kwargs) + self.value = kwargs.get('value', None) + self.next_link = kwargs.get('next_link', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_list_response_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_list_response_py3.py new file mode 100644 index 000000000000..fe5c1ce91260 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_list_response_py3.py @@ -0,0 +1,34 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class OperationListResponse(Model): + """A list of operations that can be performed by the Data Factory service. + + :param value: List of Data Factory operations supported by the Data + Factory resource provider. + :type value: list[~azure.mgmt.datafactory.models.Operation] + :param next_link: The link to the next page of results, if any remaining + results exist. + :type next_link: str + """ + + _attribute_map = { + 'value': {'key': 'value', 'type': '[Operation]'}, + 'next_link': {'key': 'nextLink', 'type': 'str'}, + } + + def __init__(self, *, value=None, next_link: str=None, **kwargs) -> None: + super(OperationListResponse, self).__init__(**kwargs) + self.value = value + self.next_link = next_link diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_log_specification.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_log_specification.py index 304707a33606..93bfaf4ed0de 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_log_specification.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_log_specification.py @@ -30,8 +30,8 @@ class OperationLogSpecification(Model): 'blob_duration': {'key': 'blobDuration', 'type': 'str'}, } - def __init__(self, name=None, display_name=None, blob_duration=None): - super(OperationLogSpecification, self).__init__() - self.name = name - self.display_name = display_name - self.blob_duration = blob_duration + def __init__(self, **kwargs): + super(OperationLogSpecification, self).__init__(**kwargs) + self.name = kwargs.get('name', None) + self.display_name = kwargs.get('display_name', None) + self.blob_duration = kwargs.get('blob_duration', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_log_specification_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_log_specification_py3.py new file mode 100644 index 000000000000..2cdd941fab7b --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_log_specification_py3.py @@ -0,0 +1,37 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class OperationLogSpecification(Model): + """Details about an operation related to logs. + + :param name: The name of the log category. + :type name: str + :param display_name: Localized display name. + :type display_name: str + :param blob_duration: Blobs created in the customer storage account, per + hour. + :type blob_duration: str + """ + + _attribute_map = { + 'name': {'key': 'name', 'type': 'str'}, + 'display_name': {'key': 'displayName', 'type': 'str'}, + 'blob_duration': {'key': 'blobDuration', 'type': 'str'}, + } + + def __init__(self, *, name: str=None, display_name: str=None, blob_duration: str=None, **kwargs) -> None: + super(OperationLogSpecification, self).__init__(**kwargs) + self.name = name + self.display_name = display_name + self.blob_duration = blob_duration diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_metric_availability.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_metric_availability.py index 2e2053aedca7..974e0cbf4b0b 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_metric_availability.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_metric_availability.py @@ -27,7 +27,7 @@ class OperationMetricAvailability(Model): 'blob_duration': {'key': 'blobDuration', 'type': 'str'}, } - def __init__(self, time_grain=None, blob_duration=None): - super(OperationMetricAvailability, self).__init__() - self.time_grain = time_grain - self.blob_duration = blob_duration + def __init__(self, **kwargs): + super(OperationMetricAvailability, self).__init__(**kwargs) + self.time_grain = kwargs.get('time_grain', None) + self.blob_duration = kwargs.get('blob_duration', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_metric_availability_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_metric_availability_py3.py new file mode 100644 index 000000000000..312b83a23701 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_metric_availability_py3.py @@ -0,0 +1,33 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class OperationMetricAvailability(Model): + """Defines how often data for a metric becomes available. + + :param time_grain: The granularity for the metric. + :type time_grain: str + :param blob_duration: Blob created in the customer storage account, per + hour. + :type blob_duration: str + """ + + _attribute_map = { + 'time_grain': {'key': 'timeGrain', 'type': 'str'}, + 'blob_duration': {'key': 'blobDuration', 'type': 'str'}, + } + + def __init__(self, *, time_grain: str=None, blob_duration: str=None, **kwargs) -> None: + super(OperationMetricAvailability, self).__init__(**kwargs) + self.time_grain = time_grain + self.blob_duration = blob_duration diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_metric_specification.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_metric_specification.py index e63942dc5931..e2a9c98580a8 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_metric_specification.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_metric_specification.py @@ -50,14 +50,14 @@ class OperationMetricSpecification(Model): 'availabilities': {'key': 'availabilities', 'type': '[OperationMetricAvailability]'}, } - def __init__(self, name=None, display_name=None, display_description=None, unit=None, aggregation_type=None, enable_regional_mdm_account=None, source_mdm_account=None, source_mdm_namespace=None, availabilities=None): - super(OperationMetricSpecification, self).__init__() - self.name = name - self.display_name = display_name - self.display_description = display_description - self.unit = unit - self.aggregation_type = aggregation_type - self.enable_regional_mdm_account = enable_regional_mdm_account - self.source_mdm_account = source_mdm_account - self.source_mdm_namespace = source_mdm_namespace - self.availabilities = availabilities + def __init__(self, **kwargs): + super(OperationMetricSpecification, self).__init__(**kwargs) + self.name = kwargs.get('name', None) + self.display_name = kwargs.get('display_name', None) + self.display_description = kwargs.get('display_description', None) + self.unit = kwargs.get('unit', None) + self.aggregation_type = kwargs.get('aggregation_type', None) + self.enable_regional_mdm_account = kwargs.get('enable_regional_mdm_account', None) + self.source_mdm_account = kwargs.get('source_mdm_account', None) + self.source_mdm_namespace = kwargs.get('source_mdm_namespace', None) + self.availabilities = kwargs.get('availabilities', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_metric_specification_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_metric_specification_py3.py new file mode 100644 index 000000000000..788106b93dfd --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_metric_specification_py3.py @@ -0,0 +1,63 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class OperationMetricSpecification(Model): + """Details about an operation related to metrics. + + :param name: The name of the metric. + :type name: str + :param display_name: Localized display name of the metric. + :type display_name: str + :param display_description: The description of the metric. + :type display_description: str + :param unit: The unit that the metric is measured in. + :type unit: str + :param aggregation_type: The type of metric aggregation. + :type aggregation_type: str + :param enable_regional_mdm_account: Whether or not the service is using + regional MDM accounts. + :type enable_regional_mdm_account: str + :param source_mdm_account: The name of the MDM account. + :type source_mdm_account: str + :param source_mdm_namespace: The name of the MDM namespace. + :type source_mdm_namespace: str + :param availabilities: Defines how often data for metrics becomes + available. + :type availabilities: + list[~azure.mgmt.datafactory.models.OperationMetricAvailability] + """ + + _attribute_map = { + 'name': {'key': 'name', 'type': 'str'}, + 'display_name': {'key': 'displayName', 'type': 'str'}, + 'display_description': {'key': 'displayDescription', 'type': 'str'}, + 'unit': {'key': 'unit', 'type': 'str'}, + 'aggregation_type': {'key': 'aggregationType', 'type': 'str'}, + 'enable_regional_mdm_account': {'key': 'enableRegionalMdmAccount', 'type': 'str'}, + 'source_mdm_account': {'key': 'sourceMdmAccount', 'type': 'str'}, + 'source_mdm_namespace': {'key': 'sourceMdmNamespace', 'type': 'str'}, + 'availabilities': {'key': 'availabilities', 'type': '[OperationMetricAvailability]'}, + } + + def __init__(self, *, name: str=None, display_name: str=None, display_description: str=None, unit: str=None, aggregation_type: str=None, enable_regional_mdm_account: str=None, source_mdm_account: str=None, source_mdm_namespace: str=None, availabilities=None, **kwargs) -> None: + super(OperationMetricSpecification, self).__init__(**kwargs) + self.name = name + self.display_name = display_name + self.display_description = display_description + self.unit = unit + self.aggregation_type = aggregation_type + self.enable_regional_mdm_account = enable_regional_mdm_account + self.source_mdm_account = source_mdm_account + self.source_mdm_namespace = source_mdm_namespace + self.availabilities = availabilities diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_py3.py new file mode 100644 index 000000000000..23305038a090 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_py3.py @@ -0,0 +1,41 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class Operation(Model): + """Azure Data Factory API operation definition. + + :param name: Operation name: {provider}/{resource}/{operation} + :type name: str + :param origin: The intended executor of the operation. + :type origin: str + :param display: Metadata associated with the operation. + :type display: ~azure.mgmt.datafactory.models.OperationDisplay + :param service_specification: Details about a service operation. + :type service_specification: + ~azure.mgmt.datafactory.models.OperationServiceSpecification + """ + + _attribute_map = { + 'name': {'key': 'name', 'type': 'str'}, + 'origin': {'key': 'origin', 'type': 'str'}, + 'display': {'key': 'display', 'type': 'OperationDisplay'}, + 'service_specification': {'key': 'properties.serviceSpecification', 'type': 'OperationServiceSpecification'}, + } + + def __init__(self, *, name: str=None, origin: str=None, display=None, service_specification=None, **kwargs) -> None: + super(Operation, self).__init__(**kwargs) + self.name = name + self.origin = origin + self.display = display + self.service_specification = service_specification diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_service_specification.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_service_specification.py index 26cac12bec97..82622a44af5a 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_service_specification.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_service_specification.py @@ -28,7 +28,7 @@ class OperationServiceSpecification(Model): 'metric_specifications': {'key': 'metricSpecifications', 'type': '[OperationMetricSpecification]'}, } - def __init__(self, log_specifications=None, metric_specifications=None): - super(OperationServiceSpecification, self).__init__() - self.log_specifications = log_specifications - self.metric_specifications = metric_specifications + def __init__(self, **kwargs): + super(OperationServiceSpecification, self).__init__(**kwargs) + self.log_specifications = kwargs.get('log_specifications', None) + self.metric_specifications = kwargs.get('metric_specifications', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_service_specification_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_service_specification_py3.py new file mode 100644 index 000000000000..4215dac6eb7f --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_service_specification_py3.py @@ -0,0 +1,34 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class OperationServiceSpecification(Model): + """Details about a service operation. + + :param log_specifications: Details about operations related to logs. + :type log_specifications: + list[~azure.mgmt.datafactory.models.OperationLogSpecification] + :param metric_specifications: Details about operations related to metrics. + :type metric_specifications: + list[~azure.mgmt.datafactory.models.OperationMetricSpecification] + """ + + _attribute_map = { + 'log_specifications': {'key': 'logSpecifications', 'type': '[OperationLogSpecification]'}, + 'metric_specifications': {'key': 'metricSpecifications', 'type': '[OperationMetricSpecification]'}, + } + + def __init__(self, *, log_specifications=None, metric_specifications=None, **kwargs) -> None: + super(OperationServiceSpecification, self).__init__(**kwargs) + self.log_specifications = log_specifications + self.metric_specifications = metric_specifications diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_linked_service.py index b7b0ce6a0230..9e78c4ef42c0 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_linked_service.py @@ -15,6 +15,8 @@ class OracleLinkedService(LinkedService): """Oracle database. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -29,9 +31,9 @@ class OracleLinkedService(LinkedService): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param connection_string: The connection string. + :param connection_string: Required. The connection string. :type connection_string: ~azure.mgmt.datafactory.models.SecretBase :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime @@ -55,8 +57,8 @@ class OracleLinkedService(LinkedService): 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, connection_string, additional_properties=None, connect_via=None, description=None, parameters=None, annotations=None, encrypted_credential=None): - super(OracleLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations) - self.connection_string = connection_string - self.encrypted_credential = encrypted_credential + def __init__(self, **kwargs): + super(OracleLinkedService, self).__init__(**kwargs) + self.connection_string = kwargs.get('connection_string', None) + self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'Oracle' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_linked_service_py3.py new file mode 100644 index 000000000000..4072d25d329c --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_linked_service_py3.py @@ -0,0 +1,64 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class OracleLinkedService(LinkedService): + """Oracle database. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param connection_string: Required. The connection string. + :type connection_string: ~azure.mgmt.datafactory.models.SecretBase + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'connection_string': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'SecretBase'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, *, connection_string, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, encrypted_credential=None, **kwargs) -> None: + super(OracleLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.connection_string = connection_string + self.encrypted_credential = encrypted_credential + self.type = 'Oracle' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_sink.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_sink.py index d8883d1d44ed..fa0e11f57553 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_sink.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_sink.py @@ -15,6 +15,8 @@ class OracleSink(CopySink): """A copy activity Oracle sink. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -32,7 +34,7 @@ class OracleSink(CopySink): resultType string), pattern: ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type sink_retry_wait: object - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param pre_copy_script: SQL pre-copy script. Type: string (or Expression with resultType string). @@ -53,7 +55,7 @@ class OracleSink(CopySink): 'pre_copy_script': {'key': 'preCopyScript', 'type': 'object'}, } - def __init__(self, additional_properties=None, write_batch_size=None, write_batch_timeout=None, sink_retry_count=None, sink_retry_wait=None, pre_copy_script=None): - super(OracleSink, self).__init__(additional_properties=additional_properties, write_batch_size=write_batch_size, write_batch_timeout=write_batch_timeout, sink_retry_count=sink_retry_count, sink_retry_wait=sink_retry_wait) - self.pre_copy_script = pre_copy_script + def __init__(self, **kwargs): + super(OracleSink, self).__init__(**kwargs) + self.pre_copy_script = kwargs.get('pre_copy_script', None) self.type = 'OracleSink' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_sink_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_sink_py3.py new file mode 100644 index 000000000000..07bd11f38775 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_sink_py3.py @@ -0,0 +1,61 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_sink import CopySink + + +class OracleSink(CopySink): + """A copy activity Oracle sink. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param write_batch_size: Write batch size. Type: integer (or Expression + with resultType integer), minimum: 0. + :type write_batch_size: object + :param write_batch_timeout: Write batch timeout. Type: string (or + Expression with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type write_batch_timeout: object + :param sink_retry_count: Sink retry count. Type: integer (or Expression + with resultType integer). + :type sink_retry_count: object + :param sink_retry_wait: Sink retry wait. Type: string (or Expression with + resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type sink_retry_wait: object + :param type: Required. Constant filled by server. + :type type: str + :param pre_copy_script: SQL pre-copy script. Type: string (or Expression + with resultType string). + :type pre_copy_script: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, + 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, + 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, + 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'pre_copy_script': {'key': 'preCopyScript', 'type': 'object'}, + } + + def __init__(self, *, additional_properties=None, write_batch_size=None, write_batch_timeout=None, sink_retry_count=None, sink_retry_wait=None, pre_copy_script=None, **kwargs) -> None: + super(OracleSink, self).__init__(additional_properties=additional_properties, write_batch_size=write_batch_size, write_batch_timeout=write_batch_timeout, sink_retry_count=sink_retry_count, sink_retry_wait=sink_retry_wait, **kwargs) + self.pre_copy_script = pre_copy_script + self.type = 'OracleSink' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_source.py index e1937a1f6b4a..3f74cf83ee7a 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_source.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_source.py @@ -15,6 +15,8 @@ class OracleSource(CopySource): """A copy activity Oracle source. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -25,7 +27,7 @@ class OracleSource(CopySource): with resultType string), pattern: ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type source_retry_wait: object - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param oracle_reader_query: Oracle reader query. Type: string (or Expression with resultType string). @@ -49,8 +51,8 @@ class OracleSource(CopySource): 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, } - def __init__(self, additional_properties=None, source_retry_count=None, source_retry_wait=None, oracle_reader_query=None, query_timeout=None): - super(OracleSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait) - self.oracle_reader_query = oracle_reader_query - self.query_timeout = query_timeout + def __init__(self, **kwargs): + super(OracleSource, self).__init__(**kwargs) + self.oracle_reader_query = kwargs.get('oracle_reader_query', None) + self.query_timeout = kwargs.get('query_timeout', None) self.type = 'OracleSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_source_py3.py new file mode 100644 index 000000000000..24af230164c5 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_source_py3.py @@ -0,0 +1,58 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class OracleSource(CopySource): + """A copy activity Oracle source. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Required. Constant filled by server. + :type type: str + :param oracle_reader_query: Oracle reader query. Type: string (or + Expression with resultType string). + :type oracle_reader_query: object + :param query_timeout: Query timeout. Type: string (or Expression with + resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type query_timeout: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'oracle_reader_query': {'key': 'oracleReaderQuery', 'type': 'object'}, + 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, + } + + def __init__(self, *, additional_properties=None, source_retry_count=None, source_retry_wait=None, oracle_reader_query=None, query_timeout=None, **kwargs) -> None: + super(OracleSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, **kwargs) + self.oracle_reader_query = oracle_reader_query + self.query_timeout = query_timeout + self.type = 'OracleSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_table_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_table_dataset.py index b9e81ec44959..78215f783ae3 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_table_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_table_dataset.py @@ -15,6 +15,8 @@ class OracleTableDataset(Dataset): """The on-premises Oracle database dataset. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -23,7 +25,7 @@ class OracleTableDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object - :param linked_service_name: Linked service reference. + :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference :param parameters: Parameters for dataset. @@ -32,10 +34,10 @@ class OracleTableDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param table_name: The table name of the on-premises Oracle database. - Type: string (or Expression with resultType string). + :param table_name: Required. The table name of the on-premises Oracle + database. Type: string (or Expression with resultType string). :type table_name: object """ @@ -56,7 +58,7 @@ class OracleTableDataset(Dataset): 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, linked_service_name, table_name, additional_properties=None, description=None, structure=None, parameters=None, annotations=None): - super(OracleTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations) - self.table_name = table_name + def __init__(self, **kwargs): + super(OracleTableDataset, self).__init__(**kwargs) + self.table_name = kwargs.get('table_name', None) self.type = 'OracleTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_table_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_table_dataset_py3.py new file mode 100644 index 000000000000..2737f7df09a3 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_table_dataset_py3.py @@ -0,0 +1,64 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class OracleTableDataset(Dataset): + """The on-premises Oracle database dataset. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Required. Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param table_name: Required. The table name of the on-premises Oracle + database. Type: string (or Expression with resultType string). + :type table_name: object + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + 'table_name': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, + } + + def __init__(self, *, linked_service_name, table_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: + super(OracleTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + self.table_name = table_name + self.type = 'OracleTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/orc_format.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/orc_format.py index b91f4d5952f7..8f0a0322062c 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/orc_format.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/orc_format.py @@ -15,6 +15,8 @@ class OrcFormat(DatasetStorageFormat): """The data stored in Optimized Row Columnar (ORC) format. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -24,7 +26,7 @@ class OrcFormat(DatasetStorageFormat): :param deserializer: Deserializer. Type: string (or Expression with resultType string). :type deserializer: object - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str """ @@ -32,6 +34,13 @@ class OrcFormat(DatasetStorageFormat): 'type': {'required': True}, } - def __init__(self, additional_properties=None, serializer=None, deserializer=None): - super(OrcFormat, self).__init__(additional_properties=additional_properties, serializer=serializer, deserializer=deserializer) + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'serializer': {'key': 'serializer', 'type': 'object'}, + 'deserializer': {'key': 'deserializer', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, **kwargs): + super(OrcFormat, self).__init__(**kwargs) self.type = 'OrcFormat' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/orc_format_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/orc_format_py3.py new file mode 100644 index 000000000000..17db530edc64 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/orc_format_py3.py @@ -0,0 +1,46 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset_storage_format import DatasetStorageFormat + + +class OrcFormat(DatasetStorageFormat): + """The data stored in Optimized Row Columnar (ORC) format. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param serializer: Serializer. Type: string (or Expression with resultType + string). + :type serializer: object + :param deserializer: Deserializer. Type: string (or Expression with + resultType string). + :type deserializer: object + :param type: Required. Constant filled by server. + :type type: str + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'serializer': {'key': 'serializer', 'type': 'object'}, + 'deserializer': {'key': 'deserializer', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, *, additional_properties=None, serializer=None, deserializer=None, **kwargs) -> None: + super(OrcFormat, self).__init__(additional_properties=additional_properties, serializer=serializer, deserializer=deserializer, **kwargs) + self.type = 'OrcFormat' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/parameter_specification.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/parameter_specification.py index 5747dc8efedf..aef855d955f0 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/parameter_specification.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/parameter_specification.py @@ -15,8 +15,10 @@ class ParameterSpecification(Model): """Definition of a single parameter for an entity. - :param type: Parameter type. Possible values include: 'Object', 'String', - 'Int', 'Float', 'Bool', 'Array', 'SecureString' + All required parameters must be populated in order to send to Azure. + + :param type: Required. Parameter type. Possible values include: 'Object', + 'String', 'Int', 'Float', 'Bool', 'Array', 'SecureString' :type type: str or ~azure.mgmt.datafactory.models.ParameterType :param default_value: Default value of parameter. :type default_value: object @@ -31,7 +33,7 @@ class ParameterSpecification(Model): 'default_value': {'key': 'defaultValue', 'type': 'object'}, } - def __init__(self, type, default_value=None): - super(ParameterSpecification, self).__init__() - self.type = type - self.default_value = default_value + def __init__(self, **kwargs): + super(ParameterSpecification, self).__init__(**kwargs) + self.type = kwargs.get('type', None) + self.default_value = kwargs.get('default_value', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/parameter_specification_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/parameter_specification_py3.py new file mode 100644 index 000000000000..d5b6f981d365 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/parameter_specification_py3.py @@ -0,0 +1,39 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class ParameterSpecification(Model): + """Definition of a single parameter for an entity. + + All required parameters must be populated in order to send to Azure. + + :param type: Required. Parameter type. Possible values include: 'Object', + 'String', 'Int', 'Float', 'Bool', 'Array', 'SecureString' + :type type: str or ~azure.mgmt.datafactory.models.ParameterType + :param default_value: Default value of parameter. + :type default_value: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'type': {'key': 'type', 'type': 'str'}, + 'default_value': {'key': 'defaultValue', 'type': 'object'}, + } + + def __init__(self, *, type, default_value=None, **kwargs) -> None: + super(ParameterSpecification, self).__init__(**kwargs) + self.type = type + self.default_value = default_value diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/parquet_format.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/parquet_format.py index 4d953b3a4e78..d742ff24b522 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/parquet_format.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/parquet_format.py @@ -15,6 +15,8 @@ class ParquetFormat(DatasetStorageFormat): """The data stored in Parquet format. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -24,7 +26,7 @@ class ParquetFormat(DatasetStorageFormat): :param deserializer: Deserializer. Type: string (or Expression with resultType string). :type deserializer: object - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str """ @@ -32,6 +34,13 @@ class ParquetFormat(DatasetStorageFormat): 'type': {'required': True}, } - def __init__(self, additional_properties=None, serializer=None, deserializer=None): - super(ParquetFormat, self).__init__(additional_properties=additional_properties, serializer=serializer, deserializer=deserializer) + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'serializer': {'key': 'serializer', 'type': 'object'}, + 'deserializer': {'key': 'deserializer', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, **kwargs): + super(ParquetFormat, self).__init__(**kwargs) self.type = 'ParquetFormat' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/parquet_format_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/parquet_format_py3.py new file mode 100644 index 000000000000..ba2b2428f140 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/parquet_format_py3.py @@ -0,0 +1,46 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset_storage_format import DatasetStorageFormat + + +class ParquetFormat(DatasetStorageFormat): + """The data stored in Parquet format. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param serializer: Serializer. Type: string (or Expression with resultType + string). + :type serializer: object + :param deserializer: Deserializer. Type: string (or Expression with + resultType string). + :type deserializer: object + :param type: Required. Constant filled by server. + :type type: str + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'serializer': {'key': 'serializer', 'type': 'object'}, + 'deserializer': {'key': 'deserializer', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, *, additional_properties=None, serializer=None, deserializer=None, **kwargs) -> None: + super(ParquetFormat, self).__init__(additional_properties=additional_properties, serializer=serializer, deserializer=deserializer, **kwargs) + self.type = 'ParquetFormat' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_linked_service.py index c59f7146d169..fa7ba90e1e77 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_linked_service.py @@ -15,6 +15,8 @@ class PaypalLinkedService(LinkedService): """Paypal Serivce linked service. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -29,11 +31,13 @@ class PaypalLinkedService(LinkedService): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param host: The URL of the PayPal instance. (i.e. api.sandbox.paypal.com) + :param host: Required. The URL of the PayPal instance. (i.e. + api.sandbox.paypal.com) :type host: object - :param client_id: The client ID associated with your PayPal application. + :param client_id: Required. The client ID associated with your PayPal + application. :type client_id: object :param client_secret: The client secret associated with your PayPal application. @@ -76,13 +80,13 @@ class PaypalLinkedService(LinkedService): 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, host, client_id, additional_properties=None, connect_via=None, description=None, parameters=None, annotations=None, client_secret=None, use_encrypted_endpoints=None, use_host_verification=None, use_peer_verification=None, encrypted_credential=None): - super(PaypalLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations) - self.host = host - self.client_id = client_id - self.client_secret = client_secret - self.use_encrypted_endpoints = use_encrypted_endpoints - self.use_host_verification = use_host_verification - self.use_peer_verification = use_peer_verification - self.encrypted_credential = encrypted_credential + def __init__(self, **kwargs): + super(PaypalLinkedService, self).__init__(**kwargs) + self.host = kwargs.get('host', None) + self.client_id = kwargs.get('client_id', None) + self.client_secret = kwargs.get('client_secret', None) + self.use_encrypted_endpoints = kwargs.get('use_encrypted_endpoints', None) + self.use_host_verification = kwargs.get('use_host_verification', None) + self.use_peer_verification = kwargs.get('use_peer_verification', None) + self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'Paypal' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_linked_service_py3.py new file mode 100644 index 000000000000..39c83d422068 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_linked_service_py3.py @@ -0,0 +1,92 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class PaypalLinkedService(LinkedService): + """Paypal Serivce linked service. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param host: Required. The URL of the PayPal instance. (i.e. + api.sandbox.paypal.com) + :type host: object + :param client_id: Required. The client ID associated with your PayPal + application. + :type client_id: object + :param client_secret: The client secret associated with your PayPal + application. + :type client_secret: ~azure.mgmt.datafactory.models.SecretBase + :param use_encrypted_endpoints: Specifies whether the data source + endpoints are encrypted using HTTPS. The default value is true. + :type use_encrypted_endpoints: object + :param use_host_verification: Specifies whether to require the host name + in the server's certificate to match the host name of the server when + connecting over SSL. The default value is true. + :type use_host_verification: object + :param use_peer_verification: Specifies whether to verify the identity of + the server when connecting over SSL. The default value is true. + :type use_peer_verification: object + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'host': {'required': True}, + 'client_id': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'host': {'key': 'typeProperties.host', 'type': 'object'}, + 'client_id': {'key': 'typeProperties.clientId', 'type': 'object'}, + 'client_secret': {'key': 'typeProperties.clientSecret', 'type': 'SecretBase'}, + 'use_encrypted_endpoints': {'key': 'typeProperties.useEncryptedEndpoints', 'type': 'object'}, + 'use_host_verification': {'key': 'typeProperties.useHostVerification', 'type': 'object'}, + 'use_peer_verification': {'key': 'typeProperties.usePeerVerification', 'type': 'object'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, *, host, client_id, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, client_secret=None, use_encrypted_endpoints=None, use_host_verification=None, use_peer_verification=None, encrypted_credential=None, **kwargs) -> None: + super(PaypalLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.host = host + self.client_id = client_id + self.client_secret = client_secret + self.use_encrypted_endpoints = use_encrypted_endpoints + self.use_host_verification = use_host_verification + self.use_peer_verification = use_peer_verification + self.encrypted_credential = encrypted_credential + self.type = 'Paypal' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_object_dataset.py index d65a546a2006..6d48d7fb99c3 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_object_dataset.py @@ -15,6 +15,8 @@ class PaypalObjectDataset(Dataset): """Paypal Serivce dataset. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -23,7 +25,7 @@ class PaypalObjectDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object - :param linked_service_name: Linked service reference. + :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference :param parameters: Parameters for dataset. @@ -32,7 +34,7 @@ class PaypalObjectDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str """ @@ -41,6 +43,16 @@ class PaypalObjectDataset(Dataset): 'type': {'required': True}, } - def __init__(self, linked_service_name, additional_properties=None, description=None, structure=None, parameters=None, annotations=None): - super(PaypalObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations) + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, **kwargs): + super(PaypalObjectDataset, self).__init__(**kwargs) self.type = 'PaypalObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_object_dataset_py3.py new file mode 100644 index 000000000000..475c579602a7 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_object_dataset_py3.py @@ -0,0 +1,58 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class PaypalObjectDataset(Dataset): + """Paypal Serivce dataset. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Required. Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: + super(PaypalObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + self.type = 'PaypalObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_source.py index 1cc34f2decfa..61f89f28bc59 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_source.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_source.py @@ -15,6 +15,8 @@ class PaypalSource(CopySource): """A copy activity Paypal Serivce source. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -25,7 +27,7 @@ class PaypalSource(CopySource): with resultType string), pattern: ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type source_retry_wait: object - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). @@ -44,7 +46,7 @@ class PaypalSource(CopySource): 'query': {'key': 'query', 'type': 'object'}, } - def __init__(self, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None): - super(PaypalSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait) - self.query = query + def __init__(self, **kwargs): + super(PaypalSource, self).__init__(**kwargs) + self.query = kwargs.get('query', None) self.type = 'PaypalSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_source_py3.py new file mode 100644 index 000000000000..9d6d7e6f30fa --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_source_py3.py @@ -0,0 +1,52 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class PaypalSource(CopySource): + """A copy activity Paypal Serivce source. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Required. Constant filled by server. + :type type: str + :param query: A query to retrieve data from source. Type: string (or + Expression with resultType string). + :type query: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'query': {'key': 'query', 'type': 'object'}, + } + + def __init__(self, *, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None, **kwargs) -> None: + super(PaypalSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, **kwargs) + self.query = query + self.type = 'PaypalSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/phoenix_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/phoenix_linked_service.py index e2ffa8ca2ccd..b9d16bc32c56 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/phoenix_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/phoenix_linked_service.py @@ -15,6 +15,8 @@ class PhoenixLinkedService(LinkedService): """Phoenix server linked service. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -29,10 +31,10 @@ class PhoenixLinkedService(LinkedService): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param host: The IP address or host name of the Phoenix server. (i.e. - 192.168.222.160) + :param host: Required. The IP address or host name of the Phoenix server. + (i.e. 192.168.222.160) :type host: object :param port: The TCP port that the Phoenix server uses to listen for client connections. The default value is 8765. @@ -41,8 +43,8 @@ class PhoenixLinkedService(LinkedService): (i.e. /gateway/sandbox/phoenix/version). The default value is hbasephoenix if using WindowsAzureHDInsightService. :type http_path: object - :param authentication_type: The authentication mechanism used to connect - to the Phoenix server. Possible values include: 'Anonymous', + :param authentication_type: Required. The authentication mechanism used to + connect to the Phoenix server. Possible values include: 'Anonymous', 'UsernameAndPassword', 'WindowsAzureHDInsightService' :type authentication_type: str or ~azure.mgmt.datafactory.models.PhoenixAuthenticationType @@ -102,18 +104,18 @@ class PhoenixLinkedService(LinkedService): 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, host, authentication_type, additional_properties=None, connect_via=None, description=None, parameters=None, annotations=None, port=None, http_path=None, username=None, password=None, enable_ssl=None, trusted_cert_path=None, use_system_trust_store=None, allow_host_name_cn_mismatch=None, allow_self_signed_server_cert=None, encrypted_credential=None): - super(PhoenixLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations) - self.host = host - self.port = port - self.http_path = http_path - self.authentication_type = authentication_type - self.username = username - self.password = password - self.enable_ssl = enable_ssl - self.trusted_cert_path = trusted_cert_path - self.use_system_trust_store = use_system_trust_store - self.allow_host_name_cn_mismatch = allow_host_name_cn_mismatch - self.allow_self_signed_server_cert = allow_self_signed_server_cert - self.encrypted_credential = encrypted_credential + def __init__(self, **kwargs): + super(PhoenixLinkedService, self).__init__(**kwargs) + self.host = kwargs.get('host', None) + self.port = kwargs.get('port', None) + self.http_path = kwargs.get('http_path', None) + self.authentication_type = kwargs.get('authentication_type', None) + self.username = kwargs.get('username', None) + self.password = kwargs.get('password', None) + self.enable_ssl = kwargs.get('enable_ssl', None) + self.trusted_cert_path = kwargs.get('trusted_cert_path', None) + self.use_system_trust_store = kwargs.get('use_system_trust_store', None) + self.allow_host_name_cn_mismatch = kwargs.get('allow_host_name_cn_mismatch', None) + self.allow_self_signed_server_cert = kwargs.get('allow_self_signed_server_cert', None) + self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'Phoenix' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/phoenix_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/phoenix_linked_service_py3.py new file mode 100644 index 000000000000..fe18934da90a --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/phoenix_linked_service_py3.py @@ -0,0 +1,121 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class PhoenixLinkedService(LinkedService): + """Phoenix server linked service. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param host: Required. The IP address or host name of the Phoenix server. + (i.e. 192.168.222.160) + :type host: object + :param port: The TCP port that the Phoenix server uses to listen for + client connections. The default value is 8765. + :type port: object + :param http_path: The partial URL corresponding to the Phoenix server. + (i.e. /gateway/sandbox/phoenix/version). The default value is hbasephoenix + if using WindowsAzureHDInsightService. + :type http_path: object + :param authentication_type: Required. The authentication mechanism used to + connect to the Phoenix server. Possible values include: 'Anonymous', + 'UsernameAndPassword', 'WindowsAzureHDInsightService' + :type authentication_type: str or + ~azure.mgmt.datafactory.models.PhoenixAuthenticationType + :param username: The user name used to connect to the Phoenix server. + :type username: object + :param password: The password corresponding to the user name. + :type password: ~azure.mgmt.datafactory.models.SecretBase + :param enable_ssl: Specifies whether the connections to the server are + encrypted using SSL. The default value is false. + :type enable_ssl: object + :param trusted_cert_path: The full path of the .pem file containing + trusted CA certificates for verifying the server when connecting over SSL. + This property can only be set when using SSL on self-hosted IR. The + default value is the cacerts.pem file installed with the IR. + :type trusted_cert_path: object + :param use_system_trust_store: Specifies whether to use a CA certificate + from the system trust store or from a specified PEM file. The default + value is false. + :type use_system_trust_store: object + :param allow_host_name_cn_mismatch: Specifies whether to require a + CA-issued SSL certificate name to match the host name of the server when + connecting over SSL. The default value is false. + :type allow_host_name_cn_mismatch: object + :param allow_self_signed_server_cert: Specifies whether to allow + self-signed certificates from the server. The default value is false. + :type allow_self_signed_server_cert: object + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'host': {'required': True}, + 'authentication_type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'host': {'key': 'typeProperties.host', 'type': 'object'}, + 'port': {'key': 'typeProperties.port', 'type': 'object'}, + 'http_path': {'key': 'typeProperties.httpPath', 'type': 'object'}, + 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, + 'username': {'key': 'typeProperties.username', 'type': 'object'}, + 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, + 'enable_ssl': {'key': 'typeProperties.enableSsl', 'type': 'object'}, + 'trusted_cert_path': {'key': 'typeProperties.trustedCertPath', 'type': 'object'}, + 'use_system_trust_store': {'key': 'typeProperties.useSystemTrustStore', 'type': 'object'}, + 'allow_host_name_cn_mismatch': {'key': 'typeProperties.allowHostNameCNMismatch', 'type': 'object'}, + 'allow_self_signed_server_cert': {'key': 'typeProperties.allowSelfSignedServerCert', 'type': 'object'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, *, host, authentication_type, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, port=None, http_path=None, username=None, password=None, enable_ssl=None, trusted_cert_path=None, use_system_trust_store=None, allow_host_name_cn_mismatch=None, allow_self_signed_server_cert=None, encrypted_credential=None, **kwargs) -> None: + super(PhoenixLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.host = host + self.port = port + self.http_path = http_path + self.authentication_type = authentication_type + self.username = username + self.password = password + self.enable_ssl = enable_ssl + self.trusted_cert_path = trusted_cert_path + self.use_system_trust_store = use_system_trust_store + self.allow_host_name_cn_mismatch = allow_host_name_cn_mismatch + self.allow_self_signed_server_cert = allow_self_signed_server_cert + self.encrypted_credential = encrypted_credential + self.type = 'Phoenix' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/phoenix_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/phoenix_object_dataset.py index c6b7e51b75b6..b1cd7cf37049 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/phoenix_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/phoenix_object_dataset.py @@ -15,6 +15,8 @@ class PhoenixObjectDataset(Dataset): """Phoenix server dataset. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -23,7 +25,7 @@ class PhoenixObjectDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object - :param linked_service_name: Linked service reference. + :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference :param parameters: Parameters for dataset. @@ -32,7 +34,7 @@ class PhoenixObjectDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str """ @@ -41,6 +43,16 @@ class PhoenixObjectDataset(Dataset): 'type': {'required': True}, } - def __init__(self, linked_service_name, additional_properties=None, description=None, structure=None, parameters=None, annotations=None): - super(PhoenixObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations) + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, **kwargs): + super(PhoenixObjectDataset, self).__init__(**kwargs) self.type = 'PhoenixObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/phoenix_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/phoenix_object_dataset_py3.py new file mode 100644 index 000000000000..a279a7763bd6 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/phoenix_object_dataset_py3.py @@ -0,0 +1,58 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class PhoenixObjectDataset(Dataset): + """Phoenix server dataset. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Required. Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: + super(PhoenixObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + self.type = 'PhoenixObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/phoenix_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/phoenix_source.py index 670bd7ddca85..daad6ec41c31 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/phoenix_source.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/phoenix_source.py @@ -15,6 +15,8 @@ class PhoenixSource(CopySource): """A copy activity Phoenix server source. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -25,7 +27,7 @@ class PhoenixSource(CopySource): with resultType string), pattern: ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type source_retry_wait: object - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). @@ -44,7 +46,7 @@ class PhoenixSource(CopySource): 'query': {'key': 'query', 'type': 'object'}, } - def __init__(self, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None): - super(PhoenixSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait) - self.query = query + def __init__(self, **kwargs): + super(PhoenixSource, self).__init__(**kwargs) + self.query = kwargs.get('query', None) self.type = 'PhoenixSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/phoenix_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/phoenix_source_py3.py new file mode 100644 index 000000000000..6497d3dfc80e --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/phoenix_source_py3.py @@ -0,0 +1,52 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class PhoenixSource(CopySource): + """A copy activity Phoenix server source. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Required. Constant filled by server. + :type type: str + :param query: A query to retrieve data from source. Type: string (or + Expression with resultType string). + :type query: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'query': {'key': 'query', 'type': 'object'}, + } + + def __init__(self, *, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None, **kwargs) -> None: + super(PhoenixSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, **kwargs) + self.query = query + self.type = 'PhoenixSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_reference.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_reference.py index 1d39beea8145..aa8b23e62932 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_reference.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_reference.py @@ -18,9 +18,12 @@ class PipelineReference(Model): Variables are only populated by the server, and will be ignored when sending a request. - :ivar type: Pipeline reference type. Default value: "PipelineReference" . + All required parameters must be populated in order to send to Azure. + + :ivar type: Required. Pipeline reference type. Default value: + "PipelineReference" . :vartype type: str - :param reference_name: Reference pipeline name. + :param reference_name: Required. Reference pipeline name. :type reference_name: str :param name: Reference name. :type name: str @@ -39,7 +42,7 @@ class PipelineReference(Model): type = "PipelineReference" - def __init__(self, reference_name, name=None): - super(PipelineReference, self).__init__() - self.reference_name = reference_name - self.name = name + def __init__(self, **kwargs): + super(PipelineReference, self).__init__(**kwargs) + self.reference_name = kwargs.get('reference_name', None) + self.name = kwargs.get('name', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_reference_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_reference_py3.py new file mode 100644 index 000000000000..ce63f06092d1 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_reference_py3.py @@ -0,0 +1,48 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class PipelineReference(Model): + """Pipeline reference type. + + Variables are only populated by the server, and will be ignored when + sending a request. + + All required parameters must be populated in order to send to Azure. + + :ivar type: Required. Pipeline reference type. Default value: + "PipelineReference" . + :vartype type: str + :param reference_name: Required. Reference pipeline name. + :type reference_name: str + :param name: Reference name. + :type name: str + """ + + _validation = { + 'type': {'required': True, 'constant': True}, + 'reference_name': {'required': True}, + } + + _attribute_map = { + 'type': {'key': 'type', 'type': 'str'}, + 'reference_name': {'key': 'referenceName', 'type': 'str'}, + 'name': {'key': 'name', 'type': 'str'}, + } + + type = "PipelineReference" + + def __init__(self, *, reference_name: str, name: str=None, **kwargs) -> None: + super(PipelineReference, self).__init__(**kwargs) + self.reference_name = reference_name + self.name = name diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_resource.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_resource.py index dedd0c1af9e6..2531c35a3838 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_resource.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_resource.py @@ -64,11 +64,11 @@ class PipelineResource(SubResource): 'annotations': {'key': 'properties.annotations', 'type': '[object]'}, } - def __init__(self, additional_properties=None, description=None, activities=None, parameters=None, concurrency=None, annotations=None): - super(PipelineResource, self).__init__() - self.additional_properties = additional_properties - self.description = description - self.activities = activities - self.parameters = parameters - self.concurrency = concurrency - self.annotations = annotations + def __init__(self, **kwargs): + super(PipelineResource, self).__init__(**kwargs) + self.additional_properties = kwargs.get('additional_properties', None) + self.description = kwargs.get('description', None) + self.activities = kwargs.get('activities', None) + self.parameters = kwargs.get('parameters', None) + self.concurrency = kwargs.get('concurrency', None) + self.annotations = kwargs.get('annotations', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_resource_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_resource_py3.py new file mode 100644 index 000000000000..4da049f1f3b5 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_resource_py3.py @@ -0,0 +1,74 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .sub_resource import SubResource + + +class PipelineResource(SubResource): + """Pipeline resource type. + + Variables are only populated by the server, and will be ignored when + sending a request. + + :ivar id: The resource identifier. + :vartype id: str + :ivar name: The resource name. + :vartype name: str + :ivar type: The resource type. + :vartype type: str + :ivar etag: Etag identifies change in the resource. + :vartype etag: str + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param description: The description of the pipeline. + :type description: str + :param activities: List of activities in pipeline. + :type activities: list[~azure.mgmt.datafactory.models.Activity] + :param parameters: List of parameters for pipeline. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param concurrency: The max number of concurrent runs for the pipeline. + :type concurrency: int + :param annotations: List of tags that can be used for describing the + Pipeline. + :type annotations: list[object] + """ + + _validation = { + 'id': {'readonly': True}, + 'name': {'readonly': True}, + 'type': {'readonly': True}, + 'etag': {'readonly': True}, + 'concurrency': {'minimum': 1}, + } + + _attribute_map = { + 'id': {'key': 'id', 'type': 'str'}, + 'name': {'key': 'name', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'str'}, + 'etag': {'key': 'etag', 'type': 'str'}, + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'properties.description', 'type': 'str'}, + 'activities': {'key': 'properties.activities', 'type': '[Activity]'}, + 'parameters': {'key': 'properties.parameters', 'type': '{ParameterSpecification}'}, + 'concurrency': {'key': 'properties.concurrency', 'type': 'int'}, + 'annotations': {'key': 'properties.annotations', 'type': '[object]'}, + } + + def __init__(self, *, additional_properties=None, description: str=None, activities=None, parameters=None, concurrency: int=None, annotations=None, **kwargs) -> None: + super(PipelineResource, self).__init__(, **kwargs) + self.additional_properties = additional_properties + self.description = description + self.activities = activities + self.parameters = parameters + self.concurrency = concurrency + self.annotations = annotations diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run.py index 2939bb2762d9..3ae4beb48ff1 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run.py @@ -72,9 +72,9 @@ class PipelineRun(Model): 'message': {'key': 'message', 'type': 'str'}, } - def __init__(self, additional_properties=None): - super(PipelineRun, self).__init__() - self.additional_properties = additional_properties + def __init__(self, **kwargs): + super(PipelineRun, self).__init__(**kwargs) + self.additional_properties = kwargs.get('additional_properties', None) self.run_id = None self.pipeline_name = None self.parameters = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run_filter_parameters.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run_filter_parameters.py index 5c0b83f4c8b0..5e1341649f19 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run_filter_parameters.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run_filter_parameters.py @@ -15,14 +15,16 @@ class PipelineRunFilterParameters(Model): """Query parameters for listing pipeline runs. + All required parameters must be populated in order to send to Azure. + :param continuation_token: The continuation token for getting the next page of results. Null for first page. :type continuation_token: str - :param last_updated_after: The time at or after which the pipeline run - event was updated in 'ISO 8601' format. + :param last_updated_after: Required. The time at or after which the + pipeline run event was updated in 'ISO 8601' format. :type last_updated_after: datetime - :param last_updated_before: The time at or before which the pipeline run - event was updated in 'ISO 8601' format. + :param last_updated_before: Required. The time at or before which the + pipeline run event was updated in 'ISO 8601' format. :type last_updated_before: datetime :param filters: List of filters. :type filters: list[~azure.mgmt.datafactory.models.PipelineRunQueryFilter] @@ -44,10 +46,10 @@ class PipelineRunFilterParameters(Model): 'order_by': {'key': 'orderBy', 'type': '[PipelineRunQueryOrderBy]'}, } - def __init__(self, last_updated_after, last_updated_before, continuation_token=None, filters=None, order_by=None): - super(PipelineRunFilterParameters, self).__init__() - self.continuation_token = continuation_token - self.last_updated_after = last_updated_after - self.last_updated_before = last_updated_before - self.filters = filters - self.order_by = order_by + def __init__(self, **kwargs): + super(PipelineRunFilterParameters, self).__init__(**kwargs) + self.continuation_token = kwargs.get('continuation_token', None) + self.last_updated_after = kwargs.get('last_updated_after', None) + self.last_updated_before = kwargs.get('last_updated_before', None) + self.filters = kwargs.get('filters', None) + self.order_by = kwargs.get('order_by', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run_filter_parameters_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run_filter_parameters_py3.py new file mode 100644 index 000000000000..a436b8e4023c --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run_filter_parameters_py3.py @@ -0,0 +1,55 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class PipelineRunFilterParameters(Model): + """Query parameters for listing pipeline runs. + + All required parameters must be populated in order to send to Azure. + + :param continuation_token: The continuation token for getting the next + page of results. Null for first page. + :type continuation_token: str + :param last_updated_after: Required. The time at or after which the + pipeline run event was updated in 'ISO 8601' format. + :type last_updated_after: datetime + :param last_updated_before: Required. The time at or before which the + pipeline run event was updated in 'ISO 8601' format. + :type last_updated_before: datetime + :param filters: List of filters. + :type filters: list[~azure.mgmt.datafactory.models.PipelineRunQueryFilter] + :param order_by: List of OrderBy option. + :type order_by: + list[~azure.mgmt.datafactory.models.PipelineRunQueryOrderBy] + """ + + _validation = { + 'last_updated_after': {'required': True}, + 'last_updated_before': {'required': True}, + } + + _attribute_map = { + 'continuation_token': {'key': 'continuationToken', 'type': 'str'}, + 'last_updated_after': {'key': 'lastUpdatedAfter', 'type': 'iso-8601'}, + 'last_updated_before': {'key': 'lastUpdatedBefore', 'type': 'iso-8601'}, + 'filters': {'key': 'filters', 'type': '[PipelineRunQueryFilter]'}, + 'order_by': {'key': 'orderBy', 'type': '[PipelineRunQueryOrderBy]'}, + } + + def __init__(self, *, last_updated_after, last_updated_before, continuation_token: str=None, filters=None, order_by=None, **kwargs) -> None: + super(PipelineRunFilterParameters, self).__init__(**kwargs) + self.continuation_token = continuation_token + self.last_updated_after = last_updated_after + self.last_updated_before = last_updated_before + self.filters = filters + self.order_by = order_by diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run_invoked_by.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run_invoked_by.py index 9f4336505173..c43b642159d8 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run_invoked_by.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run_invoked_by.py @@ -34,7 +34,7 @@ class PipelineRunInvokedBy(Model): 'id': {'key': 'id', 'type': 'str'}, } - def __init__(self): - super(PipelineRunInvokedBy, self).__init__() + def __init__(self, **kwargs): + super(PipelineRunInvokedBy, self).__init__(**kwargs) self.name = None self.id = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run_invoked_by_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run_invoked_by_py3.py new file mode 100644 index 000000000000..d7540cdd0b0f --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run_invoked_by_py3.py @@ -0,0 +1,40 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class PipelineRunInvokedBy(Model): + """Provides entity name and id that started the pipeline run. + + Variables are only populated by the server, and will be ignored when + sending a request. + + :ivar name: Name of the entity that started the pipeline run. + :vartype name: str + :ivar id: The ID of the entity that started the run. + :vartype id: str + """ + + _validation = { + 'name': {'readonly': True}, + 'id': {'readonly': True}, + } + + _attribute_map = { + 'name': {'key': 'name', 'type': 'str'}, + 'id': {'key': 'id', 'type': 'str'}, + } + + def __init__(self, **kwargs) -> None: + super(PipelineRunInvokedBy, self).__init__(**kwargs) + self.name = None + self.id = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run_py3.py new file mode 100644 index 000000000000..aed5dd0466d2 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run_py3.py @@ -0,0 +1,87 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class PipelineRun(Model): + """Information about a pipeline run. + + Variables are only populated by the server, and will be ignored when + sending a request. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :ivar run_id: Identifier of a run. + :vartype run_id: str + :ivar pipeline_name: The pipeline name. + :vartype pipeline_name: str + :ivar parameters: The full or partial list of parameter name, value pair + used in the pipeline run. + :vartype parameters: dict[str, str] + :ivar invoked_by: Entity that started the pipeline run. + :vartype invoked_by: ~azure.mgmt.datafactory.models.PipelineRunInvokedBy + :ivar last_updated: The last updated timestamp for the pipeline run event + in ISO8601 format. + :vartype last_updated: datetime + :ivar run_start: The start time of a pipeline run in ISO8601 format. + :vartype run_start: datetime + :ivar run_end: The end time of a pipeline run in ISO8601 format. + :vartype run_end: datetime + :ivar duration_in_ms: The duration of a pipeline run. + :vartype duration_in_ms: int + :ivar status: The status of a pipeline run. + :vartype status: str + :ivar message: The message from a pipeline run. + :vartype message: str + """ + + _validation = { + 'run_id': {'readonly': True}, + 'pipeline_name': {'readonly': True}, + 'parameters': {'readonly': True}, + 'invoked_by': {'readonly': True}, + 'last_updated': {'readonly': True}, + 'run_start': {'readonly': True}, + 'run_end': {'readonly': True}, + 'duration_in_ms': {'readonly': True}, + 'status': {'readonly': True}, + 'message': {'readonly': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'run_id': {'key': 'runId', 'type': 'str'}, + 'pipeline_name': {'key': 'pipelineName', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{str}'}, + 'invoked_by': {'key': 'invokedBy', 'type': 'PipelineRunInvokedBy'}, + 'last_updated': {'key': 'lastUpdated', 'type': 'iso-8601'}, + 'run_start': {'key': 'runStart', 'type': 'iso-8601'}, + 'run_end': {'key': 'runEnd', 'type': 'iso-8601'}, + 'duration_in_ms': {'key': 'durationInMs', 'type': 'int'}, + 'status': {'key': 'status', 'type': 'str'}, + 'message': {'key': 'message', 'type': 'str'}, + } + + def __init__(self, *, additional_properties=None, **kwargs) -> None: + super(PipelineRun, self).__init__(**kwargs) + self.additional_properties = additional_properties + self.run_id = None + self.pipeline_name = None + self.parameters = None + self.invoked_by = None + self.last_updated = None + self.run_start = None + self.run_end = None + self.duration_in_ms = None + self.status = None + self.message = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run_query_filter.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run_query_filter.py index ba65ea65df00..bca4438a6ed6 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run_query_filter.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run_query_filter.py @@ -15,15 +15,17 @@ class PipelineRunQueryFilter(Model): """Query filter option for listing pipeline runs. - :param operand: Parameter name to be used for filter. Possible values - include: 'PipelineName', 'Status', 'RunStart', 'RunEnd' + All required parameters must be populated in order to send to Azure. + + :param operand: Required. Parameter name to be used for filter. Possible + values include: 'PipelineName', 'Status', 'RunStart', 'RunEnd' :type operand: str or ~azure.mgmt.datafactory.models.PipelineRunQueryFilterOperand - :param operator: Operator to be used for filter. Possible values include: - 'Equals', 'NotEquals', 'In', 'NotIn' + :param operator: Required. Operator to be used for filter. Possible values + include: 'Equals', 'NotEquals', 'In', 'NotIn' :type operator: str or ~azure.mgmt.datafactory.models.PipelineRunQueryFilterOperator - :param values: List of filter values. + :param values: Required. List of filter values. :type values: list[str] """ @@ -39,8 +41,8 @@ class PipelineRunQueryFilter(Model): 'values': {'key': 'values', 'type': '[str]'}, } - def __init__(self, operand, operator, values): - super(PipelineRunQueryFilter, self).__init__() - self.operand = operand - self.operator = operator - self.values = values + def __init__(self, **kwargs): + super(PipelineRunQueryFilter, self).__init__(**kwargs) + self.operand = kwargs.get('operand', None) + self.operator = kwargs.get('operator', None) + self.values = kwargs.get('values', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run_query_filter_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run_query_filter_py3.py new file mode 100644 index 000000000000..3e6d24a3ac73 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run_query_filter_py3.py @@ -0,0 +1,48 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class PipelineRunQueryFilter(Model): + """Query filter option for listing pipeline runs. + + All required parameters must be populated in order to send to Azure. + + :param operand: Required. Parameter name to be used for filter. Possible + values include: 'PipelineName', 'Status', 'RunStart', 'RunEnd' + :type operand: str or + ~azure.mgmt.datafactory.models.PipelineRunQueryFilterOperand + :param operator: Required. Operator to be used for filter. Possible values + include: 'Equals', 'NotEquals', 'In', 'NotIn' + :type operator: str or + ~azure.mgmt.datafactory.models.PipelineRunQueryFilterOperator + :param values: Required. List of filter values. + :type values: list[str] + """ + + _validation = { + 'operand': {'required': True}, + 'operator': {'required': True}, + 'values': {'required': True}, + } + + _attribute_map = { + 'operand': {'key': 'operand', 'type': 'str'}, + 'operator': {'key': 'operator', 'type': 'str'}, + 'values': {'key': 'values', 'type': '[str]'}, + } + + def __init__(self, *, operand, operator, values, **kwargs) -> None: + super(PipelineRunQueryFilter, self).__init__(**kwargs) + self.operand = operand + self.operator = operator + self.values = values diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run_query_order_by.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run_query_order_by.py index 9e0be821cc9e..27da9656c88b 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run_query_order_by.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run_query_order_by.py @@ -15,12 +15,14 @@ class PipelineRunQueryOrderBy(Model): """An object to provide order by options for listing pipeline runs. - :param order_by: Parameter name to be used for order by. Possible values - include: 'RunStart', 'RunEnd' + All required parameters must be populated in order to send to Azure. + + :param order_by: Required. Parameter name to be used for order by. + Possible values include: 'RunStart', 'RunEnd' :type order_by: str or ~azure.mgmt.datafactory.models.PipelineRunQueryOrderByField - :param order: Sorting order of the parameter. Possible values include: - 'ASC', 'DESC' + :param order: Required. Sorting order of the parameter. Possible values + include: 'ASC', 'DESC' :type order: str or ~azure.mgmt.datafactory.models.PipelineRunQueryOrder """ @@ -34,7 +36,7 @@ class PipelineRunQueryOrderBy(Model): 'order': {'key': 'order', 'type': 'str'}, } - def __init__(self, order_by, order): - super(PipelineRunQueryOrderBy, self).__init__() - self.order_by = order_by - self.order = order + def __init__(self, **kwargs): + super(PipelineRunQueryOrderBy, self).__init__(**kwargs) + self.order_by = kwargs.get('order_by', None) + self.order = kwargs.get('order', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run_query_order_by_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run_query_order_by_py3.py new file mode 100644 index 000000000000..4edaf9ba99de --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run_query_order_by_py3.py @@ -0,0 +1,42 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class PipelineRunQueryOrderBy(Model): + """An object to provide order by options for listing pipeline runs. + + All required parameters must be populated in order to send to Azure. + + :param order_by: Required. Parameter name to be used for order by. + Possible values include: 'RunStart', 'RunEnd' + :type order_by: str or + ~azure.mgmt.datafactory.models.PipelineRunQueryOrderByField + :param order: Required. Sorting order of the parameter. Possible values + include: 'ASC', 'DESC' + :type order: str or ~azure.mgmt.datafactory.models.PipelineRunQueryOrder + """ + + _validation = { + 'order_by': {'required': True}, + 'order': {'required': True}, + } + + _attribute_map = { + 'order_by': {'key': 'orderBy', 'type': 'str'}, + 'order': {'key': 'order', 'type': 'str'}, + } + + def __init__(self, *, order_by, order, **kwargs) -> None: + super(PipelineRunQueryOrderBy, self).__init__(**kwargs) + self.order_by = order_by + self.order = order diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run_query_response.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run_query_response.py index 8badc10ce056..50639429fc78 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run_query_response.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run_query_response.py @@ -15,7 +15,9 @@ class PipelineRunQueryResponse(Model): """A list pipeline runs. - :param value: List of pipeline runs. + All required parameters must be populated in order to send to Azure. + + :param value: Required. List of pipeline runs. :type value: list[~azure.mgmt.datafactory.models.PipelineRun] :param continuation_token: The continuation token for getting the next page of results, if any remaining results exist, null otherwise. @@ -31,7 +33,7 @@ class PipelineRunQueryResponse(Model): 'continuation_token': {'key': 'continuationToken', 'type': 'str'}, } - def __init__(self, value, continuation_token=None): - super(PipelineRunQueryResponse, self).__init__() - self.value = value - self.continuation_token = continuation_token + def __init__(self, **kwargs): + super(PipelineRunQueryResponse, self).__init__(**kwargs) + self.value = kwargs.get('value', None) + self.continuation_token = kwargs.get('continuation_token', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run_query_response_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run_query_response_py3.py new file mode 100644 index 000000000000..f0aef3ea8bc8 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run_query_response_py3.py @@ -0,0 +1,39 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class PipelineRunQueryResponse(Model): + """A list pipeline runs. + + All required parameters must be populated in order to send to Azure. + + :param value: Required. List of pipeline runs. + :type value: list[~azure.mgmt.datafactory.models.PipelineRun] + :param continuation_token: The continuation token for getting the next + page of results, if any remaining results exist, null otherwise. + :type continuation_token: str + """ + + _validation = { + 'value': {'required': True}, + } + + _attribute_map = { + 'value': {'key': 'value', 'type': '[PipelineRun]'}, + 'continuation_token': {'key': 'continuationToken', 'type': 'str'}, + } + + def __init__(self, *, value, continuation_token: str=None, **kwargs) -> None: + super(PipelineRunQueryResponse, self).__init__(**kwargs) + self.value = value + self.continuation_token = continuation_token diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/polybase_settings.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/polybase_settings.py index 0055d320b598..5a261d8fea84 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/polybase_settings.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/polybase_settings.py @@ -44,10 +44,10 @@ class PolybaseSettings(Model): 'use_type_default': {'key': 'useTypeDefault', 'type': 'object'}, } - def __init__(self, additional_properties=None, reject_type=None, reject_value=None, reject_sample_value=None, use_type_default=None): - super(PolybaseSettings, self).__init__() - self.additional_properties = additional_properties - self.reject_type = reject_type - self.reject_value = reject_value - self.reject_sample_value = reject_sample_value - self.use_type_default = use_type_default + def __init__(self, **kwargs): + super(PolybaseSettings, self).__init__(**kwargs) + self.additional_properties = kwargs.get('additional_properties', None) + self.reject_type = kwargs.get('reject_type', None) + self.reject_value = kwargs.get('reject_value', None) + self.reject_sample_value = kwargs.get('reject_sample_value', None) + self.use_type_default = kwargs.get('use_type_default', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/polybase_settings_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/polybase_settings_py3.py new file mode 100644 index 000000000000..baae78b14c5f --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/polybase_settings_py3.py @@ -0,0 +1,53 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class PolybaseSettings(Model): + """PolyBase settings. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param reject_type: Reject type. Possible values include: 'value', + 'percentage' + :type reject_type: str or + ~azure.mgmt.datafactory.models.PolybaseSettingsRejectType + :param reject_value: Specifies the value or the percentage of rows that + can be rejected before the query fails. Type: number (or Expression with + resultType number), minimum: 0. + :type reject_value: object + :param reject_sample_value: Determines the number of rows to attempt to + retrieve before the PolyBase recalculates the percentage of rejected rows. + Type: integer (or Expression with resultType integer), minimum: 0. + :type reject_sample_value: object + :param use_type_default: Specifies how to handle missing values in + delimited text files when PolyBase retrieves data from the text file. + Type: boolean (or Expression with resultType boolean). + :type use_type_default: object + """ + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'reject_type': {'key': 'rejectType', 'type': 'str'}, + 'reject_value': {'key': 'rejectValue', 'type': 'object'}, + 'reject_sample_value': {'key': 'rejectSampleValue', 'type': 'object'}, + 'use_type_default': {'key': 'useTypeDefault', 'type': 'object'}, + } + + def __init__(self, *, additional_properties=None, reject_type=None, reject_value=None, reject_sample_value=None, use_type_default=None, **kwargs) -> None: + super(PolybaseSettings, self).__init__(**kwargs) + self.additional_properties = additional_properties + self.reject_type = reject_type + self.reject_value = reject_value + self.reject_sample_value = reject_sample_value + self.use_type_default = use_type_default diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/postgre_sql_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/postgre_sql_linked_service.py index 27f23ae0e840..7a96d59e87e1 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/postgre_sql_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/postgre_sql_linked_service.py @@ -15,6 +15,8 @@ class PostgreSqlLinkedService(LinkedService): """Linked service for PostgreSQL data source. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -29,13 +31,13 @@ class PostgreSqlLinkedService(LinkedService): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param server: Server name for connection. Type: string (or Expression - with resultType string). + :param server: Required. Server name for connection. Type: string (or + Expression with resultType string). :type server: object - :param database: Database name for connection. Type: string (or Expression - with resultType string). + :param database: Required. Database name for connection. Type: string (or + Expression with resultType string). :type database: object :param schema: Schema name for connection. Type: string (or Expression with resultType string). @@ -72,12 +74,12 @@ class PostgreSqlLinkedService(LinkedService): 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, server, database, additional_properties=None, connect_via=None, description=None, parameters=None, annotations=None, schema=None, username=None, password=None, encrypted_credential=None): - super(PostgreSqlLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations) - self.server = server - self.database = database - self.schema = schema - self.username = username - self.password = password - self.encrypted_credential = encrypted_credential + def __init__(self, **kwargs): + super(PostgreSqlLinkedService, self).__init__(**kwargs) + self.server = kwargs.get('server', None) + self.database = kwargs.get('database', None) + self.schema = kwargs.get('schema', None) + self.username = kwargs.get('username', None) + self.password = kwargs.get('password', None) + self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'PostgreSql' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/postgre_sql_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/postgre_sql_linked_service_py3.py new file mode 100644 index 000000000000..976e29aebbde --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/postgre_sql_linked_service_py3.py @@ -0,0 +1,85 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class PostgreSqlLinkedService(LinkedService): + """Linked service for PostgreSQL data source. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param server: Required. Server name for connection. Type: string (or + Expression with resultType string). + :type server: object + :param database: Required. Database name for connection. Type: string (or + Expression with resultType string). + :type database: object + :param schema: Schema name for connection. Type: string (or Expression + with resultType string). + :type schema: object + :param username: Username for authentication. Type: string (or Expression + with resultType string). + :type username: object + :param password: Password for authentication. + :type password: ~azure.mgmt.datafactory.models.SecretBase + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'server': {'required': True}, + 'database': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'server': {'key': 'typeProperties.server', 'type': 'object'}, + 'database': {'key': 'typeProperties.database', 'type': 'object'}, + 'schema': {'key': 'typeProperties.schema', 'type': 'object'}, + 'username': {'key': 'typeProperties.username', 'type': 'object'}, + 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, *, server, database, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, schema=None, username=None, password=None, encrypted_credential=None, **kwargs) -> None: + super(PostgreSqlLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.server = server + self.database = database + self.schema = schema + self.username = username + self.password = password + self.encrypted_credential = encrypted_credential + self.type = 'PostgreSql' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/presto_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/presto_linked_service.py index 6bf59ae71e6f..abf4adde8515 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/presto_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/presto_linked_service.py @@ -15,6 +15,8 @@ class PrestoLinkedService(LinkedService): """Presto server linked service. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -29,20 +31,22 @@ class PrestoLinkedService(LinkedService): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param host: The IP address or host name of the Presto server. (i.e. - 192.168.222.160) + :param host: Required. The IP address or host name of the Presto server. + (i.e. 192.168.222.160) :type host: object - :param server_version: The version of the Presto server. (i.e. 0.148-t) + :param server_version: Required. The version of the Presto server. (i.e. + 0.148-t) :type server_version: object - :param catalog: The catalog context for all request against the server. + :param catalog: Required. The catalog context for all request against the + server. :type catalog: object :param port: The TCP port that the Presto server uses to listen for client connections. The default value is 8080. :type port: object - :param authentication_type: The authentication mechanism used to connect - to the Presto server. Possible values include: 'Anonymous', 'LDAP' + :param authentication_type: Required. The authentication mechanism used to + connect to the Presto server. Possible values include: 'Anonymous', 'LDAP' :type authentication_type: str or ~azure.mgmt.datafactory.models.PrestoAuthenticationType :param username: The user name used to connect to the Presto server. @@ -109,20 +113,20 @@ class PrestoLinkedService(LinkedService): 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, host, server_version, catalog, authentication_type, additional_properties=None, connect_via=None, description=None, parameters=None, annotations=None, port=None, username=None, password=None, enable_ssl=None, trusted_cert_path=None, use_system_trust_store=None, allow_host_name_cn_mismatch=None, allow_self_signed_server_cert=None, time_zone_id=None, encrypted_credential=None): - super(PrestoLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations) - self.host = host - self.server_version = server_version - self.catalog = catalog - self.port = port - self.authentication_type = authentication_type - self.username = username - self.password = password - self.enable_ssl = enable_ssl - self.trusted_cert_path = trusted_cert_path - self.use_system_trust_store = use_system_trust_store - self.allow_host_name_cn_mismatch = allow_host_name_cn_mismatch - self.allow_self_signed_server_cert = allow_self_signed_server_cert - self.time_zone_id = time_zone_id - self.encrypted_credential = encrypted_credential + def __init__(self, **kwargs): + super(PrestoLinkedService, self).__init__(**kwargs) + self.host = kwargs.get('host', None) + self.server_version = kwargs.get('server_version', None) + self.catalog = kwargs.get('catalog', None) + self.port = kwargs.get('port', None) + self.authentication_type = kwargs.get('authentication_type', None) + self.username = kwargs.get('username', None) + self.password = kwargs.get('password', None) + self.enable_ssl = kwargs.get('enable_ssl', None) + self.trusted_cert_path = kwargs.get('trusted_cert_path', None) + self.use_system_trust_store = kwargs.get('use_system_trust_store', None) + self.allow_host_name_cn_mismatch = kwargs.get('allow_host_name_cn_mismatch', None) + self.allow_self_signed_server_cert = kwargs.get('allow_self_signed_server_cert', None) + self.time_zone_id = kwargs.get('time_zone_id', None) + self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'Presto' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/presto_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/presto_linked_service_py3.py new file mode 100644 index 000000000000..e9721fcba562 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/presto_linked_service_py3.py @@ -0,0 +1,132 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class PrestoLinkedService(LinkedService): + """Presto server linked service. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param host: Required. The IP address or host name of the Presto server. + (i.e. 192.168.222.160) + :type host: object + :param server_version: Required. The version of the Presto server. (i.e. + 0.148-t) + :type server_version: object + :param catalog: Required. The catalog context for all request against the + server. + :type catalog: object + :param port: The TCP port that the Presto server uses to listen for client + connections. The default value is 8080. + :type port: object + :param authentication_type: Required. The authentication mechanism used to + connect to the Presto server. Possible values include: 'Anonymous', 'LDAP' + :type authentication_type: str or + ~azure.mgmt.datafactory.models.PrestoAuthenticationType + :param username: The user name used to connect to the Presto server. + :type username: object + :param password: The password corresponding to the user name. + :type password: ~azure.mgmt.datafactory.models.SecretBase + :param enable_ssl: Specifies whether the connections to the server are + encrypted using SSL. The default value is false. + :type enable_ssl: object + :param trusted_cert_path: The full path of the .pem file containing + trusted CA certificates for verifying the server when connecting over SSL. + This property can only be set when using SSL on self-hosted IR. The + default value is the cacerts.pem file installed with the IR. + :type trusted_cert_path: object + :param use_system_trust_store: Specifies whether to use a CA certificate + from the system trust store or from a specified PEM file. The default + value is false. + :type use_system_trust_store: object + :param allow_host_name_cn_mismatch: Specifies whether to require a + CA-issued SSL certificate name to match the host name of the server when + connecting over SSL. The default value is false. + :type allow_host_name_cn_mismatch: object + :param allow_self_signed_server_cert: Specifies whether to allow + self-signed certificates from the server. The default value is false. + :type allow_self_signed_server_cert: object + :param time_zone_id: The local time zone used by the connection. Valid + values for this option are specified in the IANA Time Zone Database. The + default value is the system time zone. + :type time_zone_id: object + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'host': {'required': True}, + 'server_version': {'required': True}, + 'catalog': {'required': True}, + 'authentication_type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'host': {'key': 'typeProperties.host', 'type': 'object'}, + 'server_version': {'key': 'typeProperties.serverVersion', 'type': 'object'}, + 'catalog': {'key': 'typeProperties.catalog', 'type': 'object'}, + 'port': {'key': 'typeProperties.port', 'type': 'object'}, + 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, + 'username': {'key': 'typeProperties.username', 'type': 'object'}, + 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, + 'enable_ssl': {'key': 'typeProperties.enableSsl', 'type': 'object'}, + 'trusted_cert_path': {'key': 'typeProperties.trustedCertPath', 'type': 'object'}, + 'use_system_trust_store': {'key': 'typeProperties.useSystemTrustStore', 'type': 'object'}, + 'allow_host_name_cn_mismatch': {'key': 'typeProperties.allowHostNameCNMismatch', 'type': 'object'}, + 'allow_self_signed_server_cert': {'key': 'typeProperties.allowSelfSignedServerCert', 'type': 'object'}, + 'time_zone_id': {'key': 'typeProperties.timeZoneID', 'type': 'object'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, *, host, server_version, catalog, authentication_type, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, port=None, username=None, password=None, enable_ssl=None, trusted_cert_path=None, use_system_trust_store=None, allow_host_name_cn_mismatch=None, allow_self_signed_server_cert=None, time_zone_id=None, encrypted_credential=None, **kwargs) -> None: + super(PrestoLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.host = host + self.server_version = server_version + self.catalog = catalog + self.port = port + self.authentication_type = authentication_type + self.username = username + self.password = password + self.enable_ssl = enable_ssl + self.trusted_cert_path = trusted_cert_path + self.use_system_trust_store = use_system_trust_store + self.allow_host_name_cn_mismatch = allow_host_name_cn_mismatch + self.allow_self_signed_server_cert = allow_self_signed_server_cert + self.time_zone_id = time_zone_id + self.encrypted_credential = encrypted_credential + self.type = 'Presto' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/presto_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/presto_object_dataset.py index 68034f6379bb..863d8851701d 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/presto_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/presto_object_dataset.py @@ -15,6 +15,8 @@ class PrestoObjectDataset(Dataset): """Presto server dataset. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -23,7 +25,7 @@ class PrestoObjectDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object - :param linked_service_name: Linked service reference. + :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference :param parameters: Parameters for dataset. @@ -32,7 +34,7 @@ class PrestoObjectDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str """ @@ -41,6 +43,16 @@ class PrestoObjectDataset(Dataset): 'type': {'required': True}, } - def __init__(self, linked_service_name, additional_properties=None, description=None, structure=None, parameters=None, annotations=None): - super(PrestoObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations) + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, **kwargs): + super(PrestoObjectDataset, self).__init__(**kwargs) self.type = 'PrestoObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/presto_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/presto_object_dataset_py3.py new file mode 100644 index 000000000000..a0dce3cd789b --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/presto_object_dataset_py3.py @@ -0,0 +1,58 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class PrestoObjectDataset(Dataset): + """Presto server dataset. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Required. Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: + super(PrestoObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + self.type = 'PrestoObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/presto_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/presto_source.py index 82a11441bef6..333a4e6dca9e 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/presto_source.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/presto_source.py @@ -15,6 +15,8 @@ class PrestoSource(CopySource): """A copy activity Presto server source. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -25,7 +27,7 @@ class PrestoSource(CopySource): with resultType string), pattern: ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type source_retry_wait: object - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). @@ -44,7 +46,7 @@ class PrestoSource(CopySource): 'query': {'key': 'query', 'type': 'object'}, } - def __init__(self, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None): - super(PrestoSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait) - self.query = query + def __init__(self, **kwargs): + super(PrestoSource, self).__init__(**kwargs) + self.query = kwargs.get('query', None) self.type = 'PrestoSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/presto_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/presto_source_py3.py new file mode 100644 index 000000000000..be957bd2bfcd --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/presto_source_py3.py @@ -0,0 +1,52 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class PrestoSource(CopySource): + """A copy activity Presto server source. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Required. Constant filled by server. + :type type: str + :param query: A query to retrieve data from source. Type: string (or + Expression with resultType string). + :type query: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'query': {'key': 'query', 'type': 'object'}, + } + + def __init__(self, *, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None, **kwargs) -> None: + super(PrestoSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, **kwargs) + self.query = query + self.type = 'PrestoSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_linked_service.py index b13c2f90f595..a199f9763f70 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_linked_service.py @@ -15,6 +15,8 @@ class QuickBooksLinkedService(LinkedService): """QuickBooks server linked service. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -29,12 +31,13 @@ class QuickBooksLinkedService(LinkedService): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param endpoint: The endpoint of the QuickBooks server. (i.e. + :param endpoint: Required. The endpoint of the QuickBooks server. (i.e. quickbooks.api.intuit.com) :type endpoint: object - :param company_id: The company ID of the QuickBooks company to authorize. + :param company_id: Required. The company ID of the QuickBooks company to + authorize. :type company_id: object :param access_token: The access token for OAuth 1.0 authentication. :type access_token: ~azure.mgmt.datafactory.models.SecretBase @@ -71,12 +74,12 @@ class QuickBooksLinkedService(LinkedService): 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, endpoint, company_id, additional_properties=None, connect_via=None, description=None, parameters=None, annotations=None, access_token=None, access_token_secret=None, use_encrypted_endpoints=None, encrypted_credential=None): - super(QuickBooksLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations) - self.endpoint = endpoint - self.company_id = company_id - self.access_token = access_token - self.access_token_secret = access_token_secret - self.use_encrypted_endpoints = use_encrypted_endpoints - self.encrypted_credential = encrypted_credential + def __init__(self, **kwargs): + super(QuickBooksLinkedService, self).__init__(**kwargs) + self.endpoint = kwargs.get('endpoint', None) + self.company_id = kwargs.get('company_id', None) + self.access_token = kwargs.get('access_token', None) + self.access_token_secret = kwargs.get('access_token_secret', None) + self.use_encrypted_endpoints = kwargs.get('use_encrypted_endpoints', None) + self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'QuickBooks' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_linked_service_py3.py new file mode 100644 index 000000000000..b56e34367553 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_linked_service_py3.py @@ -0,0 +1,85 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class QuickBooksLinkedService(LinkedService): + """QuickBooks server linked service. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param endpoint: Required. The endpoint of the QuickBooks server. (i.e. + quickbooks.api.intuit.com) + :type endpoint: object + :param company_id: Required. The company ID of the QuickBooks company to + authorize. + :type company_id: object + :param access_token: The access token for OAuth 1.0 authentication. + :type access_token: ~azure.mgmt.datafactory.models.SecretBase + :param access_token_secret: The access token secret for OAuth 1.0 + authentication. + :type access_token_secret: ~azure.mgmt.datafactory.models.SecretBase + :param use_encrypted_endpoints: Specifies whether the data source + endpoints are encrypted using HTTPS. The default value is true. + :type use_encrypted_endpoints: object + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'endpoint': {'required': True}, + 'company_id': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'endpoint': {'key': 'typeProperties.endpoint', 'type': 'object'}, + 'company_id': {'key': 'typeProperties.companyId', 'type': 'object'}, + 'access_token': {'key': 'typeProperties.accessToken', 'type': 'SecretBase'}, + 'access_token_secret': {'key': 'typeProperties.accessTokenSecret', 'type': 'SecretBase'}, + 'use_encrypted_endpoints': {'key': 'typeProperties.useEncryptedEndpoints', 'type': 'object'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, *, endpoint, company_id, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, access_token=None, access_token_secret=None, use_encrypted_endpoints=None, encrypted_credential=None, **kwargs) -> None: + super(QuickBooksLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.endpoint = endpoint + self.company_id = company_id + self.access_token = access_token + self.access_token_secret = access_token_secret + self.use_encrypted_endpoints = use_encrypted_endpoints + self.encrypted_credential = encrypted_credential + self.type = 'QuickBooks' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_object_dataset.py index 7420d28b53f3..56e68336c469 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_object_dataset.py @@ -15,6 +15,8 @@ class QuickBooksObjectDataset(Dataset): """QuickBooks server dataset. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -23,7 +25,7 @@ class QuickBooksObjectDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object - :param linked_service_name: Linked service reference. + :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference :param parameters: Parameters for dataset. @@ -32,7 +34,7 @@ class QuickBooksObjectDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str """ @@ -41,6 +43,16 @@ class QuickBooksObjectDataset(Dataset): 'type': {'required': True}, } - def __init__(self, linked_service_name, additional_properties=None, description=None, structure=None, parameters=None, annotations=None): - super(QuickBooksObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations) + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, **kwargs): + super(QuickBooksObjectDataset, self).__init__(**kwargs) self.type = 'QuickBooksObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_object_dataset_py3.py new file mode 100644 index 000000000000..e852f5bcd624 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_object_dataset_py3.py @@ -0,0 +1,58 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class QuickBooksObjectDataset(Dataset): + """QuickBooks server dataset. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Required. Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: + super(QuickBooksObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + self.type = 'QuickBooksObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_source.py index d43f206b1407..b8567cd772ed 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_source.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_source.py @@ -15,6 +15,8 @@ class QuickBooksSource(CopySource): """A copy activity QuickBooks server source. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -25,7 +27,7 @@ class QuickBooksSource(CopySource): with resultType string), pattern: ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type source_retry_wait: object - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). @@ -44,7 +46,7 @@ class QuickBooksSource(CopySource): 'query': {'key': 'query', 'type': 'object'}, } - def __init__(self, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None): - super(QuickBooksSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait) - self.query = query + def __init__(self, **kwargs): + super(QuickBooksSource, self).__init__(**kwargs) + self.query = kwargs.get('query', None) self.type = 'QuickBooksSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_source_py3.py new file mode 100644 index 000000000000..e2ecafefc3bf --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_source_py3.py @@ -0,0 +1,52 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class QuickBooksSource(CopySource): + """A copy activity QuickBooks server source. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Required. Constant filled by server. + :type type: str + :param query: A query to retrieve data from source. Type: string (or + Expression with resultType string). + :type query: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'query': {'key': 'query', 'type': 'object'}, + } + + def __init__(self, *, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None, **kwargs) -> None: + super(QuickBooksSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, **kwargs) + self.query = query + self.type = 'QuickBooksSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/recurrence_schedule.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/recurrence_schedule.py index a36216b08620..f23d452392b0 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/recurrence_schedule.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/recurrence_schedule.py @@ -40,11 +40,11 @@ class RecurrenceSchedule(Model): 'monthly_occurrences': {'key': 'monthlyOccurrences', 'type': '[RecurrenceScheduleOccurrence]'}, } - def __init__(self, additional_properties=None, minutes=None, hours=None, week_days=None, month_days=None, monthly_occurrences=None): - super(RecurrenceSchedule, self).__init__() - self.additional_properties = additional_properties - self.minutes = minutes - self.hours = hours - self.week_days = week_days - self.month_days = month_days - self.monthly_occurrences = monthly_occurrences + def __init__(self, **kwargs): + super(RecurrenceSchedule, self).__init__(**kwargs) + self.additional_properties = kwargs.get('additional_properties', None) + self.minutes = kwargs.get('minutes', None) + self.hours = kwargs.get('hours', None) + self.week_days = kwargs.get('week_days', None) + self.month_days = kwargs.get('month_days', None) + self.monthly_occurrences = kwargs.get('monthly_occurrences', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/recurrence_schedule_occurrence.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/recurrence_schedule_occurrence.py index 35118f257baf..f42136545ac2 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/recurrence_schedule_occurrence.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/recurrence_schedule_occurrence.py @@ -31,8 +31,8 @@ class RecurrenceScheduleOccurrence(Model): 'occurrence': {'key': 'occurrence', 'type': 'int'}, } - def __init__(self, additional_properties=None, day=None, occurrence=None): - super(RecurrenceScheduleOccurrence, self).__init__() - self.additional_properties = additional_properties - self.day = day - self.occurrence = occurrence + def __init__(self, **kwargs): + super(RecurrenceScheduleOccurrence, self).__init__(**kwargs) + self.additional_properties = kwargs.get('additional_properties', None) + self.day = kwargs.get('day', None) + self.occurrence = kwargs.get('occurrence', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/recurrence_schedule_occurrence_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/recurrence_schedule_occurrence_py3.py new file mode 100644 index 000000000000..e3173e47e8e4 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/recurrence_schedule_occurrence_py3.py @@ -0,0 +1,38 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class RecurrenceScheduleOccurrence(Model): + """The recurrence schedule occurence. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param day: The day of the week. Possible values include: 'Sunday', + 'Monday', 'Tuesday', 'Wednesday', 'Thursday', 'Friday', 'Saturday' + :type day: str or ~azure.mgmt.datafactory.models.DayOfWeek + :param occurrence: The occurrence. + :type occurrence: int + """ + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'day': {'key': 'day', 'type': 'DayOfWeek'}, + 'occurrence': {'key': 'occurrence', 'type': 'int'}, + } + + def __init__(self, *, additional_properties=None, day=None, occurrence: int=None, **kwargs) -> None: + super(RecurrenceScheduleOccurrence, self).__init__(**kwargs) + self.additional_properties = additional_properties + self.day = day + self.occurrence = occurrence diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/recurrence_schedule_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/recurrence_schedule_py3.py new file mode 100644 index 000000000000..fbe44fa3f021 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/recurrence_schedule_py3.py @@ -0,0 +1,50 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class RecurrenceSchedule(Model): + """The recurrence schedule. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param minutes: The minutes. + :type minutes: list[int] + :param hours: The hours. + :type hours: list[int] + :param week_days: The days of the week. + :type week_days: list[str or ~azure.mgmt.datafactory.models.DaysOfWeek] + :param month_days: The month days. + :type month_days: list[int] + :param monthly_occurrences: The monthly occurrences. + :type monthly_occurrences: + list[~azure.mgmt.datafactory.models.RecurrenceScheduleOccurrence] + """ + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'minutes': {'key': 'minutes', 'type': '[int]'}, + 'hours': {'key': 'hours', 'type': '[int]'}, + 'week_days': {'key': 'weekDays', 'type': '[DaysOfWeek]'}, + 'month_days': {'key': 'monthDays', 'type': '[int]'}, + 'monthly_occurrences': {'key': 'monthlyOccurrences', 'type': '[RecurrenceScheduleOccurrence]'}, + } + + def __init__(self, *, additional_properties=None, minutes=None, hours=None, week_days=None, month_days=None, monthly_occurrences=None, **kwargs) -> None: + super(RecurrenceSchedule, self).__init__(**kwargs) + self.additional_properties = additional_properties + self.minutes = minutes + self.hours = hours + self.week_days = week_days + self.month_days = month_days + self.monthly_occurrences = monthly_occurrences diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/redirect_incompatible_row_settings.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/redirect_incompatible_row_settings.py index 8d5aba7e3ed0..a2e3bddb9425 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/redirect_incompatible_row_settings.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/redirect_incompatible_row_settings.py @@ -15,13 +15,15 @@ class RedirectIncompatibleRowSettings(Model): """Redirect incompatible row settings. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] - :param linked_service_name: Name of the Azure Storage, Storage SAS, or - Azure Data Lake Store linked service used for redirecting incompatible - row. Must be specified if redirectIncompatibleRowSettings is specified. - Type: string (or Expression with resultType string). + :param linked_service_name: Required. Name of the Azure Storage, Storage + SAS, or Azure Data Lake Store linked service used for redirecting + incompatible row. Must be specified if redirectIncompatibleRowSettings is + specified. Type: string (or Expression with resultType string). :type linked_service_name: object :param path: The path for storing the redirect incompatible row data. Type: string (or Expression with resultType string). @@ -38,8 +40,8 @@ class RedirectIncompatibleRowSettings(Model): 'path': {'key': 'path', 'type': 'object'}, } - def __init__(self, linked_service_name, additional_properties=None, path=None): - super(RedirectIncompatibleRowSettings, self).__init__() - self.additional_properties = additional_properties - self.linked_service_name = linked_service_name - self.path = path + def __init__(self, **kwargs): + super(RedirectIncompatibleRowSettings, self).__init__(**kwargs) + self.additional_properties = kwargs.get('additional_properties', None) + self.linked_service_name = kwargs.get('linked_service_name', None) + self.path = kwargs.get('path', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/redirect_incompatible_row_settings_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/redirect_incompatible_row_settings_py3.py new file mode 100644 index 000000000000..b47878ef4354 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/redirect_incompatible_row_settings_py3.py @@ -0,0 +1,47 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class RedirectIncompatibleRowSettings(Model): + """Redirect incompatible row settings. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param linked_service_name: Required. Name of the Azure Storage, Storage + SAS, or Azure Data Lake Store linked service used for redirecting + incompatible row. Must be specified if redirectIncompatibleRowSettings is + specified. Type: string (or Expression with resultType string). + :type linked_service_name: object + :param path: The path for storing the redirect incompatible row data. + Type: string (or Expression with resultType string). + :type path: object + """ + + _validation = { + 'linked_service_name': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'object'}, + 'path': {'key': 'path', 'type': 'object'}, + } + + def __init__(self, *, linked_service_name, additional_properties=None, path=None, **kwargs) -> None: + super(RedirectIncompatibleRowSettings, self).__init__(**kwargs) + self.additional_properties = additional_properties + self.linked_service_name = linked_service_name + self.path = path diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/redshift_unload_settings.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/redshift_unload_settings.py index 46552265701d..7114b85e10db 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/redshift_unload_settings.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/redshift_unload_settings.py @@ -18,15 +18,17 @@ class RedshiftUnloadSettings(Model): will be unloaded into S3 first and then copied into the targeted sink from the interim S3. - :param s3_linked_service_name: The name of the Amazon S3 linked service - which will be used for the unload operation when copying from the Amazon - Redshift source. + All required parameters must be populated in order to send to Azure. + + :param s3_linked_service_name: Required. The name of the Amazon S3 linked + service which will be used for the unload operation when copying from the + Amazon Redshift source. :type s3_linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference - :param bucket_name: The bucket of the interim Amazon S3 which will be used - to store the unloaded data from Amazon Redshift source. The bucket must be - in the same region as the Amazon Redshift source. Type: string (or - Expression with resultType string). + :param bucket_name: Required. The bucket of the interim Amazon S3 which + will be used to store the unloaded data from Amazon Redshift source. The + bucket must be in the same region as the Amazon Redshift source. Type: + string (or Expression with resultType string). :type bucket_name: object """ @@ -40,7 +42,7 @@ class RedshiftUnloadSettings(Model): 'bucket_name': {'key': 'bucketName', 'type': 'object'}, } - def __init__(self, s3_linked_service_name, bucket_name): - super(RedshiftUnloadSettings, self).__init__() - self.s3_linked_service_name = s3_linked_service_name - self.bucket_name = bucket_name + def __init__(self, **kwargs): + super(RedshiftUnloadSettings, self).__init__(**kwargs) + self.s3_linked_service_name = kwargs.get('s3_linked_service_name', None) + self.bucket_name = kwargs.get('bucket_name', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/redshift_unload_settings_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/redshift_unload_settings_py3.py new file mode 100644 index 000000000000..a40d014a32f9 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/redshift_unload_settings_py3.py @@ -0,0 +1,48 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class RedshiftUnloadSettings(Model): + """The Amazon S3 settings needed for the interim Amazon S3 when copying from + Amazon Redshift with unload. With this, data from Amazon Redshift source + will be unloaded into S3 first and then copied into the targeted sink from + the interim S3. + + All required parameters must be populated in order to send to Azure. + + :param s3_linked_service_name: Required. The name of the Amazon S3 linked + service which will be used for the unload operation when copying from the + Amazon Redshift source. + :type s3_linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param bucket_name: Required. The bucket of the interim Amazon S3 which + will be used to store the unloaded data from Amazon Redshift source. The + bucket must be in the same region as the Amazon Redshift source. Type: + string (or Expression with resultType string). + :type bucket_name: object + """ + + _validation = { + 's3_linked_service_name': {'required': True}, + 'bucket_name': {'required': True}, + } + + _attribute_map = { + 's3_linked_service_name': {'key': 's3LinkedServiceName', 'type': 'LinkedServiceReference'}, + 'bucket_name': {'key': 'bucketName', 'type': 'object'}, + } + + def __init__(self, *, s3_linked_service_name, bucket_name, **kwargs) -> None: + super(RedshiftUnloadSettings, self).__init__(**kwargs) + self.s3_linked_service_name = s3_linked_service_name + self.bucket_name = bucket_name diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/relational_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/relational_source.py index 93582c8536ec..1dc8ff198eb8 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/relational_source.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/relational_source.py @@ -15,6 +15,8 @@ class RelationalSource(CopySource): """A copy activity source for various relational databases. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -25,7 +27,7 @@ class RelationalSource(CopySource): with resultType string), pattern: ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type source_retry_wait: object - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param query: Database query. Type: string (or Expression with resultType string). @@ -44,7 +46,7 @@ class RelationalSource(CopySource): 'query': {'key': 'query', 'type': 'object'}, } - def __init__(self, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None): - super(RelationalSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait) - self.query = query + def __init__(self, **kwargs): + super(RelationalSource, self).__init__(**kwargs) + self.query = kwargs.get('query', None) self.type = 'RelationalSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/relational_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/relational_source_py3.py new file mode 100644 index 000000000000..efac5c7b9050 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/relational_source_py3.py @@ -0,0 +1,52 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class RelationalSource(CopySource): + """A copy activity source for various relational databases. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Required. Constant filled by server. + :type type: str + :param query: Database query. Type: string (or Expression with resultType + string). + :type query: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'query': {'key': 'query', 'type': 'object'}, + } + + def __init__(self, *, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None, **kwargs) -> None: + super(RelationalSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, **kwargs) + self.query = query + self.type = 'RelationalSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/relational_table_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/relational_table_dataset.py index ea5cd0fe5591..1037b936b32f 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/relational_table_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/relational_table_dataset.py @@ -15,6 +15,8 @@ class RelationalTableDataset(Dataset): """The relational table dataset. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -23,7 +25,7 @@ class RelationalTableDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object - :param linked_service_name: Linked service reference. + :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference :param parameters: Parameters for dataset. @@ -32,7 +34,7 @@ class RelationalTableDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param table_name: The relational table name. Type: string (or Expression with resultType string). @@ -55,7 +57,7 @@ class RelationalTableDataset(Dataset): 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, linked_service_name, additional_properties=None, description=None, structure=None, parameters=None, annotations=None, table_name=None): - super(RelationalTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations) - self.table_name = table_name + def __init__(self, **kwargs): + super(RelationalTableDataset, self).__init__(**kwargs) + self.table_name = kwargs.get('table_name', None) self.type = 'RelationalTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/relational_table_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/relational_table_dataset_py3.py new file mode 100644 index 000000000000..908ca1d8657d --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/relational_table_dataset_py3.py @@ -0,0 +1,63 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class RelationalTableDataset(Dataset): + """The relational table dataset. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Required. Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param table_name: The relational table name. Type: string (or Expression + with resultType string). + :type table_name: object + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, + } + + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, table_name=None, **kwargs) -> None: + super(RelationalTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + self.table_name = table_name + self.type = 'RelationalTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/resource.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/resource.py index 88bea69ad1a1..38d454c351e6 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/resource.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/resource.py @@ -44,10 +44,10 @@ class Resource(Model): 'tags': {'key': 'tags', 'type': '{str}'}, } - def __init__(self, location=None, tags=None): - super(Resource, self).__init__() + def __init__(self, **kwargs): + super(Resource, self).__init__(**kwargs) self.id = None self.name = None self.type = None - self.location = location - self.tags = tags + self.location = kwargs.get('location', None) + self.tags = kwargs.get('tags', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/resource_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/resource_py3.py new file mode 100644 index 000000000000..47cbb7d6aa4b --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/resource_py3.py @@ -0,0 +1,53 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class Resource(Model): + """Azure Data Factory top-level resource. + + Variables are only populated by the server, and will be ignored when + sending a request. + + :ivar id: The resource identifier. + :vartype id: str + :ivar name: The resource name. + :vartype name: str + :ivar type: The resource type. + :vartype type: str + :param location: The resource location. + :type location: str + :param tags: The resource tags. + :type tags: dict[str, str] + """ + + _validation = { + 'id': {'readonly': True}, + 'name': {'readonly': True}, + 'type': {'readonly': True}, + } + + _attribute_map = { + 'id': {'key': 'id', 'type': 'str'}, + 'name': {'key': 'name', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'str'}, + 'location': {'key': 'location', 'type': 'str'}, + 'tags': {'key': 'tags', 'type': '{str}'}, + } + + def __init__(self, *, location: str=None, tags=None, **kwargs) -> None: + super(Resource, self).__init__(**kwargs) + self.id = None + self.name = None + self.type = None + self.location = location + self.tags = tags diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_linked_service.py new file mode 100644 index 000000000000..9c1b8e4c3cbd --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_linked_service.py @@ -0,0 +1,94 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class ResponsysLinkedService(LinkedService): + """Responsys linked service. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param endpoint: Required. The endpoint of the Responsys server. + :type endpoint: object + :param client_id: Required. The client ID associated with the Responsys + application. Type: string (or Expression with resultType string). + :type client_id: object + :param client_secret: The client secret associated with the Responsys + application. Type: string (or Expression with resultType string). + :type client_secret: ~azure.mgmt.datafactory.models.SecretBase + :param use_encrypted_endpoints: Specifies whether the data source + endpoints are encrypted using HTTPS. The default value is true. Type: + boolean (or Expression with resultType boolean). + :type use_encrypted_endpoints: object + :param use_host_verification: Specifies whether to require the host name + in the server's certificate to match the host name of the server when + connecting over SSL. The default value is true. Type: boolean (or + Expression with resultType boolean). + :type use_host_verification: object + :param use_peer_verification: Specifies whether to verify the identity of + the server when connecting over SSL. The default value is true. Type: + boolean (or Expression with resultType boolean). + :type use_peer_verification: object + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'endpoint': {'required': True}, + 'client_id': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'endpoint': {'key': 'typeProperties.endpoint', 'type': 'object'}, + 'client_id': {'key': 'typeProperties.clientId', 'type': 'object'}, + 'client_secret': {'key': 'typeProperties.clientSecret', 'type': 'SecretBase'}, + 'use_encrypted_endpoints': {'key': 'typeProperties.useEncryptedEndpoints', 'type': 'object'}, + 'use_host_verification': {'key': 'typeProperties.useHostVerification', 'type': 'object'}, + 'use_peer_verification': {'key': 'typeProperties.usePeerVerification', 'type': 'object'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, **kwargs): + super(ResponsysLinkedService, self).__init__(**kwargs) + self.endpoint = kwargs.get('endpoint', None) + self.client_id = kwargs.get('client_id', None) + self.client_secret = kwargs.get('client_secret', None) + self.use_encrypted_endpoints = kwargs.get('use_encrypted_endpoints', None) + self.use_host_verification = kwargs.get('use_host_verification', None) + self.use_peer_verification = kwargs.get('use_peer_verification', None) + self.encrypted_credential = kwargs.get('encrypted_credential', None) + self.type = 'Responsys' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_linked_service_py3.py new file mode 100644 index 000000000000..efad0984a99f --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_linked_service_py3.py @@ -0,0 +1,94 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class ResponsysLinkedService(LinkedService): + """Responsys linked service. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param endpoint: Required. The endpoint of the Responsys server. + :type endpoint: object + :param client_id: Required. The client ID associated with the Responsys + application. Type: string (or Expression with resultType string). + :type client_id: object + :param client_secret: The client secret associated with the Responsys + application. Type: string (or Expression with resultType string). + :type client_secret: ~azure.mgmt.datafactory.models.SecretBase + :param use_encrypted_endpoints: Specifies whether the data source + endpoints are encrypted using HTTPS. The default value is true. Type: + boolean (or Expression with resultType boolean). + :type use_encrypted_endpoints: object + :param use_host_verification: Specifies whether to require the host name + in the server's certificate to match the host name of the server when + connecting over SSL. The default value is true. Type: boolean (or + Expression with resultType boolean). + :type use_host_verification: object + :param use_peer_verification: Specifies whether to verify the identity of + the server when connecting over SSL. The default value is true. Type: + boolean (or Expression with resultType boolean). + :type use_peer_verification: object + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'endpoint': {'required': True}, + 'client_id': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'endpoint': {'key': 'typeProperties.endpoint', 'type': 'object'}, + 'client_id': {'key': 'typeProperties.clientId', 'type': 'object'}, + 'client_secret': {'key': 'typeProperties.clientSecret', 'type': 'SecretBase'}, + 'use_encrypted_endpoints': {'key': 'typeProperties.useEncryptedEndpoints', 'type': 'object'}, + 'use_host_verification': {'key': 'typeProperties.useHostVerification', 'type': 'object'}, + 'use_peer_verification': {'key': 'typeProperties.usePeerVerification', 'type': 'object'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, *, endpoint, client_id, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, client_secret=None, use_encrypted_endpoints=None, use_host_verification=None, use_peer_verification=None, encrypted_credential=None, **kwargs) -> None: + super(ResponsysLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.endpoint = endpoint + self.client_id = client_id + self.client_secret = client_secret + self.use_encrypted_endpoints = use_encrypted_endpoints + self.use_host_verification = use_host_verification + self.use_peer_verification = use_peer_verification + self.encrypted_credential = encrypted_credential + self.type = 'Responsys' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_object_dataset.py new file mode 100644 index 000000000000..eef47fbf10a8 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_object_dataset.py @@ -0,0 +1,58 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class ResponsysObjectDataset(Dataset): + """Responsys dataset. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Required. Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, **kwargs): + super(ResponsysObjectDataset, self).__init__(**kwargs) + self.type = 'ResponsysObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_object_dataset_py3.py new file mode 100644 index 000000000000..e7b88bd52cec --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_object_dataset_py3.py @@ -0,0 +1,58 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class ResponsysObjectDataset(Dataset): + """Responsys dataset. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Required. Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: + super(ResponsysObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + self.type = 'ResponsysObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_source.py new file mode 100644 index 000000000000..1e1a9397a6ba --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_source.py @@ -0,0 +1,52 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class ResponsysSource(CopySource): + """A copy activity Responsys source. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Required. Constant filled by server. + :type type: str + :param query: A query to retrieve data from source. Type: string (or + Expression with resultType string). + :type query: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'query': {'key': 'query', 'type': 'object'}, + } + + def __init__(self, **kwargs): + super(ResponsysSource, self).__init__(**kwargs) + self.query = kwargs.get('query', None) + self.type = 'ResponsysSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_source_py3.py new file mode 100644 index 000000000000..1934f830d515 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_source_py3.py @@ -0,0 +1,52 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class ResponsysSource(CopySource): + """A copy activity Responsys source. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Required. Constant filled by server. + :type type: str + :param query: A query to retrieve data from source. Type: string (or + Expression with resultType string). + :type query: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'query': {'key': 'query', 'type': 'object'}, + } + + def __init__(self, *, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None, **kwargs) -> None: + super(ResponsysSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, **kwargs) + self.query = query + self.type = 'ResponsysSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/retry_policy.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/retry_policy.py index dd6aa0d11d76..e6f5b1876259 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/retry_policy.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/retry_policy.py @@ -32,7 +32,7 @@ class RetryPolicy(Model): 'interval_in_seconds': {'key': 'intervalInSeconds', 'type': 'int'}, } - def __init__(self, count=None, interval_in_seconds=None): - super(RetryPolicy, self).__init__() - self.count = count - self.interval_in_seconds = interval_in_seconds + def __init__(self, **kwargs): + super(RetryPolicy, self).__init__(**kwargs) + self.count = kwargs.get('count', None) + self.interval_in_seconds = kwargs.get('interval_in_seconds', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/retry_policy_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/retry_policy_py3.py new file mode 100644 index 000000000000..b51b87a49938 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/retry_policy_py3.py @@ -0,0 +1,38 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class RetryPolicy(Model): + """Execution policy for an activity. + + :param count: Maximum ordinary retry attempts. Default is 0. Type: integer + (or Expression with resultType integer), minimum: 0. + :type count: object + :param interval_in_seconds: Interval between retries in seconds. Default + is 30. + :type interval_in_seconds: int + """ + + _validation = { + 'interval_in_seconds': {'maximum': 86400, 'minimum': 30}, + } + + _attribute_map = { + 'count': {'key': 'count', 'type': 'object'}, + 'interval_in_seconds': {'key': 'intervalInSeconds', 'type': 'int'}, + } + + def __init__(self, *, count=None, interval_in_seconds: int=None, **kwargs) -> None: + super(RetryPolicy, self).__init__(**kwargs) + self.count = count + self.interval_in_seconds = interval_in_seconds diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_linked_service.py index 0aeadaecafd5..5804e779d1ef 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_linked_service.py @@ -15,6 +15,8 @@ class SalesforceLinkedService(LinkedService): """Linked service for Salesforce. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -29,7 +31,7 @@ class SalesforceLinkedService(LinkedService): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param environment_url: The URL of Salesforce instance. Default is 'https://login.salesforce.com'. To copy data from sandbox, specify @@ -70,11 +72,11 @@ class SalesforceLinkedService(LinkedService): 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, additional_properties=None, connect_via=None, description=None, parameters=None, annotations=None, environment_url=None, username=None, password=None, security_token=None, encrypted_credential=None): - super(SalesforceLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations) - self.environment_url = environment_url - self.username = username - self.password = password - self.security_token = security_token - self.encrypted_credential = encrypted_credential + def __init__(self, **kwargs): + super(SalesforceLinkedService, self).__init__(**kwargs) + self.environment_url = kwargs.get('environment_url', None) + self.username = kwargs.get('username', None) + self.password = kwargs.get('password', None) + self.security_token = kwargs.get('security_token', None) + self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'Salesforce' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_linked_service_py3.py new file mode 100644 index 000000000000..2668ed113708 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_linked_service_py3.py @@ -0,0 +1,82 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class SalesforceLinkedService(LinkedService): + """Linked service for Salesforce. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param environment_url: The URL of Salesforce instance. Default is + 'https://login.salesforce.com'. To copy data from sandbox, specify + 'https://test.salesforce.com'. To copy data from custom domain, specify, + for example, 'https://[domain].my.salesforce.com'. Type: string (or + Expression with resultType string). + :type environment_url: object + :param username: The username for Basic authentication of the Salesforce + instance. Type: string (or Expression with resultType string). + :type username: object + :param password: The password for Basic authentication of the Salesforce + instance. + :type password: ~azure.mgmt.datafactory.models.SecretBase + :param security_token: The security token is required to remotely access + Salesforce instance. + :type security_token: ~azure.mgmt.datafactory.models.SecretBase + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'environment_url': {'key': 'typeProperties.environmentUrl', 'type': 'object'}, + 'username': {'key': 'typeProperties.username', 'type': 'object'}, + 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, + 'security_token': {'key': 'typeProperties.securityToken', 'type': 'SecretBase'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, *, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, environment_url=None, username=None, password=None, security_token=None, encrypted_credential=None, **kwargs) -> None: + super(SalesforceLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.environment_url = environment_url + self.username = username + self.password = password + self.security_token = security_token + self.encrypted_credential = encrypted_credential + self.type = 'Salesforce' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_marketing_cloud_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_marketing_cloud_linked_service.py index 2a178a681bd3..f3d2861576e4 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_marketing_cloud_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_marketing_cloud_linked_service.py @@ -15,6 +15,8 @@ class SalesforceMarketingCloudLinkedService(LinkedService): """Salesforce Marketing Cloud linked service. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -29,10 +31,11 @@ class SalesforceMarketingCloudLinkedService(LinkedService): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param client_id: The client ID associated with the Salesforce Marketing - Cloud application. Type: string (or Expression with resultType string). + :param client_id: Required. The client ID associated with the Salesforce + Marketing Cloud application. Type: string (or Expression with resultType + string). :type client_id: object :param client_secret: The client secret associated with the Salesforce Marketing Cloud application. Type: string (or Expression with resultType @@ -77,12 +80,12 @@ class SalesforceMarketingCloudLinkedService(LinkedService): 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, client_id, additional_properties=None, connect_via=None, description=None, parameters=None, annotations=None, client_secret=None, use_encrypted_endpoints=None, use_host_verification=None, use_peer_verification=None, encrypted_credential=None): - super(SalesforceMarketingCloudLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations) - self.client_id = client_id - self.client_secret = client_secret - self.use_encrypted_endpoints = use_encrypted_endpoints - self.use_host_verification = use_host_verification - self.use_peer_verification = use_peer_verification - self.encrypted_credential = encrypted_credential + def __init__(self, **kwargs): + super(SalesforceMarketingCloudLinkedService, self).__init__(**kwargs) + self.client_id = kwargs.get('client_id', None) + self.client_secret = kwargs.get('client_secret', None) + self.use_encrypted_endpoints = kwargs.get('use_encrypted_endpoints', None) + self.use_host_verification = kwargs.get('use_host_verification', None) + self.use_peer_verification = kwargs.get('use_peer_verification', None) + self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'SalesforceMarketingCloud' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_marketing_cloud_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_marketing_cloud_linked_service_py3.py new file mode 100644 index 000000000000..2cf5a947413e --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_marketing_cloud_linked_service_py3.py @@ -0,0 +1,91 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class SalesforceMarketingCloudLinkedService(LinkedService): + """Salesforce Marketing Cloud linked service. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param client_id: Required. The client ID associated with the Salesforce + Marketing Cloud application. Type: string (or Expression with resultType + string). + :type client_id: object + :param client_secret: The client secret associated with the Salesforce + Marketing Cloud application. Type: string (or Expression with resultType + string). + :type client_secret: ~azure.mgmt.datafactory.models.SecretBase + :param use_encrypted_endpoints: Specifies whether the data source + endpoints are encrypted using HTTPS. The default value is true. Type: + boolean (or Expression with resultType boolean). + :type use_encrypted_endpoints: object + :param use_host_verification: Specifies whether to require the host name + in the server's certificate to match the host name of the server when + connecting over SSL. The default value is true. Type: boolean (or + Expression with resultType boolean). + :type use_host_verification: object + :param use_peer_verification: Specifies whether to verify the identity of + the server when connecting over SSL. The default value is true. Type: + boolean (or Expression with resultType boolean). + :type use_peer_verification: object + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'client_id': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'client_id': {'key': 'typeProperties.clientId', 'type': 'object'}, + 'client_secret': {'key': 'typeProperties.clientSecret', 'type': 'SecretBase'}, + 'use_encrypted_endpoints': {'key': 'typeProperties.useEncryptedEndpoints', 'type': 'object'}, + 'use_host_verification': {'key': 'typeProperties.useHostVerification', 'type': 'object'}, + 'use_peer_verification': {'key': 'typeProperties.usePeerVerification', 'type': 'object'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, *, client_id, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, client_secret=None, use_encrypted_endpoints=None, use_host_verification=None, use_peer_verification=None, encrypted_credential=None, **kwargs) -> None: + super(SalesforceMarketingCloudLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.client_id = client_id + self.client_secret = client_secret + self.use_encrypted_endpoints = use_encrypted_endpoints + self.use_host_verification = use_host_verification + self.use_peer_verification = use_peer_verification + self.encrypted_credential = encrypted_credential + self.type = 'SalesforceMarketingCloud' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_marketing_cloud_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_marketing_cloud_object_dataset.py index 384042dbf082..7d8a49518ca6 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_marketing_cloud_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_marketing_cloud_object_dataset.py @@ -15,6 +15,8 @@ class SalesforceMarketingCloudObjectDataset(Dataset): """Salesforce Marketing Cloud dataset. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -23,7 +25,7 @@ class SalesforceMarketingCloudObjectDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object - :param linked_service_name: Linked service reference. + :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference :param parameters: Parameters for dataset. @@ -32,7 +34,7 @@ class SalesforceMarketingCloudObjectDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str """ @@ -41,6 +43,16 @@ class SalesforceMarketingCloudObjectDataset(Dataset): 'type': {'required': True}, } - def __init__(self, linked_service_name, additional_properties=None, description=None, structure=None, parameters=None, annotations=None): - super(SalesforceMarketingCloudObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations) + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, **kwargs): + super(SalesforceMarketingCloudObjectDataset, self).__init__(**kwargs) self.type = 'SalesforceMarketingCloudObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_marketing_cloud_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_marketing_cloud_object_dataset_py3.py new file mode 100644 index 000000000000..08b243e3c826 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_marketing_cloud_object_dataset_py3.py @@ -0,0 +1,58 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class SalesforceMarketingCloudObjectDataset(Dataset): + """Salesforce Marketing Cloud dataset. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Required. Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: + super(SalesforceMarketingCloudObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + self.type = 'SalesforceMarketingCloudObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_marketing_cloud_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_marketing_cloud_source.py index 7a606977d304..bf08fdaa88bf 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_marketing_cloud_source.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_marketing_cloud_source.py @@ -15,6 +15,8 @@ class SalesforceMarketingCloudSource(CopySource): """A copy activity Salesforce Marketing Cloud source. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -25,7 +27,7 @@ class SalesforceMarketingCloudSource(CopySource): with resultType string), pattern: ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type source_retry_wait: object - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). @@ -44,7 +46,7 @@ class SalesforceMarketingCloudSource(CopySource): 'query': {'key': 'query', 'type': 'object'}, } - def __init__(self, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None): - super(SalesforceMarketingCloudSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait) - self.query = query + def __init__(self, **kwargs): + super(SalesforceMarketingCloudSource, self).__init__(**kwargs) + self.query = kwargs.get('query', None) self.type = 'SalesforceMarketingCloudSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_marketing_cloud_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_marketing_cloud_source_py3.py new file mode 100644 index 000000000000..4a7e17da4f06 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_marketing_cloud_source_py3.py @@ -0,0 +1,52 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class SalesforceMarketingCloudSource(CopySource): + """A copy activity Salesforce Marketing Cloud source. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Required. Constant filled by server. + :type type: str + :param query: A query to retrieve data from source. Type: string (or + Expression with resultType string). + :type query: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'query': {'key': 'query', 'type': 'object'}, + } + + def __init__(self, *, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None, **kwargs) -> None: + super(SalesforceMarketingCloudSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, **kwargs) + self.query = query + self.type = 'SalesforceMarketingCloudSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_object_dataset.py index 1ffed292b9cc..8ea94a9f7cd1 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_object_dataset.py @@ -15,6 +15,8 @@ class SalesforceObjectDataset(Dataset): """The Salesforce object dataset. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -23,7 +25,7 @@ class SalesforceObjectDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object - :param linked_service_name: Linked service reference. + :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference :param parameters: Parameters for dataset. @@ -32,7 +34,7 @@ class SalesforceObjectDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param object_api_name: The Salesforce object API name. Type: string (or Expression with resultType string). @@ -55,7 +57,7 @@ class SalesforceObjectDataset(Dataset): 'object_api_name': {'key': 'typeProperties.objectApiName', 'type': 'object'}, } - def __init__(self, linked_service_name, additional_properties=None, description=None, structure=None, parameters=None, annotations=None, object_api_name=None): - super(SalesforceObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations) - self.object_api_name = object_api_name + def __init__(self, **kwargs): + super(SalesforceObjectDataset, self).__init__(**kwargs) + self.object_api_name = kwargs.get('object_api_name', None) self.type = 'SalesforceObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_object_dataset_py3.py new file mode 100644 index 000000000000..ccebb62fd5b3 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_object_dataset_py3.py @@ -0,0 +1,63 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class SalesforceObjectDataset(Dataset): + """The Salesforce object dataset. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Required. Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param object_api_name: The Salesforce object API name. Type: string (or + Expression with resultType string). + :type object_api_name: object + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'object_api_name': {'key': 'typeProperties.objectApiName', 'type': 'object'}, + } + + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, object_api_name=None, **kwargs) -> None: + super(SalesforceObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + self.object_api_name = object_api_name + self.type = 'SalesforceObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_sink.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_sink.py index 4df2ac10fc49..525aaccd49be 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_sink.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_sink.py @@ -15,6 +15,8 @@ class SalesforceSink(CopySink): """A copy activity Salesforce sink. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -32,7 +34,7 @@ class SalesforceSink(CopySink): resultType string), pattern: ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type sink_retry_wait: object - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param write_behavior: The write behavior for the operation. Default is Insert. Possible values include: 'Insert', 'Upsert' @@ -69,9 +71,9 @@ class SalesforceSink(CopySink): 'ignore_null_values': {'key': 'ignoreNullValues', 'type': 'object'}, } - def __init__(self, additional_properties=None, write_batch_size=None, write_batch_timeout=None, sink_retry_count=None, sink_retry_wait=None, write_behavior=None, external_id_field_name=None, ignore_null_values=None): - super(SalesforceSink, self).__init__(additional_properties=additional_properties, write_batch_size=write_batch_size, write_batch_timeout=write_batch_timeout, sink_retry_count=sink_retry_count, sink_retry_wait=sink_retry_wait) - self.write_behavior = write_behavior - self.external_id_field_name = external_id_field_name - self.ignore_null_values = ignore_null_values + def __init__(self, **kwargs): + super(SalesforceSink, self).__init__(**kwargs) + self.write_behavior = kwargs.get('write_behavior', None) + self.external_id_field_name = kwargs.get('external_id_field_name', None) + self.ignore_null_values = kwargs.get('ignore_null_values', None) self.type = 'SalesforceSink' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_sink_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_sink_py3.py new file mode 100644 index 000000000000..10f9aea9c69d --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_sink_py3.py @@ -0,0 +1,79 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_sink import CopySink + + +class SalesforceSink(CopySink): + """A copy activity Salesforce sink. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param write_batch_size: Write batch size. Type: integer (or Expression + with resultType integer), minimum: 0. + :type write_batch_size: object + :param write_batch_timeout: Write batch timeout. Type: string (or + Expression with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type write_batch_timeout: object + :param sink_retry_count: Sink retry count. Type: integer (or Expression + with resultType integer). + :type sink_retry_count: object + :param sink_retry_wait: Sink retry wait. Type: string (or Expression with + resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type sink_retry_wait: object + :param type: Required. Constant filled by server. + :type type: str + :param write_behavior: The write behavior for the operation. Default is + Insert. Possible values include: 'Insert', 'Upsert' + :type write_behavior: str or + ~azure.mgmt.datafactory.models.SalesforceSinkWriteBehavior + :param external_id_field_name: The name of the external ID field for + upsert operation. Default value is 'Id' column. Type: string (or + Expression with resultType string). + :type external_id_field_name: object + :param ignore_null_values: The flag indicating whether or not to ignore + null values from input dataset (except key fields) during write operation. + Default value is false. If set it to true, it means ADF will leave the + data in the destination object unchanged when doing upsert/update + operation and insert defined default value when doing insert operation, + versus ADF will update the data in the destination object to NULL when + doing upsert/update operation and insert NULL value when doing insert + operation. Type: boolean (or Expression with resultType boolean). + :type ignore_null_values: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, + 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, + 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, + 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'write_behavior': {'key': 'writeBehavior', 'type': 'str'}, + 'external_id_field_name': {'key': 'externalIdFieldName', 'type': 'object'}, + 'ignore_null_values': {'key': 'ignoreNullValues', 'type': 'object'}, + } + + def __init__(self, *, additional_properties=None, write_batch_size=None, write_batch_timeout=None, sink_retry_count=None, sink_retry_wait=None, write_behavior=None, external_id_field_name=None, ignore_null_values=None, **kwargs) -> None: + super(SalesforceSink, self).__init__(additional_properties=additional_properties, write_batch_size=write_batch_size, write_batch_timeout=write_batch_timeout, sink_retry_count=sink_retry_count, sink_retry_wait=sink_retry_wait, **kwargs) + self.write_behavior = write_behavior + self.external_id_field_name = external_id_field_name + self.ignore_null_values = ignore_null_values + self.type = 'SalesforceSink' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_source.py index 40507b7c10a5..8442a716c842 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_source.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_source.py @@ -15,6 +15,8 @@ class SalesforceSource(CopySource): """A copy activity Salesforce source. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -25,7 +27,7 @@ class SalesforceSource(CopySource): with resultType string), pattern: ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type source_retry_wait: object - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param query: Database query. Type: string (or Expression with resultType string). @@ -49,8 +51,8 @@ class SalesforceSource(CopySource): 'read_behavior': {'key': 'readBehavior', 'type': 'str'}, } - def __init__(self, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None, read_behavior=None): - super(SalesforceSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait) - self.query = query - self.read_behavior = read_behavior + def __init__(self, **kwargs): + super(SalesforceSource, self).__init__(**kwargs) + self.query = kwargs.get('query', None) + self.read_behavior = kwargs.get('read_behavior', None) self.type = 'SalesforceSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_source_py3.py new file mode 100644 index 000000000000..729c342b37f7 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_source_py3.py @@ -0,0 +1,58 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class SalesforceSource(CopySource): + """A copy activity Salesforce source. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Required. Constant filled by server. + :type type: str + :param query: Database query. Type: string (or Expression with resultType + string). + :type query: object + :param read_behavior: The read behavior for the operation. Default is + Query. Possible values include: 'Query', 'QueryAll' + :type read_behavior: str or + ~azure.mgmt.datafactory.models.SalesforceSourceReadBehavior + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'query': {'key': 'query', 'type': 'object'}, + 'read_behavior': {'key': 'readBehavior', 'type': 'str'}, + } + + def __init__(self, *, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None, read_behavior=None, **kwargs) -> None: + super(SalesforceSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, **kwargs) + self.query = query + self.read_behavior = read_behavior + self.type = 'SalesforceSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_bw_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_bw_linked_service.py index cc6245496b58..2fbb906559bc 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_bw_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_bw_linked_service.py @@ -15,6 +15,8 @@ class SapBWLinkedService(LinkedService): """SAP Business Warehouse Linked Service. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -29,18 +31,18 @@ class SapBWLinkedService(LinkedService): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param server: Host name of the SAP BW instance. Type: string (or - Expression with resultType string). + :param server: Required. Host name of the SAP BW instance. Type: string + (or Expression with resultType string). :type server: object - :param system_number: System number of the BW system. (Usually a two-digit - decimal number represented as a string.) Type: string (or Expression with - resultType string). - :type system_number: object - :param client_id: Client ID of the client on the BW system. (Usually a - three-digit decimal number represented as a string) Type: string (or + :param system_number: Required. System number of the BW system. (Usually a + two-digit decimal number represented as a string.) Type: string (or Expression with resultType string). + :type system_number: object + :param client_id: Required. Client ID of the client on the BW system. + (Usually a three-digit decimal number represented as a string) Type: + string (or Expression with resultType string). :type client_id: object :param user_name: Username to access the SAP BW server. Type: string (or Expression with resultType string). @@ -75,12 +77,12 @@ class SapBWLinkedService(LinkedService): 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, server, system_number, client_id, additional_properties=None, connect_via=None, description=None, parameters=None, annotations=None, user_name=None, password=None, encrypted_credential=None): - super(SapBWLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations) - self.server = server - self.system_number = system_number - self.client_id = client_id - self.user_name = user_name - self.password = password - self.encrypted_credential = encrypted_credential + def __init__(self, **kwargs): + super(SapBWLinkedService, self).__init__(**kwargs) + self.server = kwargs.get('server', None) + self.system_number = kwargs.get('system_number', None) + self.client_id = kwargs.get('client_id', None) + self.user_name = kwargs.get('user_name', None) + self.password = kwargs.get('password', None) + self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'SapBW' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_bw_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_bw_linked_service_py3.py new file mode 100644 index 000000000000..a3e686b16405 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_bw_linked_service_py3.py @@ -0,0 +1,88 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class SapBWLinkedService(LinkedService): + """SAP Business Warehouse Linked Service. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param server: Required. Host name of the SAP BW instance. Type: string + (or Expression with resultType string). + :type server: object + :param system_number: Required. System number of the BW system. (Usually a + two-digit decimal number represented as a string.) Type: string (or + Expression with resultType string). + :type system_number: object + :param client_id: Required. Client ID of the client on the BW system. + (Usually a three-digit decimal number represented as a string) Type: + string (or Expression with resultType string). + :type client_id: object + :param user_name: Username to access the SAP BW server. Type: string (or + Expression with resultType string). + :type user_name: object + :param password: Password to access the SAP BW server. + :type password: ~azure.mgmt.datafactory.models.SecretBase + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'server': {'required': True}, + 'system_number': {'required': True}, + 'client_id': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'server': {'key': 'typeProperties.server', 'type': 'object'}, + 'system_number': {'key': 'typeProperties.systemNumber', 'type': 'object'}, + 'client_id': {'key': 'typeProperties.clientId', 'type': 'object'}, + 'user_name': {'key': 'typeProperties.userName', 'type': 'object'}, + 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, *, server, system_number, client_id, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, user_name=None, password=None, encrypted_credential=None, **kwargs) -> None: + super(SapBWLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.server = server + self.system_number = system_number + self.client_id = client_id + self.user_name = user_name + self.password = password + self.encrypted_credential = encrypted_credential + self.type = 'SapBW' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_linked_service.py index 9d082074ecc1..5c9a6c2deb00 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_linked_service.py @@ -15,6 +15,8 @@ class SapCloudForCustomerLinkedService(LinkedService): """Linked service for SAP Cloud for Customer. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -29,11 +31,11 @@ class SapCloudForCustomerLinkedService(LinkedService): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param url: The URL of SAP Cloud for Customer OData API. For example, - '[https://[tenantname].crm.ondemand.com/sap/c4c/odata/v1]'. Type: string - (or Expression with resultType string). + :param url: Required. The URL of SAP Cloud for Customer OData API. For + example, '[https://[tenantname].crm.ondemand.com/sap/c4c/odata/v1]'. Type: + string (or Expression with resultType string). :type url: object :param username: The username for Basic authentication. Type: string (or Expression with resultType string). @@ -65,10 +67,10 @@ class SapCloudForCustomerLinkedService(LinkedService): 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, url, additional_properties=None, connect_via=None, description=None, parameters=None, annotations=None, username=None, password=None, encrypted_credential=None): - super(SapCloudForCustomerLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations) - self.url = url - self.username = username - self.password = password - self.encrypted_credential = encrypted_credential + def __init__(self, **kwargs): + super(SapCloudForCustomerLinkedService, self).__init__(**kwargs) + self.url = kwargs.get('url', None) + self.username = kwargs.get('username', None) + self.password = kwargs.get('password', None) + self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'SapCloudForCustomer' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_linked_service_py3.py new file mode 100644 index 000000000000..319a5f31e99a --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_linked_service_py3.py @@ -0,0 +1,76 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class SapCloudForCustomerLinkedService(LinkedService): + """Linked service for SAP Cloud for Customer. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param url: Required. The URL of SAP Cloud for Customer OData API. For + example, '[https://[tenantname].crm.ondemand.com/sap/c4c/odata/v1]'. Type: + string (or Expression with resultType string). + :type url: object + :param username: The username for Basic authentication. Type: string (or + Expression with resultType string). + :type username: object + :param password: The password for Basic authentication. + :type password: ~azure.mgmt.datafactory.models.SecretBase + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Either encryptedCredential or username/password must + be provided. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'url': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'url': {'key': 'typeProperties.url', 'type': 'object'}, + 'username': {'key': 'typeProperties.username', 'type': 'object'}, + 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, *, url, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, username=None, password=None, encrypted_credential=None, **kwargs) -> None: + super(SapCloudForCustomerLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.url = url + self.username = username + self.password = password + self.encrypted_credential = encrypted_credential + self.type = 'SapCloudForCustomer' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_resource_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_resource_dataset.py index 24da7d8224cc..65655e9ce061 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_resource_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_resource_dataset.py @@ -15,6 +15,8 @@ class SapCloudForCustomerResourceDataset(Dataset): """The path of the SAP Cloud for Customer OData entity. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -23,7 +25,7 @@ class SapCloudForCustomerResourceDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object - :param linked_service_name: Linked service reference. + :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference :param parameters: Parameters for dataset. @@ -32,10 +34,10 @@ class SapCloudForCustomerResourceDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param path: The path of the SAP Cloud for Customer OData entity. Type: - string (or Expression with resultType string). + :param path: Required. The path of the SAP Cloud for Customer OData + entity. Type: string (or Expression with resultType string). :type path: object """ @@ -56,7 +58,7 @@ class SapCloudForCustomerResourceDataset(Dataset): 'path': {'key': 'typeProperties.path', 'type': 'object'}, } - def __init__(self, linked_service_name, path, additional_properties=None, description=None, structure=None, parameters=None, annotations=None): - super(SapCloudForCustomerResourceDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations) - self.path = path + def __init__(self, **kwargs): + super(SapCloudForCustomerResourceDataset, self).__init__(**kwargs) + self.path = kwargs.get('path', None) self.type = 'SapCloudForCustomerResource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_resource_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_resource_dataset_py3.py new file mode 100644 index 000000000000..4a7c19252dd1 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_resource_dataset_py3.py @@ -0,0 +1,64 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class SapCloudForCustomerResourceDataset(Dataset): + """The path of the SAP Cloud for Customer OData entity. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Required. Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param path: Required. The path of the SAP Cloud for Customer OData + entity. Type: string (or Expression with resultType string). + :type path: object + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + 'path': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'path': {'key': 'typeProperties.path', 'type': 'object'}, + } + + def __init__(self, *, linked_service_name, path, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: + super(SapCloudForCustomerResourceDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + self.path = path + self.type = 'SapCloudForCustomerResource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_sink.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_sink.py index 40ad3d1aab20..05d98ec70eaa 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_sink.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_sink.py @@ -15,6 +15,8 @@ class SapCloudForCustomerSink(CopySink): """A copy activity SAP Cloud for Customer sink. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -32,7 +34,7 @@ class SapCloudForCustomerSink(CopySink): resultType string), pattern: ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type sink_retry_wait: object - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param write_behavior: The write behavior for the operation. Default is 'Insert'. Possible values include: 'Insert', 'Update' @@ -54,7 +56,7 @@ class SapCloudForCustomerSink(CopySink): 'write_behavior': {'key': 'writeBehavior', 'type': 'str'}, } - def __init__(self, additional_properties=None, write_batch_size=None, write_batch_timeout=None, sink_retry_count=None, sink_retry_wait=None, write_behavior=None): - super(SapCloudForCustomerSink, self).__init__(additional_properties=additional_properties, write_batch_size=write_batch_size, write_batch_timeout=write_batch_timeout, sink_retry_count=sink_retry_count, sink_retry_wait=sink_retry_wait) - self.write_behavior = write_behavior + def __init__(self, **kwargs): + super(SapCloudForCustomerSink, self).__init__(**kwargs) + self.write_behavior = kwargs.get('write_behavior', None) self.type = 'SapCloudForCustomerSink' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_sink_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_sink_py3.py new file mode 100644 index 000000000000..c366c6c00a4c --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_sink_py3.py @@ -0,0 +1,62 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_sink import CopySink + + +class SapCloudForCustomerSink(CopySink): + """A copy activity SAP Cloud for Customer sink. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param write_batch_size: Write batch size. Type: integer (or Expression + with resultType integer), minimum: 0. + :type write_batch_size: object + :param write_batch_timeout: Write batch timeout. Type: string (or + Expression with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type write_batch_timeout: object + :param sink_retry_count: Sink retry count. Type: integer (or Expression + with resultType integer). + :type sink_retry_count: object + :param sink_retry_wait: Sink retry wait. Type: string (or Expression with + resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type sink_retry_wait: object + :param type: Required. Constant filled by server. + :type type: str + :param write_behavior: The write behavior for the operation. Default is + 'Insert'. Possible values include: 'Insert', 'Update' + :type write_behavior: str or + ~azure.mgmt.datafactory.models.SapCloudForCustomerSinkWriteBehavior + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, + 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, + 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, + 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'write_behavior': {'key': 'writeBehavior', 'type': 'str'}, + } + + def __init__(self, *, additional_properties=None, write_batch_size=None, write_batch_timeout=None, sink_retry_count=None, sink_retry_wait=None, write_behavior=None, **kwargs) -> None: + super(SapCloudForCustomerSink, self).__init__(additional_properties=additional_properties, write_batch_size=write_batch_size, write_batch_timeout=write_batch_timeout, sink_retry_count=sink_retry_count, sink_retry_wait=sink_retry_wait, **kwargs) + self.write_behavior = write_behavior + self.type = 'SapCloudForCustomerSink' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_source.py index 771684a97c56..c8dedf91e188 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_source.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_source.py @@ -15,6 +15,8 @@ class SapCloudForCustomerSource(CopySource): """A copy activity source for SAP Cloud for Customer source. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -25,7 +27,7 @@ class SapCloudForCustomerSource(CopySource): with resultType string), pattern: ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type source_retry_wait: object - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param query: SAP Cloud for Customer OData query. For example, "$top=1". Type: string (or Expression with resultType string). @@ -44,7 +46,7 @@ class SapCloudForCustomerSource(CopySource): 'query': {'key': 'query', 'type': 'object'}, } - def __init__(self, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None): - super(SapCloudForCustomerSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait) - self.query = query + def __init__(self, **kwargs): + super(SapCloudForCustomerSource, self).__init__(**kwargs) + self.query = kwargs.get('query', None) self.type = 'SapCloudForCustomerSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_source_py3.py new file mode 100644 index 000000000000..8edc564d0b8d --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_source_py3.py @@ -0,0 +1,52 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class SapCloudForCustomerSource(CopySource): + """A copy activity source for SAP Cloud for Customer source. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Required. Constant filled by server. + :type type: str + :param query: SAP Cloud for Customer OData query. For example, "$top=1". + Type: string (or Expression with resultType string). + :type query: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'query': {'key': 'query', 'type': 'object'}, + } + + def __init__(self, *, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None, **kwargs) -> None: + super(SapCloudForCustomerSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, **kwargs) + self.query = query + self.type = 'SapCloudForCustomerSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_ecc_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_ecc_linked_service.py index b00caa936d75..4303b2f9cbca 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_ecc_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_ecc_linked_service.py @@ -15,6 +15,8 @@ class SapEccLinkedService(LinkedService): """Linked service for SAP ERP Central Component(SAP ECC). + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -29,9 +31,9 @@ class SapEccLinkedService(LinkedService): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param url: The URL of SAP ECC OData API. For example, + :param url: Required. The URL of SAP ECC OData API. For example, '[https://hostname:port/sap/opu/odata/sap/servicename/]'. Type: string (or Expression with resultType string). :type url: str @@ -65,10 +67,10 @@ class SapEccLinkedService(LinkedService): 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'str'}, } - def __init__(self, url, additional_properties=None, connect_via=None, description=None, parameters=None, annotations=None, username=None, password=None, encrypted_credential=None): - super(SapEccLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations) - self.url = url - self.username = username - self.password = password - self.encrypted_credential = encrypted_credential + def __init__(self, **kwargs): + super(SapEccLinkedService, self).__init__(**kwargs) + self.url = kwargs.get('url', None) + self.username = kwargs.get('username', None) + self.password = kwargs.get('password', None) + self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'SapEcc' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_ecc_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_ecc_linked_service_py3.py new file mode 100644 index 000000000000..ef2fe2e60573 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_ecc_linked_service_py3.py @@ -0,0 +1,76 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class SapEccLinkedService(LinkedService): + """Linked service for SAP ERP Central Component(SAP ECC). + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param url: Required. The URL of SAP ECC OData API. For example, + '[https://hostname:port/sap/opu/odata/sap/servicename/]'. Type: string (or + Expression with resultType string). + :type url: str + :param username: The username for Basic authentication. Type: string (or + Expression with resultType string). + :type username: str + :param password: The password for Basic authentication. + :type password: ~azure.mgmt.datafactory.models.SecretBase + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Either encryptedCredential or username/password must + be provided. Type: string (or Expression with resultType string). + :type encrypted_credential: str + """ + + _validation = { + 'type': {'required': True}, + 'url': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'url': {'key': 'typeProperties.url', 'type': 'str'}, + 'username': {'key': 'typeProperties.username', 'type': 'str'}, + 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'str'}, + } + + def __init__(self, *, url: str, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, username: str=None, password=None, encrypted_credential: str=None, **kwargs) -> None: + super(SapEccLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.url = url + self.username = username + self.password = password + self.encrypted_credential = encrypted_credential + self.type = 'SapEcc' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_ecc_resource_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_ecc_resource_dataset.py index 9e0f4fdb5baa..befc8ea726f9 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_ecc_resource_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_ecc_resource_dataset.py @@ -15,6 +15,8 @@ class SapEccResourceDataset(Dataset): """The path of the SAP ECC OData entity. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -23,7 +25,7 @@ class SapEccResourceDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object - :param linked_service_name: Linked service reference. + :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference :param parameters: Parameters for dataset. @@ -32,10 +34,10 @@ class SapEccResourceDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param path: The path of the SAP ECC OData entity. Type: string (or - Expression with resultType string). + :param path: Required. The path of the SAP ECC OData entity. Type: string + (or Expression with resultType string). :type path: str """ @@ -56,7 +58,7 @@ class SapEccResourceDataset(Dataset): 'path': {'key': 'typeProperties.path', 'type': 'str'}, } - def __init__(self, linked_service_name, path, additional_properties=None, description=None, structure=None, parameters=None, annotations=None): - super(SapEccResourceDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations) - self.path = path + def __init__(self, **kwargs): + super(SapEccResourceDataset, self).__init__(**kwargs) + self.path = kwargs.get('path', None) self.type = 'SapEccResource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_ecc_resource_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_ecc_resource_dataset_py3.py new file mode 100644 index 000000000000..321d4d85bc9a --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_ecc_resource_dataset_py3.py @@ -0,0 +1,64 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class SapEccResourceDataset(Dataset): + """The path of the SAP ECC OData entity. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Required. Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param path: Required. The path of the SAP ECC OData entity. Type: string + (or Expression with resultType string). + :type path: str + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + 'path': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'path': {'key': 'typeProperties.path', 'type': 'str'}, + } + + def __init__(self, *, linked_service_name, path: str, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: + super(SapEccResourceDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + self.path = path + self.type = 'SapEccResource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_ecc_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_ecc_source.py index 2baa90e9342d..84aa047e6d8a 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_ecc_source.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_ecc_source.py @@ -15,6 +15,8 @@ class SapEccSource(CopySource): """A copy activity source for SAP ECC source. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -25,7 +27,7 @@ class SapEccSource(CopySource): with resultType string), pattern: ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type source_retry_wait: object - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param query: SAP ECC OData query. For example, "$top=1". Type: string (or Expression with resultType string). @@ -44,7 +46,7 @@ class SapEccSource(CopySource): 'query': {'key': 'query', 'type': 'str'}, } - def __init__(self, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None): - super(SapEccSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait) - self.query = query + def __init__(self, **kwargs): + super(SapEccSource, self).__init__(**kwargs) + self.query = kwargs.get('query', None) self.type = 'SapEccSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_ecc_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_ecc_source_py3.py new file mode 100644 index 000000000000..e7ec1d6b495b --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_ecc_source_py3.py @@ -0,0 +1,52 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class SapEccSource(CopySource): + """A copy activity source for SAP ECC source. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Required. Constant filled by server. + :type type: str + :param query: SAP ECC OData query. For example, "$top=1". Type: string (or + Expression with resultType string). + :type query: str + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'query': {'key': 'query', 'type': 'str'}, + } + + def __init__(self, *, additional_properties=None, source_retry_count=None, source_retry_wait=None, query: str=None, **kwargs) -> None: + super(SapEccSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, **kwargs) + self.query = query + self.type = 'SapEccSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_hana_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_hana_linked_service.py index 9a205852d841..0c2dbec28558 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_hana_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_hana_linked_service.py @@ -15,6 +15,8 @@ class SapHanaLinkedService(LinkedService): """SAP HANA Linked Service. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -29,10 +31,10 @@ class SapHanaLinkedService(LinkedService): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param server: Host name of the SAP HANA server. Type: string (or - Expression with resultType string). + :param server: Required. Host name of the SAP HANA server. Type: string + (or Expression with resultType string). :type server: object :param authentication_type: The authentication type to be used to connect to the SAP HANA server. Possible values include: 'Basic', 'Windows' @@ -68,11 +70,11 @@ class SapHanaLinkedService(LinkedService): 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, server, additional_properties=None, connect_via=None, description=None, parameters=None, annotations=None, authentication_type=None, user_name=None, password=None, encrypted_credential=None): - super(SapHanaLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations) - self.server = server - self.authentication_type = authentication_type - self.user_name = user_name - self.password = password - self.encrypted_credential = encrypted_credential + def __init__(self, **kwargs): + super(SapHanaLinkedService, self).__init__(**kwargs) + self.server = kwargs.get('server', None) + self.authentication_type = kwargs.get('authentication_type', None) + self.user_name = kwargs.get('user_name', None) + self.password = kwargs.get('password', None) + self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'SapHana' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_hana_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_hana_linked_service_py3.py new file mode 100644 index 000000000000..60c8d0283d61 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_hana_linked_service_py3.py @@ -0,0 +1,80 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class SapHanaLinkedService(LinkedService): + """SAP HANA Linked Service. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param server: Required. Host name of the SAP HANA server. Type: string + (or Expression with resultType string). + :type server: object + :param authentication_type: The authentication type to be used to connect + to the SAP HANA server. Possible values include: 'Basic', 'Windows' + :type authentication_type: str or + ~azure.mgmt.datafactory.models.SapHanaAuthenticationType + :param user_name: Username to access the SAP HANA server. Type: string (or + Expression with resultType string). + :type user_name: object + :param password: Password to access the SAP HANA server. + :type password: ~azure.mgmt.datafactory.models.SecretBase + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'server': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'server': {'key': 'typeProperties.server', 'type': 'object'}, + 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, + 'user_name': {'key': 'typeProperties.userName', 'type': 'object'}, + 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, *, server, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, authentication_type=None, user_name=None, password=None, encrypted_credential=None, **kwargs) -> None: + super(SapHanaLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.server = server + self.authentication_type = authentication_type + self.user_name = user_name + self.password = password + self.encrypted_credential = encrypted_credential + self.type = 'SapHana' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/schedule_trigger.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/schedule_trigger.py index 2ed6c2c1045b..eaebfb4c2553 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/schedule_trigger.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/schedule_trigger.py @@ -18,6 +18,8 @@ class ScheduleTrigger(MultiplePipelineTrigger): Variables are only populated by the server, and will be ignored when sending a request. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -28,12 +30,12 @@ class ScheduleTrigger(MultiplePipelineTrigger): 'Started', 'Stopped', 'Disabled' :vartype runtime_state: str or ~azure.mgmt.datafactory.models.TriggerRuntimeState - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param pipelines: Pipelines that need to be started. :type pipelines: list[~azure.mgmt.datafactory.models.TriggerPipelineReference] - :param recurrence: Recurrence schedule configuration. + :param recurrence: Required. Recurrence schedule configuration. :type recurrence: ~azure.mgmt.datafactory.models.ScheduleTriggerRecurrence """ @@ -52,7 +54,7 @@ class ScheduleTrigger(MultiplePipelineTrigger): 'recurrence': {'key': 'typeProperties.recurrence', 'type': 'ScheduleTriggerRecurrence'}, } - def __init__(self, recurrence, additional_properties=None, description=None, pipelines=None): - super(ScheduleTrigger, self).__init__(additional_properties=additional_properties, description=description, pipelines=pipelines) - self.recurrence = recurrence + def __init__(self, **kwargs): + super(ScheduleTrigger, self).__init__(**kwargs) + self.recurrence = kwargs.get('recurrence', None) self.type = 'ScheduleTrigger' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/schedule_trigger_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/schedule_trigger_py3.py new file mode 100644 index 000000000000..217195a5930f --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/schedule_trigger_py3.py @@ -0,0 +1,60 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .multiple_pipeline_trigger import MultiplePipelineTrigger + + +class ScheduleTrigger(MultiplePipelineTrigger): + """Trigger that creates pipeline runs periodically, on schedule. + + Variables are only populated by the server, and will be ignored when + sending a request. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param description: Trigger description. + :type description: str + :ivar runtime_state: Indicates if trigger is running or not. Updated when + Start/Stop APIs are called on the Trigger. Possible values include: + 'Started', 'Stopped', 'Disabled' + :vartype runtime_state: str or + ~azure.mgmt.datafactory.models.TriggerRuntimeState + :param type: Required. Constant filled by server. + :type type: str + :param pipelines: Pipelines that need to be started. + :type pipelines: + list[~azure.mgmt.datafactory.models.TriggerPipelineReference] + :param recurrence: Required. Recurrence schedule configuration. + :type recurrence: ~azure.mgmt.datafactory.models.ScheduleTriggerRecurrence + """ + + _validation = { + 'runtime_state': {'readonly': True}, + 'type': {'required': True}, + 'recurrence': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'runtime_state': {'key': 'runtimeState', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'str'}, + 'pipelines': {'key': 'pipelines', 'type': '[TriggerPipelineReference]'}, + 'recurrence': {'key': 'typeProperties.recurrence', 'type': 'ScheduleTriggerRecurrence'}, + } + + def __init__(self, *, recurrence, additional_properties=None, description: str=None, pipelines=None, **kwargs) -> None: + super(ScheduleTrigger, self).__init__(additional_properties=additional_properties, description=description, pipelines=pipelines, **kwargs) + self.recurrence = recurrence + self.type = 'ScheduleTrigger' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/schedule_trigger_recurrence.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/schedule_trigger_recurrence.py index 021ad0afeb80..85408c45547b 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/schedule_trigger_recurrence.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/schedule_trigger_recurrence.py @@ -43,12 +43,12 @@ class ScheduleTriggerRecurrence(Model): 'schedule': {'key': 'schedule', 'type': 'RecurrenceSchedule'}, } - def __init__(self, additional_properties=None, frequency=None, interval=None, start_time=None, end_time=None, time_zone=None, schedule=None): - super(ScheduleTriggerRecurrence, self).__init__() - self.additional_properties = additional_properties - self.frequency = frequency - self.interval = interval - self.start_time = start_time - self.end_time = end_time - self.time_zone = time_zone - self.schedule = schedule + def __init__(self, **kwargs): + super(ScheduleTriggerRecurrence, self).__init__(**kwargs) + self.additional_properties = kwargs.get('additional_properties', None) + self.frequency = kwargs.get('frequency', None) + self.interval = kwargs.get('interval', None) + self.start_time = kwargs.get('start_time', None) + self.end_time = kwargs.get('end_time', None) + self.time_zone = kwargs.get('time_zone', None) + self.schedule = kwargs.get('schedule', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/schedule_trigger_recurrence_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/schedule_trigger_recurrence_py3.py new file mode 100644 index 000000000000..a9b6eded7b96 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/schedule_trigger_recurrence_py3.py @@ -0,0 +1,54 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class ScheduleTriggerRecurrence(Model): + """The workflow trigger recurrence. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param frequency: The frequency. Possible values include: 'NotSpecified', + 'Minute', 'Hour', 'Day', 'Week', 'Month', 'Year' + :type frequency: str or ~azure.mgmt.datafactory.models.RecurrenceFrequency + :param interval: The interval. + :type interval: int + :param start_time: The start time. + :type start_time: datetime + :param end_time: The end time. + :type end_time: datetime + :param time_zone: The time zone. + :type time_zone: str + :param schedule: The recurrence schedule. + :type schedule: ~azure.mgmt.datafactory.models.RecurrenceSchedule + """ + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'frequency': {'key': 'frequency', 'type': 'str'}, + 'interval': {'key': 'interval', 'type': 'int'}, + 'start_time': {'key': 'startTime', 'type': 'iso-8601'}, + 'end_time': {'key': 'endTime', 'type': 'iso-8601'}, + 'time_zone': {'key': 'timeZone', 'type': 'str'}, + 'schedule': {'key': 'schedule', 'type': 'RecurrenceSchedule'}, + } + + def __init__(self, *, additional_properties=None, frequency=None, interval: int=None, start_time=None, end_time=None, time_zone: str=None, schedule=None, **kwargs) -> None: + super(ScheduleTriggerRecurrence, self).__init__(**kwargs) + self.additional_properties = additional_properties + self.frequency = frequency + self.interval = interval + self.start_time = start_time + self.end_time = end_time + self.time_zone = time_zone + self.schedule = schedule diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/secret_base.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/secret_base.py index e7875d601f55..3d9475dd4382 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/secret_base.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/secret_base.py @@ -18,7 +18,9 @@ class SecretBase(Model): You probably want to use the sub-classes and not this class directly. Known sub-classes are: SecureString, AzureKeyVaultSecretReference - :param type: Constant filled by server. + All required parameters must be populated in order to send to Azure. + + :param type: Required. Constant filled by server. :type type: str """ @@ -34,6 +36,6 @@ class SecretBase(Model): 'type': {'SecureString': 'SecureString', 'AzureKeyVaultSecret': 'AzureKeyVaultSecretReference'} } - def __init__(self): - super(SecretBase, self).__init__() + def __init__(self, **kwargs): + super(SecretBase, self).__init__(**kwargs) self.type = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/secret_base_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/secret_base_py3.py new file mode 100644 index 000000000000..29403e61b245 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/secret_base_py3.py @@ -0,0 +1,41 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class SecretBase(Model): + """The base definition of a secret type. + + You probably want to use the sub-classes and not this class directly. Known + sub-classes are: SecureString, AzureKeyVaultSecretReference + + All required parameters must be populated in order to send to Azure. + + :param type: Required. Constant filled by server. + :type type: str + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'type': {'key': 'type', 'type': 'str'}, + } + + _subtype_map = { + 'type': {'SecureString': 'SecureString', 'AzureKeyVaultSecret': 'AzureKeyVaultSecretReference'} + } + + def __init__(self, **kwargs) -> None: + super(SecretBase, self).__init__(**kwargs) + self.type = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/secure_string.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/secure_string.py index 3cc6e7630fca..bec430fdf8a4 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/secure_string.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/secure_string.py @@ -16,9 +16,11 @@ class SecureString(SecretBase): """Azure Data Factory secure string definition. The string value will be masked with asterisks '*' during Get or List API calls. - :param type: Constant filled by server. + All required parameters must be populated in order to send to Azure. + + :param type: Required. Constant filled by server. :type type: str - :param value: Value of secure string. + :param value: Required. Value of secure string. :type value: str """ @@ -32,7 +34,7 @@ class SecureString(SecretBase): 'value': {'key': 'value', 'type': 'str'}, } - def __init__(self, value): - super(SecureString, self).__init__() - self.value = value + def __init__(self, **kwargs): + super(SecureString, self).__init__(**kwargs) + self.value = kwargs.get('value', None) self.type = 'SecureString' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/secure_string_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/secure_string_py3.py new file mode 100644 index 000000000000..1ca20053acd1 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/secure_string_py3.py @@ -0,0 +1,40 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .secret_base import SecretBase + + +class SecureString(SecretBase): + """Azure Data Factory secure string definition. The string value will be + masked with asterisks '*' during Get or List API calls. + + All required parameters must be populated in order to send to Azure. + + :param type: Required. Constant filled by server. + :type type: str + :param value: Required. Value of secure string. + :type value: str + """ + + _validation = { + 'type': {'required': True}, + 'value': {'required': True}, + } + + _attribute_map = { + 'type': {'key': 'type', 'type': 'str'}, + 'value': {'key': 'value', 'type': 'str'}, + } + + def __init__(self, *, value: str, **kwargs) -> None: + super(SecureString, self).__init__(, **kwargs) + self.value = value + self.type = 'SecureString' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime.py index 1614e0aec96c..d2b334ee11b5 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime.py @@ -15,12 +15,14 @@ class SelfHostedIntegrationRuntime(IntegrationRuntime): """Self-hosted integration runtime. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] :param description: Integration runtime description. :type description: str - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param linked_info: :type linked_info: @@ -38,7 +40,7 @@ class SelfHostedIntegrationRuntime(IntegrationRuntime): 'linked_info': {'key': 'typeProperties.linkedInfo', 'type': 'LinkedIntegrationRuntimeProperties'}, } - def __init__(self, additional_properties=None, description=None, linked_info=None): - super(SelfHostedIntegrationRuntime, self).__init__(additional_properties=additional_properties, description=description) - self.linked_info = linked_info + def __init__(self, **kwargs): + super(SelfHostedIntegrationRuntime, self).__init__(**kwargs) + self.linked_info = kwargs.get('linked_info', None) self.type = 'SelfHosted' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime_node.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime_node.py index 21d23ca9f9de..84b8586a3c8f 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime_node.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime_node.py @@ -116,9 +116,9 @@ class SelfHostedIntegrationRuntimeNode(Model): 'max_concurrent_jobs': {'key': 'maxConcurrentJobs', 'type': 'int'}, } - def __init__(self, additional_properties=None): - super(SelfHostedIntegrationRuntimeNode, self).__init__() - self.additional_properties = additional_properties + def __init__(self, **kwargs): + super(SelfHostedIntegrationRuntimeNode, self).__init__(**kwargs) + self.additional_properties = kwargs.get('additional_properties', None) self.node_name = None self.machine_name = None self.host_service_uri = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime_node_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime_node_py3.py new file mode 100644 index 000000000000..1fd742f9ffcb --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime_node_py3.py @@ -0,0 +1,139 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class SelfHostedIntegrationRuntimeNode(Model): + """Properties of Self-hosted integration runtime node. + + Variables are only populated by the server, and will be ignored when + sending a request. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :ivar node_name: Name of the integration runtime node. + :vartype node_name: str + :ivar machine_name: Machine name of the integration runtime node. + :vartype machine_name: str + :ivar host_service_uri: URI for the host machine of the integration + runtime. + :vartype host_service_uri: str + :ivar status: Status of the integration runtime node. Possible values + include: 'NeedRegistration', 'Online', 'Limited', 'Offline', 'Upgrading', + 'Initializing', 'InitializeFailed' + :vartype status: str or + ~azure.mgmt.datafactory.models.SelfHostedIntegrationRuntimeNodeStatus + :ivar capabilities: The integration runtime capabilities dictionary + :vartype capabilities: dict[str, str] + :ivar version_status: Status of the integration runtime node version. + :vartype version_status: str + :ivar version: Version of the integration runtime node. + :vartype version: str + :ivar register_time: The time at which the integration runtime node was + registered in ISO8601 format. + :vartype register_time: datetime + :ivar last_connect_time: The most recent time at which the integration + runtime was connected in ISO8601 format. + :vartype last_connect_time: datetime + :ivar expiry_time: The time at which the integration runtime will expire + in ISO8601 format. + :vartype expiry_time: datetime + :ivar last_start_time: The time the node last started up. + :vartype last_start_time: datetime + :ivar last_stop_time: The integration runtime node last stop time. + :vartype last_stop_time: datetime + :ivar last_update_result: The result of the last integration runtime node + update. Possible values include: 'Succeed', 'Fail' + :vartype last_update_result: str or + ~azure.mgmt.datafactory.models.IntegrationRuntimeUpdateResult + :ivar last_start_update_time: The last time for the integration runtime + node update start. + :vartype last_start_update_time: datetime + :ivar last_end_update_time: The last time for the integration runtime node + update end. + :vartype last_end_update_time: datetime + :ivar is_active_dispatcher: Indicates whether this node is the active + dispatcher for integration runtime requests. + :vartype is_active_dispatcher: bool + :ivar concurrent_jobs_limit: Maximum concurrent jobs on the integration + runtime node. + :vartype concurrent_jobs_limit: int + :ivar max_concurrent_jobs: The maximum concurrent jobs in this integration + runtime. + :vartype max_concurrent_jobs: int + """ + + _validation = { + 'node_name': {'readonly': True}, + 'machine_name': {'readonly': True}, + 'host_service_uri': {'readonly': True}, + 'status': {'readonly': True}, + 'capabilities': {'readonly': True}, + 'version_status': {'readonly': True}, + 'version': {'readonly': True}, + 'register_time': {'readonly': True}, + 'last_connect_time': {'readonly': True}, + 'expiry_time': {'readonly': True}, + 'last_start_time': {'readonly': True}, + 'last_stop_time': {'readonly': True}, + 'last_update_result': {'readonly': True}, + 'last_start_update_time': {'readonly': True}, + 'last_end_update_time': {'readonly': True}, + 'is_active_dispatcher': {'readonly': True}, + 'concurrent_jobs_limit': {'readonly': True}, + 'max_concurrent_jobs': {'readonly': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'node_name': {'key': 'nodeName', 'type': 'str'}, + 'machine_name': {'key': 'machineName', 'type': 'str'}, + 'host_service_uri': {'key': 'hostServiceUri', 'type': 'str'}, + 'status': {'key': 'status', 'type': 'str'}, + 'capabilities': {'key': 'capabilities', 'type': '{str}'}, + 'version_status': {'key': 'versionStatus', 'type': 'str'}, + 'version': {'key': 'version', 'type': 'str'}, + 'register_time': {'key': 'registerTime', 'type': 'iso-8601'}, + 'last_connect_time': {'key': 'lastConnectTime', 'type': 'iso-8601'}, + 'expiry_time': {'key': 'expiryTime', 'type': 'iso-8601'}, + 'last_start_time': {'key': 'lastStartTime', 'type': 'iso-8601'}, + 'last_stop_time': {'key': 'lastStopTime', 'type': 'iso-8601'}, + 'last_update_result': {'key': 'lastUpdateResult', 'type': 'str'}, + 'last_start_update_time': {'key': 'lastStartUpdateTime', 'type': 'iso-8601'}, + 'last_end_update_time': {'key': 'lastEndUpdateTime', 'type': 'iso-8601'}, + 'is_active_dispatcher': {'key': 'isActiveDispatcher', 'type': 'bool'}, + 'concurrent_jobs_limit': {'key': 'concurrentJobsLimit', 'type': 'int'}, + 'max_concurrent_jobs': {'key': 'maxConcurrentJobs', 'type': 'int'}, + } + + def __init__(self, *, additional_properties=None, **kwargs) -> None: + super(SelfHostedIntegrationRuntimeNode, self).__init__(**kwargs) + self.additional_properties = additional_properties + self.node_name = None + self.machine_name = None + self.host_service_uri = None + self.status = None + self.capabilities = None + self.version_status = None + self.version = None + self.register_time = None + self.last_connect_time = None + self.expiry_time = None + self.last_start_time = None + self.last_stop_time = None + self.last_update_result = None + self.last_start_update_time = None + self.last_end_update_time = None + self.is_active_dispatcher = None + self.concurrent_jobs_limit = None + self.max_concurrent_jobs = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime_py3.py new file mode 100644 index 000000000000..3715401fa8dd --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime_py3.py @@ -0,0 +1,46 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .integration_runtime import IntegrationRuntime + + +class SelfHostedIntegrationRuntime(IntegrationRuntime): + """Self-hosted integration runtime. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param description: Integration runtime description. + :type description: str + :param type: Required. Constant filled by server. + :type type: str + :param linked_info: + :type linked_info: + ~azure.mgmt.datafactory.models.LinkedIntegrationRuntimeProperties + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'str'}, + 'linked_info': {'key': 'typeProperties.linkedInfo', 'type': 'LinkedIntegrationRuntimeProperties'}, + } + + def __init__(self, *, additional_properties=None, description: str=None, linked_info=None, **kwargs) -> None: + super(SelfHostedIntegrationRuntime, self).__init__(additional_properties=additional_properties, description=description, **kwargs) + self.linked_info = linked_info + self.type = 'SelfHosted' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime_status.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime_status.py index 6c283be0540a..7bd67718e078 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime_status.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime_status.py @@ -18,6 +18,8 @@ class SelfHostedIntegrationRuntimeStatus(IntegrationRuntimeStatus): Variables are only populated by the server, and will be ignored when sending a request. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -29,7 +31,7 @@ class SelfHostedIntegrationRuntimeStatus(IntegrationRuntimeStatus): 'NeedRegistration', 'Online', 'Limited', 'Offline' :vartype state: str or ~azure.mgmt.datafactory.models.IntegrationRuntimeState - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :ivar create_time: The time at which the integration runtime was created, in ISO8601 format. @@ -109,13 +111,13 @@ class SelfHostedIntegrationRuntimeStatus(IntegrationRuntimeStatus): 'links': {'key': 'typeProperties.links', 'type': '[LinkedIntegrationRuntime]'}, } - def __init__(self, additional_properties=None, nodes=None, links=None): - super(SelfHostedIntegrationRuntimeStatus, self).__init__(additional_properties=additional_properties) + def __init__(self, **kwargs): + super(SelfHostedIntegrationRuntimeStatus, self).__init__(**kwargs) self.create_time = None self.task_queue_id = None self.internal_channel_encryption = None self.version = None - self.nodes = nodes + self.nodes = kwargs.get('nodes', None) self.scheduled_update_date = None self.update_delay_offset = None self.local_time_zone_offset = None @@ -123,5 +125,5 @@ def __init__(self, additional_properties=None, nodes=None, links=None): self.service_urls = None self.auto_update = None self.version_status = None - self.links = links + self.links = kwargs.get('links', None) self.type = 'SelfHosted' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime_status_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime_status_py3.py new file mode 100644 index 000000000000..e88fa5fa019d --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime_status_py3.py @@ -0,0 +1,129 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .integration_runtime_status import IntegrationRuntimeStatus + + +class SelfHostedIntegrationRuntimeStatus(IntegrationRuntimeStatus): + """Self-hosted integration runtime status. + + Variables are only populated by the server, and will be ignored when + sending a request. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :ivar data_factory_name: The data factory name which the integration + runtime belong to. + :vartype data_factory_name: str + :ivar state: The state of integration runtime. Possible values include: + 'Initial', 'Stopped', 'Started', 'Starting', 'Stopping', + 'NeedRegistration', 'Online', 'Limited', 'Offline' + :vartype state: str or + ~azure.mgmt.datafactory.models.IntegrationRuntimeState + :param type: Required. Constant filled by server. + :type type: str + :ivar create_time: The time at which the integration runtime was created, + in ISO8601 format. + :vartype create_time: datetime + :ivar task_queue_id: The task queue id of the integration runtime. + :vartype task_queue_id: str + :ivar internal_channel_encryption: It is used to set the encryption mode + for node-node communication channel (when more than 2 self-hosted + integration runtime nodes exist). Possible values include: 'NotSet', + 'SslEncrypted', 'NotEncrypted' + :vartype internal_channel_encryption: str or + ~azure.mgmt.datafactory.models.IntegrationRuntimeInternalChannelEncryptionMode + :ivar version: Version of the integration runtime. + :vartype version: str + :param nodes: The list of nodes for this integration runtime. + :type nodes: + list[~azure.mgmt.datafactory.models.SelfHostedIntegrationRuntimeNode] + :ivar scheduled_update_date: The date at which the integration runtime + will be scheduled to update, in ISO8601 format. + :vartype scheduled_update_date: datetime + :ivar update_delay_offset: The time in the date scheduled by service to + update the integration runtime, e.g., PT03H is 3 hours + :vartype update_delay_offset: str + :ivar local_time_zone_offset: The local time zone offset in hours. + :vartype local_time_zone_offset: str + :ivar capabilities: Object with additional information about integration + runtime capabilities. + :vartype capabilities: dict[str, str] + :ivar service_urls: The URLs for the services used in integration runtime + backend service. + :vartype service_urls: list[str] + :ivar auto_update: Whether Self-hosted integration runtime auto update has + been turned on. Possible values include: 'On', 'Off' + :vartype auto_update: str or + ~azure.mgmt.datafactory.models.IntegrationRuntimeAutoUpdate + :ivar version_status: Status of the integration runtime version. + :vartype version_status: str + :param links: The list of linked integration runtimes that are created to + share with this integration runtime. + :type links: list[~azure.mgmt.datafactory.models.LinkedIntegrationRuntime] + """ + + _validation = { + 'data_factory_name': {'readonly': True}, + 'state': {'readonly': True}, + 'type': {'required': True}, + 'create_time': {'readonly': True}, + 'task_queue_id': {'readonly': True}, + 'internal_channel_encryption': {'readonly': True}, + 'version': {'readonly': True}, + 'scheduled_update_date': {'readonly': True}, + 'update_delay_offset': {'readonly': True}, + 'local_time_zone_offset': {'readonly': True}, + 'capabilities': {'readonly': True}, + 'service_urls': {'readonly': True}, + 'auto_update': {'readonly': True}, + 'version_status': {'readonly': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'data_factory_name': {'key': 'dataFactoryName', 'type': 'str'}, + 'state': {'key': 'state', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'str'}, + 'create_time': {'key': 'typeProperties.createTime', 'type': 'iso-8601'}, + 'task_queue_id': {'key': 'typeProperties.taskQueueId', 'type': 'str'}, + 'internal_channel_encryption': {'key': 'typeProperties.internalChannelEncryption', 'type': 'str'}, + 'version': {'key': 'typeProperties.version', 'type': 'str'}, + 'nodes': {'key': 'typeProperties.nodes', 'type': '[SelfHostedIntegrationRuntimeNode]'}, + 'scheduled_update_date': {'key': 'typeProperties.scheduledUpdateDate', 'type': 'iso-8601'}, + 'update_delay_offset': {'key': 'typeProperties.updateDelayOffset', 'type': 'str'}, + 'local_time_zone_offset': {'key': 'typeProperties.localTimeZoneOffset', 'type': 'str'}, + 'capabilities': {'key': 'typeProperties.capabilities', 'type': '{str}'}, + 'service_urls': {'key': 'typeProperties.serviceUrls', 'type': '[str]'}, + 'auto_update': {'key': 'typeProperties.autoUpdate', 'type': 'str'}, + 'version_status': {'key': 'typeProperties.versionStatus', 'type': 'str'}, + 'links': {'key': 'typeProperties.links', 'type': '[LinkedIntegrationRuntime]'}, + } + + def __init__(self, *, additional_properties=None, nodes=None, links=None, **kwargs) -> None: + super(SelfHostedIntegrationRuntimeStatus, self).__init__(additional_properties=additional_properties, **kwargs) + self.create_time = None + self.task_queue_id = None + self.internal_channel_encryption = None + self.version = None + self.nodes = nodes + self.scheduled_update_date = None + self.update_delay_offset = None + self.local_time_zone_offset = None + self.capabilities = None + self.service_urls = None + self.auto_update = None + self.version_status = None + self.links = links + self.type = 'SelfHosted' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_linked_service.py index 1a691c632e4f..0e180ce83506 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_linked_service.py @@ -15,6 +15,8 @@ class ServiceNowLinkedService(LinkedService): """ServiceNow server linked service. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -29,13 +31,13 @@ class ServiceNowLinkedService(LinkedService): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param endpoint: The endpoint of the ServiceNow server. (i.e. + :param endpoint: Required. The endpoint of the ServiceNow server. (i.e. ServiceNowData.com) :type endpoint: object - :param authentication_type: The authentication type to use. Possible - values include: 'Basic', 'OAuth2' + :param authentication_type: Required. The authentication type to use. + Possible values include: 'Basic', 'OAuth2' :type authentication_type: str or ~azure.mgmt.datafactory.models.ServiceNowAuthenticationType :param username: The user name used to connect to the ServiceNow server @@ -89,16 +91,16 @@ class ServiceNowLinkedService(LinkedService): 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, endpoint, authentication_type, additional_properties=None, connect_via=None, description=None, parameters=None, annotations=None, username=None, password=None, client_id=None, client_secret=None, use_encrypted_endpoints=None, use_host_verification=None, use_peer_verification=None, encrypted_credential=None): - super(ServiceNowLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations) - self.endpoint = endpoint - self.authentication_type = authentication_type - self.username = username - self.password = password - self.client_id = client_id - self.client_secret = client_secret - self.use_encrypted_endpoints = use_encrypted_endpoints - self.use_host_verification = use_host_verification - self.use_peer_verification = use_peer_verification - self.encrypted_credential = encrypted_credential + def __init__(self, **kwargs): + super(ServiceNowLinkedService, self).__init__(**kwargs) + self.endpoint = kwargs.get('endpoint', None) + self.authentication_type = kwargs.get('authentication_type', None) + self.username = kwargs.get('username', None) + self.password = kwargs.get('password', None) + self.client_id = kwargs.get('client_id', None) + self.client_secret = kwargs.get('client_secret', None) + self.use_encrypted_endpoints = kwargs.get('use_encrypted_endpoints', None) + self.use_host_verification = kwargs.get('use_host_verification', None) + self.use_peer_verification = kwargs.get('use_peer_verification', None) + self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'ServiceNow' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_linked_service_py3.py new file mode 100644 index 000000000000..5aed5a07e4fc --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_linked_service_py3.py @@ -0,0 +1,106 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class ServiceNowLinkedService(LinkedService): + """ServiceNow server linked service. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param endpoint: Required. The endpoint of the ServiceNow server. (i.e. + ServiceNowData.com) + :type endpoint: object + :param authentication_type: Required. The authentication type to use. + Possible values include: 'Basic', 'OAuth2' + :type authentication_type: str or + ~azure.mgmt.datafactory.models.ServiceNowAuthenticationType + :param username: The user name used to connect to the ServiceNow server + for Basic and OAuth2 authentication. + :type username: object + :param password: The password corresponding to the user name for Basic and + OAuth2 authentication. + :type password: ~azure.mgmt.datafactory.models.SecretBase + :param client_id: The client id for OAuth2 authentication. + :type client_id: object + :param client_secret: The client secret for OAuth2 authentication. + :type client_secret: ~azure.mgmt.datafactory.models.SecretBase + :param use_encrypted_endpoints: Specifies whether the data source + endpoints are encrypted using HTTPS. The default value is true. + :type use_encrypted_endpoints: object + :param use_host_verification: Specifies whether to require the host name + in the server's certificate to match the host name of the server when + connecting over SSL. The default value is true. + :type use_host_verification: object + :param use_peer_verification: Specifies whether to verify the identity of + the server when connecting over SSL. The default value is true. + :type use_peer_verification: object + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'endpoint': {'required': True}, + 'authentication_type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'endpoint': {'key': 'typeProperties.endpoint', 'type': 'object'}, + 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, + 'username': {'key': 'typeProperties.username', 'type': 'object'}, + 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, + 'client_id': {'key': 'typeProperties.clientId', 'type': 'object'}, + 'client_secret': {'key': 'typeProperties.clientSecret', 'type': 'SecretBase'}, + 'use_encrypted_endpoints': {'key': 'typeProperties.useEncryptedEndpoints', 'type': 'object'}, + 'use_host_verification': {'key': 'typeProperties.useHostVerification', 'type': 'object'}, + 'use_peer_verification': {'key': 'typeProperties.usePeerVerification', 'type': 'object'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, *, endpoint, authentication_type, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, username=None, password=None, client_id=None, client_secret=None, use_encrypted_endpoints=None, use_host_verification=None, use_peer_verification=None, encrypted_credential=None, **kwargs) -> None: + super(ServiceNowLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.endpoint = endpoint + self.authentication_type = authentication_type + self.username = username + self.password = password + self.client_id = client_id + self.client_secret = client_secret + self.use_encrypted_endpoints = use_encrypted_endpoints + self.use_host_verification = use_host_verification + self.use_peer_verification = use_peer_verification + self.encrypted_credential = encrypted_credential + self.type = 'ServiceNow' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_object_dataset.py index 30564609bc8a..df9d7a90f3f8 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_object_dataset.py @@ -15,6 +15,8 @@ class ServiceNowObjectDataset(Dataset): """ServiceNow server dataset. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -23,7 +25,7 @@ class ServiceNowObjectDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object - :param linked_service_name: Linked service reference. + :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference :param parameters: Parameters for dataset. @@ -32,7 +34,7 @@ class ServiceNowObjectDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str """ @@ -41,6 +43,16 @@ class ServiceNowObjectDataset(Dataset): 'type': {'required': True}, } - def __init__(self, linked_service_name, additional_properties=None, description=None, structure=None, parameters=None, annotations=None): - super(ServiceNowObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations) + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, **kwargs): + super(ServiceNowObjectDataset, self).__init__(**kwargs) self.type = 'ServiceNowObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_object_dataset_py3.py new file mode 100644 index 000000000000..8d2abba32ec6 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_object_dataset_py3.py @@ -0,0 +1,58 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class ServiceNowObjectDataset(Dataset): + """ServiceNow server dataset. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Required. Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: + super(ServiceNowObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + self.type = 'ServiceNowObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_source.py index 0bc6100188d5..00068f5e5d32 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_source.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_source.py @@ -15,6 +15,8 @@ class ServiceNowSource(CopySource): """A copy activity ServiceNow server source. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -25,7 +27,7 @@ class ServiceNowSource(CopySource): with resultType string), pattern: ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type source_retry_wait: object - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). @@ -44,7 +46,7 @@ class ServiceNowSource(CopySource): 'query': {'key': 'query', 'type': 'object'}, } - def __init__(self, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None): - super(ServiceNowSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait) - self.query = query + def __init__(self, **kwargs): + super(ServiceNowSource, self).__init__(**kwargs) + self.query = kwargs.get('query', None) self.type = 'ServiceNowSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_source_py3.py new file mode 100644 index 000000000000..8010f2aea2fb --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_source_py3.py @@ -0,0 +1,52 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class ServiceNowSource(CopySource): + """A copy activity ServiceNow server source. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Required. Constant filled by server. + :type type: str + :param query: A query to retrieve data from source. Type: string (or + Expression with resultType string). + :type query: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'query': {'key': 'query', 'type': 'object'}, + } + + def __init__(self, *, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None, **kwargs) -> None: + super(ServiceNowSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, **kwargs) + self.query = query + self.type = 'ServiceNowSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sftp_server_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sftp_server_linked_service.py index 3e691e9378f3..31a9d5524f36 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sftp_server_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sftp_server_linked_service.py @@ -15,6 +15,8 @@ class SftpServerLinkedService(LinkedService): """A linked service for an SSH File Transfer Protocol (SFTP) server. . + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -29,10 +31,10 @@ class SftpServerLinkedService(LinkedService): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param host: The SFTP server host name. Type: string (or Expression with - resultType string). + :param host: Required. The SFTP server host name. Type: string (or + Expression with resultType string). :type host: object :param port: The TCP port number that the SFTP server uses to listen for client connections. Default value is 22. Type: integer (or Expression with @@ -101,17 +103,17 @@ class SftpServerLinkedService(LinkedService): 'host_key_fingerprint': {'key': 'typeProperties.hostKeyFingerprint', 'type': 'object'}, } - def __init__(self, host, additional_properties=None, connect_via=None, description=None, parameters=None, annotations=None, port=None, authentication_type=None, user_name=None, password=None, encrypted_credential=None, private_key_path=None, private_key_content=None, pass_phrase=None, skip_host_key_validation=None, host_key_fingerprint=None): - super(SftpServerLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations) - self.host = host - self.port = port - self.authentication_type = authentication_type - self.user_name = user_name - self.password = password - self.encrypted_credential = encrypted_credential - self.private_key_path = private_key_path - self.private_key_content = private_key_content - self.pass_phrase = pass_phrase - self.skip_host_key_validation = skip_host_key_validation - self.host_key_fingerprint = host_key_fingerprint + def __init__(self, **kwargs): + super(SftpServerLinkedService, self).__init__(**kwargs) + self.host = kwargs.get('host', None) + self.port = kwargs.get('port', None) + self.authentication_type = kwargs.get('authentication_type', None) + self.user_name = kwargs.get('user_name', None) + self.password = kwargs.get('password', None) + self.encrypted_credential = kwargs.get('encrypted_credential', None) + self.private_key_path = kwargs.get('private_key_path', None) + self.private_key_content = kwargs.get('private_key_content', None) + self.pass_phrase = kwargs.get('pass_phrase', None) + self.skip_host_key_validation = kwargs.get('skip_host_key_validation', None) + self.host_key_fingerprint = kwargs.get('host_key_fingerprint', None) self.type = 'Sftp' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sftp_server_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sftp_server_linked_service_py3.py new file mode 100644 index 000000000000..2c2fcca6f001 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sftp_server_linked_service_py3.py @@ -0,0 +1,119 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class SftpServerLinkedService(LinkedService): + """A linked service for an SSH File Transfer Protocol (SFTP) server. . + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param host: Required. The SFTP server host name. Type: string (or + Expression with resultType string). + :type host: object + :param port: The TCP port number that the SFTP server uses to listen for + client connections. Default value is 22. Type: integer (or Expression with + resultType integer), minimum: 0. + :type port: object + :param authentication_type: The authentication type to be used to connect + to the FTP server. Possible values include: 'Basic', 'SshPublicKey' + :type authentication_type: str or + ~azure.mgmt.datafactory.models.SftpAuthenticationType + :param user_name: The username used to log on to the SFTP server. Type: + string (or Expression with resultType string). + :type user_name: object + :param password: Password to logon the SFTP server for Basic + authentication. + :type password: ~azure.mgmt.datafactory.models.SecretBase + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + :param private_key_path: The SSH private key file path for SshPublicKey + authentication. Only valid for on-premises copy. For on-premises copy with + SshPublicKey authentication, either PrivateKeyPath or PrivateKeyContent + should be specified. SSH private key should be OpenSSH format. Type: + string (or Expression with resultType string). + :type private_key_path: object + :param private_key_content: Base64 encoded SSH private key content for + SshPublicKey authentication. For on-premises copy with SshPublicKey + authentication, either PrivateKeyPath or PrivateKeyContent should be + specified. SSH private key should be OpenSSH format. + :type private_key_content: ~azure.mgmt.datafactory.models.SecretBase + :param pass_phrase: The password to decrypt the SSH private key if the SSH + private key is encrypted. + :type pass_phrase: ~azure.mgmt.datafactory.models.SecretBase + :param skip_host_key_validation: If true, skip the SSH host key + validation. Default value is false. Type: boolean (or Expression with + resultType boolean). + :type skip_host_key_validation: object + :param host_key_fingerprint: The host key finger-print of the SFTP server. + When SkipHostKeyValidation is false, HostKeyFingerprint should be + specified. Type: string (or Expression with resultType string). + :type host_key_fingerprint: object + """ + + _validation = { + 'type': {'required': True}, + 'host': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'host': {'key': 'typeProperties.host', 'type': 'object'}, + 'port': {'key': 'typeProperties.port', 'type': 'object'}, + 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, + 'user_name': {'key': 'typeProperties.userName', 'type': 'object'}, + 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + 'private_key_path': {'key': 'typeProperties.privateKeyPath', 'type': 'object'}, + 'private_key_content': {'key': 'typeProperties.privateKeyContent', 'type': 'SecretBase'}, + 'pass_phrase': {'key': 'typeProperties.passPhrase', 'type': 'SecretBase'}, + 'skip_host_key_validation': {'key': 'typeProperties.skipHostKeyValidation', 'type': 'object'}, + 'host_key_fingerprint': {'key': 'typeProperties.hostKeyFingerprint', 'type': 'object'}, + } + + def __init__(self, *, host, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, port=None, authentication_type=None, user_name=None, password=None, encrypted_credential=None, private_key_path=None, private_key_content=None, pass_phrase=None, skip_host_key_validation=None, host_key_fingerprint=None, **kwargs) -> None: + super(SftpServerLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.host = host + self.port = port + self.authentication_type = authentication_type + self.user_name = user_name + self.password = password + self.encrypted_credential = encrypted_credential + self.private_key_path = private_key_path + self.private_key_content = private_key_content + self.pass_phrase = pass_phrase + self.skip_host_key_validation = skip_host_key_validation + self.host_key_fingerprint = host_key_fingerprint + self.type = 'Sftp' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_linked_service.py index c6fadecfd577..8d38e5d64a17 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_linked_service.py @@ -15,6 +15,8 @@ class ShopifyLinkedService(LinkedService): """Shopify Serivce linked service. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -29,9 +31,9 @@ class ShopifyLinkedService(LinkedService): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param host: The endpoint of the Shopify server. (i.e. + :param host: Required. The endpoint of the Shopify server. (i.e. mystore.myshopify.com) :type host: object :param access_token: The API access token that can be used to access @@ -73,12 +75,12 @@ class ShopifyLinkedService(LinkedService): 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, host, additional_properties=None, connect_via=None, description=None, parameters=None, annotations=None, access_token=None, use_encrypted_endpoints=None, use_host_verification=None, use_peer_verification=None, encrypted_credential=None): - super(ShopifyLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations) - self.host = host - self.access_token = access_token - self.use_encrypted_endpoints = use_encrypted_endpoints - self.use_host_verification = use_host_verification - self.use_peer_verification = use_peer_verification - self.encrypted_credential = encrypted_credential + def __init__(self, **kwargs): + super(ShopifyLinkedService, self).__init__(**kwargs) + self.host = kwargs.get('host', None) + self.access_token = kwargs.get('access_token', None) + self.use_encrypted_endpoints = kwargs.get('use_encrypted_endpoints', None) + self.use_host_verification = kwargs.get('use_host_verification', None) + self.use_peer_verification = kwargs.get('use_peer_verification', None) + self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'Shopify' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_linked_service_py3.py new file mode 100644 index 000000000000..4d1ac17b1ebe --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_linked_service_py3.py @@ -0,0 +1,86 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class ShopifyLinkedService(LinkedService): + """Shopify Serivce linked service. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param host: Required. The endpoint of the Shopify server. (i.e. + mystore.myshopify.com) + :type host: object + :param access_token: The API access token that can be used to access + Shopify’s data. The token won't expire if it is offline mode. + :type access_token: ~azure.mgmt.datafactory.models.SecretBase + :param use_encrypted_endpoints: Specifies whether the data source + endpoints are encrypted using HTTPS. The default value is true. + :type use_encrypted_endpoints: object + :param use_host_verification: Specifies whether to require the host name + in the server's certificate to match the host name of the server when + connecting over SSL. The default value is true. + :type use_host_verification: object + :param use_peer_verification: Specifies whether to verify the identity of + the server when connecting over SSL. The default value is true. + :type use_peer_verification: object + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'host': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'host': {'key': 'typeProperties.host', 'type': 'object'}, + 'access_token': {'key': 'typeProperties.accessToken', 'type': 'SecretBase'}, + 'use_encrypted_endpoints': {'key': 'typeProperties.useEncryptedEndpoints', 'type': 'object'}, + 'use_host_verification': {'key': 'typeProperties.useHostVerification', 'type': 'object'}, + 'use_peer_verification': {'key': 'typeProperties.usePeerVerification', 'type': 'object'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, *, host, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, access_token=None, use_encrypted_endpoints=None, use_host_verification=None, use_peer_verification=None, encrypted_credential=None, **kwargs) -> None: + super(ShopifyLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.host = host + self.access_token = access_token + self.use_encrypted_endpoints = use_encrypted_endpoints + self.use_host_verification = use_host_verification + self.use_peer_verification = use_peer_verification + self.encrypted_credential = encrypted_credential + self.type = 'Shopify' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_object_dataset.py index cd68975841da..2c64c3d43d8f 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_object_dataset.py @@ -15,6 +15,8 @@ class ShopifyObjectDataset(Dataset): """Shopify Serivce dataset. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -23,7 +25,7 @@ class ShopifyObjectDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object - :param linked_service_name: Linked service reference. + :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference :param parameters: Parameters for dataset. @@ -32,7 +34,7 @@ class ShopifyObjectDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str """ @@ -41,6 +43,16 @@ class ShopifyObjectDataset(Dataset): 'type': {'required': True}, } - def __init__(self, linked_service_name, additional_properties=None, description=None, structure=None, parameters=None, annotations=None): - super(ShopifyObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations) + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, **kwargs): + super(ShopifyObjectDataset, self).__init__(**kwargs) self.type = 'ShopifyObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_object_dataset_py3.py new file mode 100644 index 000000000000..b3ccec372e44 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_object_dataset_py3.py @@ -0,0 +1,58 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class ShopifyObjectDataset(Dataset): + """Shopify Serivce dataset. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Required. Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: + super(ShopifyObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + self.type = 'ShopifyObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_source.py index a156f80e6fed..cc5202bdc6bb 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_source.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_source.py @@ -15,6 +15,8 @@ class ShopifySource(CopySource): """A copy activity Shopify Serivce source. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -25,7 +27,7 @@ class ShopifySource(CopySource): with resultType string), pattern: ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type source_retry_wait: object - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). @@ -44,7 +46,7 @@ class ShopifySource(CopySource): 'query': {'key': 'query', 'type': 'object'}, } - def __init__(self, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None): - super(ShopifySource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait) - self.query = query + def __init__(self, **kwargs): + super(ShopifySource, self).__init__(**kwargs) + self.query = kwargs.get('query', None) self.type = 'ShopifySource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_source_py3.py new file mode 100644 index 000000000000..1170ae521d47 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_source_py3.py @@ -0,0 +1,52 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class ShopifySource(CopySource): + """A copy activity Shopify Serivce source. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Required. Constant filled by server. + :type type: str + :param query: A query to retrieve data from source. Type: string (or + Expression with resultType string). + :type query: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'query': {'key': 'query', 'type': 'object'}, + } + + def __init__(self, *, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None, **kwargs) -> None: + super(ShopifySource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, **kwargs) + self.query = query + self.type = 'ShopifySource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/spark_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/spark_linked_service.py index ec6464d81425..006311c492bb 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/spark_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/spark_linked_service.py @@ -15,6 +15,8 @@ class SparkLinkedService(LinkedService): """Spark Server linked service. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -29,12 +31,12 @@ class SparkLinkedService(LinkedService): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param host: IP address or host name of the Spark server + :param host: Required. IP address or host name of the Spark server :type host: object - :param port: The TCP port that the Spark server uses to listen for client - connections. + :param port: Required. The TCP port that the Spark server uses to listen + for client connections. :type port: object :param server_type: The type of Spark server. Possible values include: 'SharkServer', 'SharkServer2', 'SparkThriftServer' @@ -43,8 +45,8 @@ class SparkLinkedService(LinkedService): Thrift layer. Possible values include: 'Binary', 'SASL', 'HTTP ' :type thrift_transport_protocol: str or ~azure.mgmt.datafactory.models.SparkThriftTransportProtocol - :param authentication_type: The authentication method used to access the - Spark server. Possible values include: 'Anonymous', 'Username', + :param authentication_type: Required. The authentication method used to + access the Spark server. Possible values include: 'Anonymous', 'Username', 'UsernameAndPassword', 'WindowsAzureHDInsightService' :type authentication_type: str or ~azure.mgmt.datafactory.models.SparkAuthenticationType @@ -110,20 +112,20 @@ class SparkLinkedService(LinkedService): 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, host, port, authentication_type, additional_properties=None, connect_via=None, description=None, parameters=None, annotations=None, server_type=None, thrift_transport_protocol=None, username=None, password=None, http_path=None, enable_ssl=None, trusted_cert_path=None, use_system_trust_store=None, allow_host_name_cn_mismatch=None, allow_self_signed_server_cert=None, encrypted_credential=None): - super(SparkLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations) - self.host = host - self.port = port - self.server_type = server_type - self.thrift_transport_protocol = thrift_transport_protocol - self.authentication_type = authentication_type - self.username = username - self.password = password - self.http_path = http_path - self.enable_ssl = enable_ssl - self.trusted_cert_path = trusted_cert_path - self.use_system_trust_store = use_system_trust_store - self.allow_host_name_cn_mismatch = allow_host_name_cn_mismatch - self.allow_self_signed_server_cert = allow_self_signed_server_cert - self.encrypted_credential = encrypted_credential + def __init__(self, **kwargs): + super(SparkLinkedService, self).__init__(**kwargs) + self.host = kwargs.get('host', None) + self.port = kwargs.get('port', None) + self.server_type = kwargs.get('server_type', None) + self.thrift_transport_protocol = kwargs.get('thrift_transport_protocol', None) + self.authentication_type = kwargs.get('authentication_type', None) + self.username = kwargs.get('username', None) + self.password = kwargs.get('password', None) + self.http_path = kwargs.get('http_path', None) + self.enable_ssl = kwargs.get('enable_ssl', None) + self.trusted_cert_path = kwargs.get('trusted_cert_path', None) + self.use_system_trust_store = kwargs.get('use_system_trust_store', None) + self.allow_host_name_cn_mismatch = kwargs.get('allow_host_name_cn_mismatch', None) + self.allow_self_signed_server_cert = kwargs.get('allow_self_signed_server_cert', None) + self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'Spark' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/spark_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/spark_linked_service_py3.py new file mode 100644 index 000000000000..832b8cb0c7da --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/spark_linked_service_py3.py @@ -0,0 +1,131 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class SparkLinkedService(LinkedService): + """Spark Server linked service. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param host: Required. IP address or host name of the Spark server + :type host: object + :param port: Required. The TCP port that the Spark server uses to listen + for client connections. + :type port: object + :param server_type: The type of Spark server. Possible values include: + 'SharkServer', 'SharkServer2', 'SparkThriftServer' + :type server_type: str or ~azure.mgmt.datafactory.models.SparkServerType + :param thrift_transport_protocol: The transport protocol to use in the + Thrift layer. Possible values include: 'Binary', 'SASL', 'HTTP ' + :type thrift_transport_protocol: str or + ~azure.mgmt.datafactory.models.SparkThriftTransportProtocol + :param authentication_type: Required. The authentication method used to + access the Spark server. Possible values include: 'Anonymous', 'Username', + 'UsernameAndPassword', 'WindowsAzureHDInsightService' + :type authentication_type: str or + ~azure.mgmt.datafactory.models.SparkAuthenticationType + :param username: The user name that you use to access Spark Server. + :type username: object + :param password: The password corresponding to the user name that you + provided in the Username field + :type password: ~azure.mgmt.datafactory.models.SecretBase + :param http_path: The partial URL corresponding to the Spark server. + :type http_path: object + :param enable_ssl: Specifies whether the connections to the server are + encrypted using SSL. The default value is false. + :type enable_ssl: object + :param trusted_cert_path: The full path of the .pem file containing + trusted CA certificates for verifying the server when connecting over SSL. + This property can only be set when using SSL on self-hosted IR. The + default value is the cacerts.pem file installed with the IR. + :type trusted_cert_path: object + :param use_system_trust_store: Specifies whether to use a CA certificate + from the system trust store or from a specified PEM file. The default + value is false. + :type use_system_trust_store: object + :param allow_host_name_cn_mismatch: Specifies whether to require a + CA-issued SSL certificate name to match the host name of the server when + connecting over SSL. The default value is false. + :type allow_host_name_cn_mismatch: object + :param allow_self_signed_server_cert: Specifies whether to allow + self-signed certificates from the server. The default value is false. + :type allow_self_signed_server_cert: object + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'host': {'required': True}, + 'port': {'required': True}, + 'authentication_type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'host': {'key': 'typeProperties.host', 'type': 'object'}, + 'port': {'key': 'typeProperties.port', 'type': 'object'}, + 'server_type': {'key': 'typeProperties.serverType', 'type': 'str'}, + 'thrift_transport_protocol': {'key': 'typeProperties.thriftTransportProtocol', 'type': 'str'}, + 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, + 'username': {'key': 'typeProperties.username', 'type': 'object'}, + 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, + 'http_path': {'key': 'typeProperties.httpPath', 'type': 'object'}, + 'enable_ssl': {'key': 'typeProperties.enableSsl', 'type': 'object'}, + 'trusted_cert_path': {'key': 'typeProperties.trustedCertPath', 'type': 'object'}, + 'use_system_trust_store': {'key': 'typeProperties.useSystemTrustStore', 'type': 'object'}, + 'allow_host_name_cn_mismatch': {'key': 'typeProperties.allowHostNameCNMismatch', 'type': 'object'}, + 'allow_self_signed_server_cert': {'key': 'typeProperties.allowSelfSignedServerCert', 'type': 'object'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, *, host, port, authentication_type, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, server_type=None, thrift_transport_protocol=None, username=None, password=None, http_path=None, enable_ssl=None, trusted_cert_path=None, use_system_trust_store=None, allow_host_name_cn_mismatch=None, allow_self_signed_server_cert=None, encrypted_credential=None, **kwargs) -> None: + super(SparkLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.host = host + self.port = port + self.server_type = server_type + self.thrift_transport_protocol = thrift_transport_protocol + self.authentication_type = authentication_type + self.username = username + self.password = password + self.http_path = http_path + self.enable_ssl = enable_ssl + self.trusted_cert_path = trusted_cert_path + self.use_system_trust_store = use_system_trust_store + self.allow_host_name_cn_mismatch = allow_host_name_cn_mismatch + self.allow_self_signed_server_cert = allow_self_signed_server_cert + self.encrypted_credential = encrypted_credential + self.type = 'Spark' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/spark_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/spark_object_dataset.py index 0f998f03375c..84a7776e3091 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/spark_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/spark_object_dataset.py @@ -15,6 +15,8 @@ class SparkObjectDataset(Dataset): """Spark Server dataset. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -23,7 +25,7 @@ class SparkObjectDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object - :param linked_service_name: Linked service reference. + :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference :param parameters: Parameters for dataset. @@ -32,7 +34,7 @@ class SparkObjectDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str """ @@ -41,6 +43,16 @@ class SparkObjectDataset(Dataset): 'type': {'required': True}, } - def __init__(self, linked_service_name, additional_properties=None, description=None, structure=None, parameters=None, annotations=None): - super(SparkObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations) + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, **kwargs): + super(SparkObjectDataset, self).__init__(**kwargs) self.type = 'SparkObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/spark_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/spark_object_dataset_py3.py new file mode 100644 index 000000000000..cee1552ac9ee --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/spark_object_dataset_py3.py @@ -0,0 +1,58 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class SparkObjectDataset(Dataset): + """Spark Server dataset. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Required. Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: + super(SparkObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + self.type = 'SparkObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/spark_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/spark_source.py index 2a2c0b07b4cc..643a71610930 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/spark_source.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/spark_source.py @@ -15,6 +15,8 @@ class SparkSource(CopySource): """A copy activity Spark Server source. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -25,7 +27,7 @@ class SparkSource(CopySource): with resultType string), pattern: ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type source_retry_wait: object - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). @@ -44,7 +46,7 @@ class SparkSource(CopySource): 'query': {'key': 'query', 'type': 'object'}, } - def __init__(self, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None): - super(SparkSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait) - self.query = query + def __init__(self, **kwargs): + super(SparkSource, self).__init__(**kwargs) + self.query = kwargs.get('query', None) self.type = 'SparkSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/spark_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/spark_source_py3.py new file mode 100644 index 000000000000..8fd55d21d579 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/spark_source_py3.py @@ -0,0 +1,52 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class SparkSource(CopySource): + """A copy activity Spark Server source. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Required. Constant filled by server. + :type type: str + :param query: A query to retrieve data from source. Type: string (or + Expression with resultType string). + :type query: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'query': {'key': 'query', 'type': 'object'}, + } + + def __init__(self, *, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None, **kwargs) -> None: + super(SparkSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, **kwargs) + self.query = query + self.type = 'SparkSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_dw_sink.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_dw_sink.py index 1441b3fdc6b0..ac12b6e55e59 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_dw_sink.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_dw_sink.py @@ -15,6 +15,8 @@ class SqlDWSink(CopySink): """A copy activity SQL Data Warehouse sink. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -32,7 +34,7 @@ class SqlDWSink(CopySink): resultType string), pattern: ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type sink_retry_wait: object - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param pre_copy_script: SQL pre-copy script. Type: string (or Expression with resultType string). @@ -62,9 +64,9 @@ class SqlDWSink(CopySink): 'poly_base_settings': {'key': 'polyBaseSettings', 'type': 'PolybaseSettings'}, } - def __init__(self, additional_properties=None, write_batch_size=None, write_batch_timeout=None, sink_retry_count=None, sink_retry_wait=None, pre_copy_script=None, allow_poly_base=None, poly_base_settings=None): - super(SqlDWSink, self).__init__(additional_properties=additional_properties, write_batch_size=write_batch_size, write_batch_timeout=write_batch_timeout, sink_retry_count=sink_retry_count, sink_retry_wait=sink_retry_wait) - self.pre_copy_script = pre_copy_script - self.allow_poly_base = allow_poly_base - self.poly_base_settings = poly_base_settings + def __init__(self, **kwargs): + super(SqlDWSink, self).__init__(**kwargs) + self.pre_copy_script = kwargs.get('pre_copy_script', None) + self.allow_poly_base = kwargs.get('allow_poly_base', None) + self.poly_base_settings = kwargs.get('poly_base_settings', None) self.type = 'SqlDWSink' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_dw_sink_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_dw_sink_py3.py new file mode 100644 index 000000000000..5380b32f61d7 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_dw_sink_py3.py @@ -0,0 +1,72 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_sink import CopySink + + +class SqlDWSink(CopySink): + """A copy activity SQL Data Warehouse sink. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param write_batch_size: Write batch size. Type: integer (or Expression + with resultType integer), minimum: 0. + :type write_batch_size: object + :param write_batch_timeout: Write batch timeout. Type: string (or + Expression with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type write_batch_timeout: object + :param sink_retry_count: Sink retry count. Type: integer (or Expression + with resultType integer). + :type sink_retry_count: object + :param sink_retry_wait: Sink retry wait. Type: string (or Expression with + resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type sink_retry_wait: object + :param type: Required. Constant filled by server. + :type type: str + :param pre_copy_script: SQL pre-copy script. Type: string (or Expression + with resultType string). + :type pre_copy_script: object + :param allow_poly_base: Indicates to use PolyBase to copy data into SQL + Data Warehouse when applicable. Type: boolean (or Expression with + resultType boolean). + :type allow_poly_base: object + :param poly_base_settings: Specifies PolyBase-related settings when + allowPolyBase is true. + :type poly_base_settings: ~azure.mgmt.datafactory.models.PolybaseSettings + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, + 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, + 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, + 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'pre_copy_script': {'key': 'preCopyScript', 'type': 'object'}, + 'allow_poly_base': {'key': 'allowPolyBase', 'type': 'object'}, + 'poly_base_settings': {'key': 'polyBaseSettings', 'type': 'PolybaseSettings'}, + } + + def __init__(self, *, additional_properties=None, write_batch_size=None, write_batch_timeout=None, sink_retry_count=None, sink_retry_wait=None, pre_copy_script=None, allow_poly_base=None, poly_base_settings=None, **kwargs) -> None: + super(SqlDWSink, self).__init__(additional_properties=additional_properties, write_batch_size=write_batch_size, write_batch_timeout=write_batch_timeout, sink_retry_count=sink_retry_count, sink_retry_wait=sink_retry_wait, **kwargs) + self.pre_copy_script = pre_copy_script + self.allow_poly_base = allow_poly_base + self.poly_base_settings = poly_base_settings + self.type = 'SqlDWSink' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_dw_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_dw_source.py index c26179b18c2d..aa3f88a75938 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_dw_source.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_dw_source.py @@ -15,6 +15,8 @@ class SqlDWSource(CopySource): """A copy activity SQL Data Warehouse source. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -25,7 +27,7 @@ class SqlDWSource(CopySource): with resultType string), pattern: ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type source_retry_wait: object - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param sql_reader_query: SQL Data Warehouse reader query. Type: string (or Expression with resultType string). @@ -55,9 +57,9 @@ class SqlDWSource(CopySource): 'stored_procedure_parameters': {'key': 'storedProcedureParameters', 'type': 'object'}, } - def __init__(self, additional_properties=None, source_retry_count=None, source_retry_wait=None, sql_reader_query=None, sql_reader_stored_procedure_name=None, stored_procedure_parameters=None): - super(SqlDWSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait) - self.sql_reader_query = sql_reader_query - self.sql_reader_stored_procedure_name = sql_reader_stored_procedure_name - self.stored_procedure_parameters = stored_procedure_parameters + def __init__(self, **kwargs): + super(SqlDWSource, self).__init__(**kwargs) + self.sql_reader_query = kwargs.get('sql_reader_query', None) + self.sql_reader_stored_procedure_name = kwargs.get('sql_reader_stored_procedure_name', None) + self.stored_procedure_parameters = kwargs.get('stored_procedure_parameters', None) self.type = 'SqlDWSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_dw_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_dw_source_py3.py new file mode 100644 index 000000000000..3fb991a7bc64 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_dw_source_py3.py @@ -0,0 +1,65 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class SqlDWSource(CopySource): + """A copy activity SQL Data Warehouse source. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Required. Constant filled by server. + :type type: str + :param sql_reader_query: SQL Data Warehouse reader query. Type: string (or + Expression with resultType string). + :type sql_reader_query: object + :param sql_reader_stored_procedure_name: Name of the stored procedure for + a SQL Data Warehouse source. This cannot be used at the same time as + SqlReaderQuery. Type: string (or Expression with resultType string). + :type sql_reader_stored_procedure_name: object + :param stored_procedure_parameters: Value and type setting for stored + procedure parameters. Example: "{Parameter1: {value: "1", type: "int"}}". + Type: object (or Expression with resultType object), itemType: + StoredProcedureParameter. + :type stored_procedure_parameters: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'sql_reader_query': {'key': 'sqlReaderQuery', 'type': 'object'}, + 'sql_reader_stored_procedure_name': {'key': 'sqlReaderStoredProcedureName', 'type': 'object'}, + 'stored_procedure_parameters': {'key': 'storedProcedureParameters', 'type': 'object'}, + } + + def __init__(self, *, additional_properties=None, source_retry_count=None, source_retry_wait=None, sql_reader_query=None, sql_reader_stored_procedure_name=None, stored_procedure_parameters=None, **kwargs) -> None: + super(SqlDWSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, **kwargs) + self.sql_reader_query = sql_reader_query + self.sql_reader_stored_procedure_name = sql_reader_stored_procedure_name + self.stored_procedure_parameters = stored_procedure_parameters + self.type = 'SqlDWSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_linked_service.py index 2356a7cce5d3..5e6a25007f3f 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_linked_service.py @@ -15,6 +15,8 @@ class SqlServerLinkedService(LinkedService): """SQL Server linked service. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -29,9 +31,9 @@ class SqlServerLinkedService(LinkedService): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param connection_string: The connection string. + :param connection_string: Required. The connection string. :type connection_string: ~azure.mgmt.datafactory.models.SecretBase :param user_name: The on-premises Windows authentication user name. Type: string (or Expression with resultType string). @@ -62,10 +64,10 @@ class SqlServerLinkedService(LinkedService): 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, connection_string, additional_properties=None, connect_via=None, description=None, parameters=None, annotations=None, user_name=None, password=None, encrypted_credential=None): - super(SqlServerLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations) - self.connection_string = connection_string - self.user_name = user_name - self.password = password - self.encrypted_credential = encrypted_credential + def __init__(self, **kwargs): + super(SqlServerLinkedService, self).__init__(**kwargs) + self.connection_string = kwargs.get('connection_string', None) + self.user_name = kwargs.get('user_name', None) + self.password = kwargs.get('password', None) + self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'SqlServer' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_linked_service_py3.py new file mode 100644 index 000000000000..de0bb00b1cbd --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_linked_service_py3.py @@ -0,0 +1,73 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class SqlServerLinkedService(LinkedService): + """SQL Server linked service. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param connection_string: Required. The connection string. + :type connection_string: ~azure.mgmt.datafactory.models.SecretBase + :param user_name: The on-premises Windows authentication user name. Type: + string (or Expression with resultType string). + :type user_name: object + :param password: The on-premises Windows authentication password. + :type password: ~azure.mgmt.datafactory.models.SecretBase + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'connection_string': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'SecretBase'}, + 'user_name': {'key': 'typeProperties.userName', 'type': 'object'}, + 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, *, connection_string, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, user_name=None, password=None, encrypted_credential=None, **kwargs) -> None: + super(SqlServerLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.connection_string = connection_string + self.user_name = user_name + self.password = password + self.encrypted_credential = encrypted_credential + self.type = 'SqlServer' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_stored_procedure_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_stored_procedure_activity.py index c8d67bb25aff..4d39b9d806f1 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_stored_procedure_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_stored_procedure_activity.py @@ -15,24 +15,26 @@ class SqlServerStoredProcedureActivity(ExecutionActivity): """SQL stored procedure activity type. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] - :param name: Activity name. + :param name: Required. Activity name. :type name: str :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference :param policy: Activity policy. :type policy: ~azure.mgmt.datafactory.models.ActivityPolicy - :param stored_procedure_name: Stored procedure name. Type: string (or - Expression with resultType string). + :param stored_procedure_name: Required. Stored procedure name. Type: + string (or Expression with resultType string). :type stored_procedure_name: object :param stored_procedure_parameters: Value and type setting for stored procedure parameters. Example: "{Parameter1: {value: "1", type: "int"}}". @@ -58,8 +60,8 @@ class SqlServerStoredProcedureActivity(ExecutionActivity): 'stored_procedure_parameters': {'key': 'typeProperties.storedProcedureParameters', 'type': '{StoredProcedureParameter}'}, } - def __init__(self, name, stored_procedure_name, additional_properties=None, description=None, depends_on=None, linked_service_name=None, policy=None, stored_procedure_parameters=None): - super(SqlServerStoredProcedureActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, linked_service_name=linked_service_name, policy=policy) - self.stored_procedure_name = stored_procedure_name - self.stored_procedure_parameters = stored_procedure_parameters + def __init__(self, **kwargs): + super(SqlServerStoredProcedureActivity, self).__init__(**kwargs) + self.stored_procedure_name = kwargs.get('stored_procedure_name', None) + self.stored_procedure_parameters = kwargs.get('stored_procedure_parameters', None) self.type = 'SqlServerStoredProcedure' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_stored_procedure_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_stored_procedure_activity_py3.py new file mode 100644 index 000000000000..a03d1500425a --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_stored_procedure_activity_py3.py @@ -0,0 +1,67 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .execution_activity import ExecutionActivity + + +class SqlServerStoredProcedureActivity(ExecutionActivity): + """SQL stored procedure activity type. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param name: Required. Activity name. + :type name: str + :param description: Activity description. + :type description: str + :param depends_on: Activity depends on condition. + :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param type: Required. Constant filled by server. + :type type: str + :param linked_service_name: Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param policy: Activity policy. + :type policy: ~azure.mgmt.datafactory.models.ActivityPolicy + :param stored_procedure_name: Required. Stored procedure name. Type: + string (or Expression with resultType string). + :type stored_procedure_name: object + :param stored_procedure_parameters: Value and type setting for stored + procedure parameters. Example: "{Parameter1: {value: "1", type: "int"}}". + :type stored_procedure_parameters: dict[str, + ~azure.mgmt.datafactory.models.StoredProcedureParameter] + """ + + _validation = { + 'name': {'required': True}, + 'type': {'required': True}, + 'stored_procedure_name': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'name': {'key': 'name', 'type': 'str'}, + 'description': {'key': 'description', 'type': 'str'}, + 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, + 'stored_procedure_name': {'key': 'typeProperties.storedProcedureName', 'type': 'object'}, + 'stored_procedure_parameters': {'key': 'typeProperties.storedProcedureParameters', 'type': '{StoredProcedureParameter}'}, + } + + def __init__(self, *, name: str, stored_procedure_name, additional_properties=None, description: str=None, depends_on=None, linked_service_name=None, policy=None, stored_procedure_parameters=None, **kwargs) -> None: + super(SqlServerStoredProcedureActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, linked_service_name=linked_service_name, policy=policy, **kwargs) + self.stored_procedure_name = stored_procedure_name + self.stored_procedure_parameters = stored_procedure_parameters + self.type = 'SqlServerStoredProcedure' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_table_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_table_dataset.py index 4fcde4bc8df9..3e058a5d9506 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_table_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_table_dataset.py @@ -15,6 +15,8 @@ class SqlServerTableDataset(Dataset): """The on-premises SQL Server dataset. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -23,7 +25,7 @@ class SqlServerTableDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object - :param linked_service_name: Linked service reference. + :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference :param parameters: Parameters for dataset. @@ -32,10 +34,10 @@ class SqlServerTableDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param table_name: The table name of the SQL Server dataset. Type: string - (or Expression with resultType string). + :param table_name: Required. The table name of the SQL Server dataset. + Type: string (or Expression with resultType string). :type table_name: object """ @@ -56,7 +58,7 @@ class SqlServerTableDataset(Dataset): 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, linked_service_name, table_name, additional_properties=None, description=None, structure=None, parameters=None, annotations=None): - super(SqlServerTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations) - self.table_name = table_name + def __init__(self, **kwargs): + super(SqlServerTableDataset, self).__init__(**kwargs) + self.table_name = kwargs.get('table_name', None) self.type = 'SqlServerTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_table_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_table_dataset_py3.py new file mode 100644 index 000000000000..fe75db9ab096 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_table_dataset_py3.py @@ -0,0 +1,64 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class SqlServerTableDataset(Dataset): + """The on-premises SQL Server dataset. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Required. Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param table_name: Required. The table name of the SQL Server dataset. + Type: string (or Expression with resultType string). + :type table_name: object + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + 'table_name': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, + } + + def __init__(self, *, linked_service_name, table_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: + super(SqlServerTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + self.table_name = table_name + self.type = 'SqlServerTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_sink.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_sink.py index 88837785773d..77692817100d 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_sink.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_sink.py @@ -15,6 +15,8 @@ class SqlSink(CopySink): """A copy activity SQL sink. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -32,7 +34,7 @@ class SqlSink(CopySink): resultType string), pattern: ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type sink_retry_wait: object - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param sql_writer_stored_procedure_name: SQL writer stored procedure name. Type: string (or Expression with resultType string). @@ -65,10 +67,10 @@ class SqlSink(CopySink): 'stored_procedure_parameters': {'key': 'storedProcedureParameters', 'type': '{StoredProcedureParameter}'}, } - def __init__(self, additional_properties=None, write_batch_size=None, write_batch_timeout=None, sink_retry_count=None, sink_retry_wait=None, sql_writer_stored_procedure_name=None, sql_writer_table_type=None, pre_copy_script=None, stored_procedure_parameters=None): - super(SqlSink, self).__init__(additional_properties=additional_properties, write_batch_size=write_batch_size, write_batch_timeout=write_batch_timeout, sink_retry_count=sink_retry_count, sink_retry_wait=sink_retry_wait) - self.sql_writer_stored_procedure_name = sql_writer_stored_procedure_name - self.sql_writer_table_type = sql_writer_table_type - self.pre_copy_script = pre_copy_script - self.stored_procedure_parameters = stored_procedure_parameters + def __init__(self, **kwargs): + super(SqlSink, self).__init__(**kwargs) + self.sql_writer_stored_procedure_name = kwargs.get('sql_writer_stored_procedure_name', None) + self.sql_writer_table_type = kwargs.get('sql_writer_table_type', None) + self.pre_copy_script = kwargs.get('pre_copy_script', None) + self.stored_procedure_parameters = kwargs.get('stored_procedure_parameters', None) self.type = 'SqlSink' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_sink_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_sink_py3.py new file mode 100644 index 000000000000..e12a17c18ece --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_sink_py3.py @@ -0,0 +1,76 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_sink import CopySink + + +class SqlSink(CopySink): + """A copy activity SQL sink. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param write_batch_size: Write batch size. Type: integer (or Expression + with resultType integer), minimum: 0. + :type write_batch_size: object + :param write_batch_timeout: Write batch timeout. Type: string (or + Expression with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type write_batch_timeout: object + :param sink_retry_count: Sink retry count. Type: integer (or Expression + with resultType integer). + :type sink_retry_count: object + :param sink_retry_wait: Sink retry wait. Type: string (or Expression with + resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type sink_retry_wait: object + :param type: Required. Constant filled by server. + :type type: str + :param sql_writer_stored_procedure_name: SQL writer stored procedure name. + Type: string (or Expression with resultType string). + :type sql_writer_stored_procedure_name: object + :param sql_writer_table_type: SQL writer table type. Type: string (or + Expression with resultType string). + :type sql_writer_table_type: object + :param pre_copy_script: SQL pre-copy script. Type: string (or Expression + with resultType string). + :type pre_copy_script: object + :param stored_procedure_parameters: SQL stored procedure parameters. + :type stored_procedure_parameters: dict[str, + ~azure.mgmt.datafactory.models.StoredProcedureParameter] + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, + 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, + 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, + 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'sql_writer_stored_procedure_name': {'key': 'sqlWriterStoredProcedureName', 'type': 'object'}, + 'sql_writer_table_type': {'key': 'sqlWriterTableType', 'type': 'object'}, + 'pre_copy_script': {'key': 'preCopyScript', 'type': 'object'}, + 'stored_procedure_parameters': {'key': 'storedProcedureParameters', 'type': '{StoredProcedureParameter}'}, + } + + def __init__(self, *, additional_properties=None, write_batch_size=None, write_batch_timeout=None, sink_retry_count=None, sink_retry_wait=None, sql_writer_stored_procedure_name=None, sql_writer_table_type=None, pre_copy_script=None, stored_procedure_parameters=None, **kwargs) -> None: + super(SqlSink, self).__init__(additional_properties=additional_properties, write_batch_size=write_batch_size, write_batch_timeout=write_batch_timeout, sink_retry_count=sink_retry_count, sink_retry_wait=sink_retry_wait, **kwargs) + self.sql_writer_stored_procedure_name = sql_writer_stored_procedure_name + self.sql_writer_table_type = sql_writer_table_type + self.pre_copy_script = pre_copy_script + self.stored_procedure_parameters = stored_procedure_parameters + self.type = 'SqlSink' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_source.py index dee6953a64b3..3f374b19f072 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_source.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_source.py @@ -15,6 +15,8 @@ class SqlSource(CopySource): """A copy activity SQL source. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -25,7 +27,7 @@ class SqlSource(CopySource): with resultType string), pattern: ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type source_retry_wait: object - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param sql_reader_query: SQL reader query. Type: string (or Expression with resultType string). @@ -54,9 +56,9 @@ class SqlSource(CopySource): 'stored_procedure_parameters': {'key': 'storedProcedureParameters', 'type': '{StoredProcedureParameter}'}, } - def __init__(self, additional_properties=None, source_retry_count=None, source_retry_wait=None, sql_reader_query=None, sql_reader_stored_procedure_name=None, stored_procedure_parameters=None): - super(SqlSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait) - self.sql_reader_query = sql_reader_query - self.sql_reader_stored_procedure_name = sql_reader_stored_procedure_name - self.stored_procedure_parameters = stored_procedure_parameters + def __init__(self, **kwargs): + super(SqlSource, self).__init__(**kwargs) + self.sql_reader_query = kwargs.get('sql_reader_query', None) + self.sql_reader_stored_procedure_name = kwargs.get('sql_reader_stored_procedure_name', None) + self.stored_procedure_parameters = kwargs.get('stored_procedure_parameters', None) self.type = 'SqlSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_source_py3.py new file mode 100644 index 000000000000..34be7054e4e8 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_source_py3.py @@ -0,0 +1,64 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class SqlSource(CopySource): + """A copy activity SQL source. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Required. Constant filled by server. + :type type: str + :param sql_reader_query: SQL reader query. Type: string (or Expression + with resultType string). + :type sql_reader_query: object + :param sql_reader_stored_procedure_name: Name of the stored procedure for + a SQL Database source. This cannot be used at the same time as + SqlReaderQuery. Type: string (or Expression with resultType string). + :type sql_reader_stored_procedure_name: object + :param stored_procedure_parameters: Value and type setting for stored + procedure parameters. Example: "{Parameter1: {value: "1", type: "int"}}". + :type stored_procedure_parameters: dict[str, + ~azure.mgmt.datafactory.models.StoredProcedureParameter] + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'sql_reader_query': {'key': 'sqlReaderQuery', 'type': 'object'}, + 'sql_reader_stored_procedure_name': {'key': 'sqlReaderStoredProcedureName', 'type': 'object'}, + 'stored_procedure_parameters': {'key': 'storedProcedureParameters', 'type': '{StoredProcedureParameter}'}, + } + + def __init__(self, *, additional_properties=None, source_retry_count=None, source_retry_wait=None, sql_reader_query=None, sql_reader_stored_procedure_name=None, stored_procedure_parameters=None, **kwargs) -> None: + super(SqlSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, **kwargs) + self.sql_reader_query = sql_reader_query + self.sql_reader_stored_procedure_name = sql_reader_stored_procedure_name + self.stored_procedure_parameters = stored_procedure_parameters + self.type = 'SqlSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_linked_service.py index bda1682c7018..2b175ead12df 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_linked_service.py @@ -15,6 +15,8 @@ class SquareLinkedService(LinkedService): """Square Serivce linked service. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -29,17 +31,19 @@ class SquareLinkedService(LinkedService): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param host: The URL of the Square instance. (i.e. mystore.mysquare.com) + :param host: Required. The URL of the Square instance. (i.e. + mystore.mysquare.com) :type host: object - :param client_id: The client ID associated with your Square application. + :param client_id: Required. The client ID associated with your Square + application. :type client_id: object :param client_secret: The client secret associated with your Square application. :type client_secret: ~azure.mgmt.datafactory.models.SecretBase - :param redirect_uri: The redirect URL assigned in the Square application - dashboard. (i.e. http://localhost:2500) + :param redirect_uri: Required. The redirect URL assigned in the Square + application dashboard. (i.e. http://localhost:2500) :type redirect_uri: object :param use_encrypted_endpoints: Specifies whether the data source endpoints are encrypted using HTTPS. The default value is true. @@ -81,14 +85,14 @@ class SquareLinkedService(LinkedService): 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, host, client_id, redirect_uri, additional_properties=None, connect_via=None, description=None, parameters=None, annotations=None, client_secret=None, use_encrypted_endpoints=None, use_host_verification=None, use_peer_verification=None, encrypted_credential=None): - super(SquareLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations) - self.host = host - self.client_id = client_id - self.client_secret = client_secret - self.redirect_uri = redirect_uri - self.use_encrypted_endpoints = use_encrypted_endpoints - self.use_host_verification = use_host_verification - self.use_peer_verification = use_peer_verification - self.encrypted_credential = encrypted_credential + def __init__(self, **kwargs): + super(SquareLinkedService, self).__init__(**kwargs) + self.host = kwargs.get('host', None) + self.client_id = kwargs.get('client_id', None) + self.client_secret = kwargs.get('client_secret', None) + self.redirect_uri = kwargs.get('redirect_uri', None) + self.use_encrypted_endpoints = kwargs.get('use_encrypted_endpoints', None) + self.use_host_verification = kwargs.get('use_host_verification', None) + self.use_peer_verification = kwargs.get('use_peer_verification', None) + self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'Square' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_linked_service_py3.py new file mode 100644 index 000000000000..08cb36e5a54f --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_linked_service_py3.py @@ -0,0 +1,98 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class SquareLinkedService(LinkedService): + """Square Serivce linked service. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param host: Required. The URL of the Square instance. (i.e. + mystore.mysquare.com) + :type host: object + :param client_id: Required. The client ID associated with your Square + application. + :type client_id: object + :param client_secret: The client secret associated with your Square + application. + :type client_secret: ~azure.mgmt.datafactory.models.SecretBase + :param redirect_uri: Required. The redirect URL assigned in the Square + application dashboard. (i.e. http://localhost:2500) + :type redirect_uri: object + :param use_encrypted_endpoints: Specifies whether the data source + endpoints are encrypted using HTTPS. The default value is true. + :type use_encrypted_endpoints: object + :param use_host_verification: Specifies whether to require the host name + in the server's certificate to match the host name of the server when + connecting over SSL. The default value is true. + :type use_host_verification: object + :param use_peer_verification: Specifies whether to verify the identity of + the server when connecting over SSL. The default value is true. + :type use_peer_verification: object + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'host': {'required': True}, + 'client_id': {'required': True}, + 'redirect_uri': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'host': {'key': 'typeProperties.host', 'type': 'object'}, + 'client_id': {'key': 'typeProperties.clientId', 'type': 'object'}, + 'client_secret': {'key': 'typeProperties.clientSecret', 'type': 'SecretBase'}, + 'redirect_uri': {'key': 'typeProperties.redirectUri', 'type': 'object'}, + 'use_encrypted_endpoints': {'key': 'typeProperties.useEncryptedEndpoints', 'type': 'object'}, + 'use_host_verification': {'key': 'typeProperties.useHostVerification', 'type': 'object'}, + 'use_peer_verification': {'key': 'typeProperties.usePeerVerification', 'type': 'object'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, *, host, client_id, redirect_uri, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, client_secret=None, use_encrypted_endpoints=None, use_host_verification=None, use_peer_verification=None, encrypted_credential=None, **kwargs) -> None: + super(SquareLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.host = host + self.client_id = client_id + self.client_secret = client_secret + self.redirect_uri = redirect_uri + self.use_encrypted_endpoints = use_encrypted_endpoints + self.use_host_verification = use_host_verification + self.use_peer_verification = use_peer_verification + self.encrypted_credential = encrypted_credential + self.type = 'Square' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_object_dataset.py index e6f95997d532..789fec1809d2 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_object_dataset.py @@ -15,6 +15,8 @@ class SquareObjectDataset(Dataset): """Square Serivce dataset. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -23,7 +25,7 @@ class SquareObjectDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object - :param linked_service_name: Linked service reference. + :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference :param parameters: Parameters for dataset. @@ -32,7 +34,7 @@ class SquareObjectDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str """ @@ -41,6 +43,16 @@ class SquareObjectDataset(Dataset): 'type': {'required': True}, } - def __init__(self, linked_service_name, additional_properties=None, description=None, structure=None, parameters=None, annotations=None): - super(SquareObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations) + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, **kwargs): + super(SquareObjectDataset, self).__init__(**kwargs) self.type = 'SquareObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_object_dataset_py3.py new file mode 100644 index 000000000000..0fd0dd4abbfd --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_object_dataset_py3.py @@ -0,0 +1,58 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class SquareObjectDataset(Dataset): + """Square Serivce dataset. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Required. Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: + super(SquareObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + self.type = 'SquareObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_source.py index 1325235cdde6..98765da2fea8 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_source.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_source.py @@ -15,6 +15,8 @@ class SquareSource(CopySource): """A copy activity Square Serivce source. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -25,7 +27,7 @@ class SquareSource(CopySource): with resultType string), pattern: ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type source_retry_wait: object - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). @@ -44,7 +46,7 @@ class SquareSource(CopySource): 'query': {'key': 'query', 'type': 'object'}, } - def __init__(self, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None): - super(SquareSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait) - self.query = query + def __init__(self, **kwargs): + super(SquareSource, self).__init__(**kwargs) + self.query = kwargs.get('query', None) self.type = 'SquareSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_source_py3.py new file mode 100644 index 000000000000..81fbbeb9c150 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_source_py3.py @@ -0,0 +1,52 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class SquareSource(CopySource): + """A copy activity Square Serivce source. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Required. Constant filled by server. + :type type: str + :param query: A query to retrieve data from source. Type: string (or + Expression with resultType string). + :type query: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'query': {'key': 'query', 'type': 'object'}, + } + + def __init__(self, *, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None, **kwargs) -> None: + super(SquareSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, **kwargs) + self.query = query + self.type = 'SquareSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/ssis_package_location.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/ssis_package_location.py index 46971b3e5f1e..f81a9fafec37 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/ssis_package_location.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/ssis_package_location.py @@ -15,7 +15,9 @@ class SSISPackageLocation(Model): """SSIS package location. - :param package_path: The SSIS package path. + All required parameters must be populated in order to send to Azure. + + :param package_path: Required. The SSIS package path. :type package_path: str """ @@ -27,6 +29,6 @@ class SSISPackageLocation(Model): 'package_path': {'key': 'packagePath', 'type': 'str'}, } - def __init__(self, package_path): - super(SSISPackageLocation, self).__init__() - self.package_path = package_path + def __init__(self, **kwargs): + super(SSISPackageLocation, self).__init__(**kwargs) + self.package_path = kwargs.get('package_path', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/ssis_package_location_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/ssis_package_location_py3.py new file mode 100644 index 000000000000..0449d4ff2943 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/ssis_package_location_py3.py @@ -0,0 +1,34 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class SSISPackageLocation(Model): + """SSIS package location. + + All required parameters must be populated in order to send to Azure. + + :param package_path: Required. The SSIS package path. + :type package_path: str + """ + + _validation = { + 'package_path': {'required': True}, + } + + _attribute_map = { + 'package_path': {'key': 'packagePath', 'type': 'str'}, + } + + def __init__(self, *, package_path: str, **kwargs) -> None: + super(SSISPackageLocation, self).__init__(**kwargs) + self.package_path = package_path diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/staging_settings.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/staging_settings.py index 61efe881513e..05ca8dff2c52 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/staging_settings.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/staging_settings.py @@ -15,10 +15,12 @@ class StagingSettings(Model): """Staging settings. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] - :param linked_service_name: Staging linked service reference. + :param linked_service_name: Required. Staging linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference :param path: The path to storage for storing the interim data. Type: @@ -41,9 +43,9 @@ class StagingSettings(Model): 'enable_compression': {'key': 'enableCompression', 'type': 'object'}, } - def __init__(self, linked_service_name, additional_properties=None, path=None, enable_compression=None): - super(StagingSettings, self).__init__() - self.additional_properties = additional_properties - self.linked_service_name = linked_service_name - self.path = path - self.enable_compression = enable_compression + def __init__(self, **kwargs): + super(StagingSettings, self).__init__(**kwargs) + self.additional_properties = kwargs.get('additional_properties', None) + self.linked_service_name = kwargs.get('linked_service_name', None) + self.path = kwargs.get('path', None) + self.enable_compression = kwargs.get('enable_compression', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/staging_settings_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/staging_settings_py3.py new file mode 100644 index 000000000000..13b4353963a3 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/staging_settings_py3.py @@ -0,0 +1,51 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class StagingSettings(Model): + """Staging settings. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param linked_service_name: Required. Staging linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param path: The path to storage for storing the interim data. Type: + string (or Expression with resultType string). + :type path: object + :param enable_compression: Specifies whether to use compression when + copying data via an interim staging. Default value is false. Type: boolean + (or Expression with resultType boolean). + :type enable_compression: object + """ + + _validation = { + 'linked_service_name': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'path': {'key': 'path', 'type': 'object'}, + 'enable_compression': {'key': 'enableCompression', 'type': 'object'}, + } + + def __init__(self, *, linked_service_name, additional_properties=None, path=None, enable_compression=None, **kwargs) -> None: + super(StagingSettings, self).__init__(**kwargs) + self.additional_properties = additional_properties + self.linked_service_name = linked_service_name + self.path = path + self.enable_compression = enable_compression diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/stored_procedure_parameter.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/stored_procedure_parameter.py index 9f073a8d1c4a..63d2c3b80aa5 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/stored_procedure_parameter.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/stored_procedure_parameter.py @@ -15,7 +15,9 @@ class StoredProcedureParameter(Model): """SQL stored procedure parameter. - :param value: Stored procedure parameter value. Type: string (or + All required parameters must be populated in order to send to Azure. + + :param value: Required. Stored procedure parameter value. Type: string (or Expression with resultType string). :type value: object :param type: Stored procedure parameter type. Possible values include: @@ -33,7 +35,7 @@ class StoredProcedureParameter(Model): 'type': {'key': 'type', 'type': 'str'}, } - def __init__(self, value, type=None): - super(StoredProcedureParameter, self).__init__() - self.value = value - self.type = type + def __init__(self, **kwargs): + super(StoredProcedureParameter, self).__init__(**kwargs) + self.value = kwargs.get('value', None) + self.type = kwargs.get('type', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/stored_procedure_parameter_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/stored_procedure_parameter_py3.py new file mode 100644 index 000000000000..f5c55777e0c3 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/stored_procedure_parameter_py3.py @@ -0,0 +1,41 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class StoredProcedureParameter(Model): + """SQL stored procedure parameter. + + All required parameters must be populated in order to send to Azure. + + :param value: Required. Stored procedure parameter value. Type: string (or + Expression with resultType string). + :type value: object + :param type: Stored procedure parameter type. Possible values include: + 'String', 'Int', 'Decimal', 'Guid', 'Boolean', 'Date' + :type type: str or + ~azure.mgmt.datafactory.models.StoredProcedureParameterType + """ + + _validation = { + 'value': {'required': True}, + } + + _attribute_map = { + 'value': {'key': 'value', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, *, value, type=None, **kwargs) -> None: + super(StoredProcedureParameter, self).__init__(**kwargs) + self.value = value + self.type = type diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sub_resource.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sub_resource.py index 89a27ab2a0c3..c80b531db7d1 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sub_resource.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sub_resource.py @@ -42,8 +42,8 @@ class SubResource(Model): 'etag': {'key': 'etag', 'type': 'str'}, } - def __init__(self): - super(SubResource, self).__init__() + def __init__(self, **kwargs): + super(SubResource, self).__init__(**kwargs) self.id = None self.name = None self.type = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sub_resource_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sub_resource_py3.py new file mode 100644 index 000000000000..3b2d9ec62366 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sub_resource_py3.py @@ -0,0 +1,50 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class SubResource(Model): + """Azure Data Factory nested resource, which belongs to a factory. + + Variables are only populated by the server, and will be ignored when + sending a request. + + :ivar id: The resource identifier. + :vartype id: str + :ivar name: The resource name. + :vartype name: str + :ivar type: The resource type. + :vartype type: str + :ivar etag: Etag identifies change in the resource. + :vartype etag: str + """ + + _validation = { + 'id': {'readonly': True}, + 'name': {'readonly': True}, + 'type': {'readonly': True}, + 'etag': {'readonly': True}, + } + + _attribute_map = { + 'id': {'key': 'id', 'type': 'str'}, + 'name': {'key': 'name', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'str'}, + 'etag': {'key': 'etag', 'type': 'str'}, + } + + def __init__(self, **kwargs) -> None: + super(SubResource, self).__init__(**kwargs) + self.id = None + self.name = None + self.type = None + self.etag = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sybase_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sybase_linked_service.py index 27d8b3d179bf..634b4268bdb5 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sybase_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sybase_linked_service.py @@ -15,6 +15,8 @@ class SybaseLinkedService(LinkedService): """Linked service for Sybase data source. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -29,13 +31,13 @@ class SybaseLinkedService(LinkedService): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param server: Server name for connection. Type: string (or Expression - with resultType string). + :param server: Required. Server name for connection. Type: string (or + Expression with resultType string). :type server: object - :param database: Database name for connection. Type: string (or Expression - with resultType string). + :param database: Required. Database name for connection. Type: string (or + Expression with resultType string). :type database: object :param schema: Schema name for connection. Type: string (or Expression with resultType string). @@ -77,13 +79,13 @@ class SybaseLinkedService(LinkedService): 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, server, database, additional_properties=None, connect_via=None, description=None, parameters=None, annotations=None, schema=None, authentication_type=None, username=None, password=None, encrypted_credential=None): - super(SybaseLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations) - self.server = server - self.database = database - self.schema = schema - self.authentication_type = authentication_type - self.username = username - self.password = password - self.encrypted_credential = encrypted_credential + def __init__(self, **kwargs): + super(SybaseLinkedService, self).__init__(**kwargs) + self.server = kwargs.get('server', None) + self.database = kwargs.get('database', None) + self.schema = kwargs.get('schema', None) + self.authentication_type = kwargs.get('authentication_type', None) + self.username = kwargs.get('username', None) + self.password = kwargs.get('password', None) + self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'Sybase' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sybase_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sybase_linked_service_py3.py new file mode 100644 index 000000000000..b5c9c8788f25 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sybase_linked_service_py3.py @@ -0,0 +1,91 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class SybaseLinkedService(LinkedService): + """Linked service for Sybase data source. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param server: Required. Server name for connection. Type: string (or + Expression with resultType string). + :type server: object + :param database: Required. Database name for connection. Type: string (or + Expression with resultType string). + :type database: object + :param schema: Schema name for connection. Type: string (or Expression + with resultType string). + :type schema: object + :param authentication_type: AuthenticationType to be used for connection. + Possible values include: 'Basic', 'Windows' + :type authentication_type: str or + ~azure.mgmt.datafactory.models.SybaseAuthenticationType + :param username: Username for authentication. Type: string (or Expression + with resultType string). + :type username: object + :param password: Password for authentication. + :type password: ~azure.mgmt.datafactory.models.SecretBase + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'server': {'required': True}, + 'database': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'server': {'key': 'typeProperties.server', 'type': 'object'}, + 'database': {'key': 'typeProperties.database', 'type': 'object'}, + 'schema': {'key': 'typeProperties.schema', 'type': 'object'}, + 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, + 'username': {'key': 'typeProperties.username', 'type': 'object'}, + 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, *, server, database, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, schema=None, authentication_type=None, username=None, password=None, encrypted_credential=None, **kwargs) -> None: + super(SybaseLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.server = server + self.database = database + self.schema = schema + self.authentication_type = authentication_type + self.username = username + self.password = password + self.encrypted_credential = encrypted_credential + self.type = 'Sybase' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/tabular_translator.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/tabular_translator.py index 3278863cd8c7..ec499cceee24 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/tabular_translator.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/tabular_translator.py @@ -15,10 +15,12 @@ class TabularTranslator(CopyTranslator): """A copy activity tabular translator. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param column_mappings: Column mappings. Type: string (or Expression with resultType string). @@ -35,7 +37,7 @@ class TabularTranslator(CopyTranslator): 'column_mappings': {'key': 'columnMappings', 'type': 'object'}, } - def __init__(self, additional_properties=None, column_mappings=None): - super(TabularTranslator, self).__init__(additional_properties=additional_properties) - self.column_mappings = column_mappings + def __init__(self, **kwargs): + super(TabularTranslator, self).__init__(**kwargs) + self.column_mappings = kwargs.get('column_mappings', None) self.type = 'TabularTranslator' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/tabular_translator_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/tabular_translator_py3.py new file mode 100644 index 000000000000..eddbf396c0d6 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/tabular_translator_py3.py @@ -0,0 +1,43 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_translator import CopyTranslator + + +class TabularTranslator(CopyTranslator): + """A copy activity tabular translator. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param type: Required. Constant filled by server. + :type type: str + :param column_mappings: Column mappings. Type: string (or Expression with + resultType string). + :type column_mappings: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'type': {'key': 'type', 'type': 'str'}, + 'column_mappings': {'key': 'columnMappings', 'type': 'object'}, + } + + def __init__(self, *, additional_properties=None, column_mappings=None, **kwargs) -> None: + super(TabularTranslator, self).__init__(additional_properties=additional_properties, **kwargs) + self.column_mappings = column_mappings + self.type = 'TabularTranslator' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/teradata_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/teradata_linked_service.py index 7a1141496821..0df87c4f1487 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/teradata_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/teradata_linked_service.py @@ -15,6 +15,8 @@ class TeradataLinkedService(LinkedService): """Linked service for Teradata data source. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -29,10 +31,10 @@ class TeradataLinkedService(LinkedService): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param server: Server name for connection. Type: string (or Expression - with resultType string). + :param server: Required. Server name for connection. Type: string (or + Expression with resultType string). :type server: object :param schema: Schema name for connection. Type: string (or Expression with resultType string). @@ -72,12 +74,12 @@ class TeradataLinkedService(LinkedService): 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, server, additional_properties=None, connect_via=None, description=None, parameters=None, annotations=None, schema=None, authentication_type=None, username=None, password=None, encrypted_credential=None): - super(TeradataLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations) - self.server = server - self.schema = schema - self.authentication_type = authentication_type - self.username = username - self.password = password - self.encrypted_credential = encrypted_credential + def __init__(self, **kwargs): + super(TeradataLinkedService, self).__init__(**kwargs) + self.server = kwargs.get('server', None) + self.schema = kwargs.get('schema', None) + self.authentication_type = kwargs.get('authentication_type', None) + self.username = kwargs.get('username', None) + self.password = kwargs.get('password', None) + self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'Teradata' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/teradata_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/teradata_linked_service_py3.py new file mode 100644 index 000000000000..1fecbababcb2 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/teradata_linked_service_py3.py @@ -0,0 +1,85 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class TeradataLinkedService(LinkedService): + """Linked service for Teradata data source. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param server: Required. Server name for connection. Type: string (or + Expression with resultType string). + :type server: object + :param schema: Schema name for connection. Type: string (or Expression + with resultType string). + :type schema: object + :param authentication_type: AuthenticationType to be used for connection. + Possible values include: 'Basic', 'Windows' + :type authentication_type: str or + ~azure.mgmt.datafactory.models.TeradataAuthenticationType + :param username: Username for authentication. Type: string (or Expression + with resultType string). + :type username: object + :param password: Password for authentication. + :type password: ~azure.mgmt.datafactory.models.SecretBase + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'server': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'server': {'key': 'typeProperties.server', 'type': 'object'}, + 'schema': {'key': 'typeProperties.schema', 'type': 'object'}, + 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, + 'username': {'key': 'typeProperties.username', 'type': 'object'}, + 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, *, server, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, schema=None, authentication_type=None, username=None, password=None, encrypted_credential=None, **kwargs) -> None: + super(TeradataLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.server = server + self.schema = schema + self.authentication_type = authentication_type + self.username = username + self.password = password + self.encrypted_credential = encrypted_credential + self.type = 'Teradata' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/text_format.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/text_format.py index 60d963930e56..48f32bf10133 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/text_format.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/text_format.py @@ -15,6 +15,8 @@ class TextFormat(DatasetStorageFormat): """The data stored in text format. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -24,7 +26,7 @@ class TextFormat(DatasetStorageFormat): :param deserializer: Deserializer. Type: string (or Expression with resultType string). :type deserializer: object - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param column_delimiter: The column delimiter. Type: string (or Expression with resultType string). @@ -83,15 +85,15 @@ class TextFormat(DatasetStorageFormat): 'first_row_as_header': {'key': 'firstRowAsHeader', 'type': 'object'}, } - def __init__(self, additional_properties=None, serializer=None, deserializer=None, column_delimiter=None, row_delimiter=None, escape_char=None, quote_char=None, null_value=None, encoding_name=None, treat_empty_as_null=None, skip_line_count=None, first_row_as_header=None): - super(TextFormat, self).__init__(additional_properties=additional_properties, serializer=serializer, deserializer=deserializer) - self.column_delimiter = column_delimiter - self.row_delimiter = row_delimiter - self.escape_char = escape_char - self.quote_char = quote_char - self.null_value = null_value - self.encoding_name = encoding_name - self.treat_empty_as_null = treat_empty_as_null - self.skip_line_count = skip_line_count - self.first_row_as_header = first_row_as_header + def __init__(self, **kwargs): + super(TextFormat, self).__init__(**kwargs) + self.column_delimiter = kwargs.get('column_delimiter', None) + self.row_delimiter = kwargs.get('row_delimiter', None) + self.escape_char = kwargs.get('escape_char', None) + self.quote_char = kwargs.get('quote_char', None) + self.null_value = kwargs.get('null_value', None) + self.encoding_name = kwargs.get('encoding_name', None) + self.treat_empty_as_null = kwargs.get('treat_empty_as_null', None) + self.skip_line_count = kwargs.get('skip_line_count', None) + self.first_row_as_header = kwargs.get('first_row_as_header', None) self.type = 'TextFormat' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/text_format_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/text_format_py3.py new file mode 100644 index 000000000000..17a7b55327c4 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/text_format_py3.py @@ -0,0 +1,99 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset_storage_format import DatasetStorageFormat + + +class TextFormat(DatasetStorageFormat): + """The data stored in text format. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param serializer: Serializer. Type: string (or Expression with resultType + string). + :type serializer: object + :param deserializer: Deserializer. Type: string (or Expression with + resultType string). + :type deserializer: object + :param type: Required. Constant filled by server. + :type type: str + :param column_delimiter: The column delimiter. Type: string (or Expression + with resultType string). + :type column_delimiter: object + :param row_delimiter: The row delimiter. Type: string (or Expression with + resultType string). + :type row_delimiter: object + :param escape_char: The escape character. Type: string (or Expression with + resultType string). + :type escape_char: object + :param quote_char: The quote character. Type: string (or Expression with + resultType string). + :type quote_char: object + :param null_value: The null value string. Type: string (or Expression with + resultType string). + :type null_value: object + :param encoding_name: The code page name of the preferred encoding. If + miss, the default value is ΓÇ£utf-8ΓÇ¥, unless BOM denotes another Unicode + encoding. Refer to the ΓÇ£NameΓÇ¥ column of the table in the following + link to set supported values: + https://msdn.microsoft.com/library/system.text.encoding.aspx. Type: string + (or Expression with resultType string). + :type encoding_name: object + :param treat_empty_as_null: Treat empty column values in the text file as + null. The default value is true. Type: boolean (or Expression with + resultType boolean). + :type treat_empty_as_null: object + :param skip_line_count: The number of lines/rows to be skipped when + parsing text files. The default value is 0. Type: integer (or Expression + with resultType integer). + :type skip_line_count: object + :param first_row_as_header: When used as input, treat the first row of + data as headers. When used as output,write the headers into the output as + the first row of data. The default value is false. Type: boolean (or + Expression with resultType boolean). + :type first_row_as_header: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'serializer': {'key': 'serializer', 'type': 'object'}, + 'deserializer': {'key': 'deserializer', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'column_delimiter': {'key': 'columnDelimiter', 'type': 'object'}, + 'row_delimiter': {'key': 'rowDelimiter', 'type': 'object'}, + 'escape_char': {'key': 'escapeChar', 'type': 'object'}, + 'quote_char': {'key': 'quoteChar', 'type': 'object'}, + 'null_value': {'key': 'nullValue', 'type': 'object'}, + 'encoding_name': {'key': 'encodingName', 'type': 'object'}, + 'treat_empty_as_null': {'key': 'treatEmptyAsNull', 'type': 'object'}, + 'skip_line_count': {'key': 'skipLineCount', 'type': 'object'}, + 'first_row_as_header': {'key': 'firstRowAsHeader', 'type': 'object'}, + } + + def __init__(self, *, additional_properties=None, serializer=None, deserializer=None, column_delimiter=None, row_delimiter=None, escape_char=None, quote_char=None, null_value=None, encoding_name=None, treat_empty_as_null=None, skip_line_count=None, first_row_as_header=None, **kwargs) -> None: + super(TextFormat, self).__init__(additional_properties=additional_properties, serializer=serializer, deserializer=deserializer, **kwargs) + self.column_delimiter = column_delimiter + self.row_delimiter = row_delimiter + self.escape_char = escape_char + self.quote_char = quote_char + self.null_value = null_value + self.encoding_name = encoding_name + self.treat_empty_as_null = treat_empty_as_null + self.skip_line_count = skip_line_count + self.first_row_as_header = first_row_as_header + self.type = 'TextFormat' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger.py index dc630ae2ef06..0e7882159e95 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger.py @@ -22,6 +22,8 @@ class Trigger(Model): Variables are only populated by the server, and will be ignored when sending a request. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -32,7 +34,7 @@ class Trigger(Model): 'Started', 'Stopped', 'Disabled' :vartype runtime_state: str or ~azure.mgmt.datafactory.models.TriggerRuntimeState - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str """ @@ -52,9 +54,9 @@ class Trigger(Model): 'type': {'TumblingWindowTrigger': 'TumblingWindowTrigger', 'MultiplePipelineTrigger': 'MultiplePipelineTrigger'} } - def __init__(self, additional_properties=None, description=None): - super(Trigger, self).__init__() - self.additional_properties = additional_properties - self.description = description + def __init__(self, **kwargs): + super(Trigger, self).__init__(**kwargs) + self.additional_properties = kwargs.get('additional_properties', None) + self.description = kwargs.get('description', None) self.runtime_state = None self.type = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger_pipeline_reference.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger_pipeline_reference.py index 52d0dc2609bf..70c9f2904347 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger_pipeline_reference.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger_pipeline_reference.py @@ -26,7 +26,7 @@ class TriggerPipelineReference(Model): 'parameters': {'key': 'parameters', 'type': '{object}'}, } - def __init__(self, pipeline_reference=None, parameters=None): - super(TriggerPipelineReference, self).__init__() - self.pipeline_reference = pipeline_reference - self.parameters = parameters + def __init__(self, **kwargs): + super(TriggerPipelineReference, self).__init__(**kwargs) + self.pipeline_reference = kwargs.get('pipeline_reference', None) + self.parameters = kwargs.get('parameters', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger_pipeline_reference_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger_pipeline_reference_py3.py new file mode 100644 index 000000000000..e32af8006326 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger_pipeline_reference_py3.py @@ -0,0 +1,32 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class TriggerPipelineReference(Model): + """Pipeline that needs to be triggered with the given parameters. + + :param pipeline_reference: Pipeline reference. + :type pipeline_reference: ~azure.mgmt.datafactory.models.PipelineReference + :param parameters: Pipeline parameters. + :type parameters: dict[str, object] + """ + + _attribute_map = { + 'pipeline_reference': {'key': 'pipelineReference', 'type': 'PipelineReference'}, + 'parameters': {'key': 'parameters', 'type': '{object}'}, + } + + def __init__(self, *, pipeline_reference=None, parameters=None, **kwargs) -> None: + super(TriggerPipelineReference, self).__init__(**kwargs) + self.pipeline_reference = pipeline_reference + self.parameters = parameters diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger_py3.py new file mode 100644 index 000000000000..3e232b149f0b --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger_py3.py @@ -0,0 +1,62 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class Trigger(Model): + """Azure data factory nested object which contains information about creating + pipeline run. + + You probably want to use the sub-classes and not this class directly. Known + sub-classes are: TumblingWindowTrigger, MultiplePipelineTrigger + + Variables are only populated by the server, and will be ignored when + sending a request. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param description: Trigger description. + :type description: str + :ivar runtime_state: Indicates if trigger is running or not. Updated when + Start/Stop APIs are called on the Trigger. Possible values include: + 'Started', 'Stopped', 'Disabled' + :vartype runtime_state: str or + ~azure.mgmt.datafactory.models.TriggerRuntimeState + :param type: Required. Constant filled by server. + :type type: str + """ + + _validation = { + 'runtime_state': {'readonly': True}, + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'runtime_state': {'key': 'runtimeState', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + _subtype_map = { + 'type': {'TumblingWindowTrigger': 'TumblingWindowTrigger', 'MultiplePipelineTrigger': 'MultiplePipelineTrigger'} + } + + def __init__(self, *, additional_properties=None, description: str=None, **kwargs) -> None: + super(Trigger, self).__init__(**kwargs) + self.additional_properties = additional_properties + self.description = description + self.runtime_state = None + self.type = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger_resource.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger_resource.py index ea35206bbed7..539ac4775350 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger_resource.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger_resource.py @@ -18,6 +18,8 @@ class TriggerResource(SubResource): Variables are only populated by the server, and will be ignored when sending a request. + All required parameters must be populated in order to send to Azure. + :ivar id: The resource identifier. :vartype id: str :ivar name: The resource name. @@ -26,7 +28,7 @@ class TriggerResource(SubResource): :vartype type: str :ivar etag: Etag identifies change in the resource. :vartype etag: str - :param properties: Properties of the trigger. + :param properties: Required. Properties of the trigger. :type properties: ~azure.mgmt.datafactory.models.Trigger """ @@ -46,6 +48,6 @@ class TriggerResource(SubResource): 'properties': {'key': 'properties', 'type': 'Trigger'}, } - def __init__(self, properties): - super(TriggerResource, self).__init__() - self.properties = properties + def __init__(self, **kwargs): + super(TriggerResource, self).__init__(**kwargs) + self.properties = kwargs.get('properties', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger_resource_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger_resource_py3.py new file mode 100644 index 000000000000..e90d1a8010fb --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger_resource_py3.py @@ -0,0 +1,53 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .sub_resource import SubResource + + +class TriggerResource(SubResource): + """Trigger resource type. + + Variables are only populated by the server, and will be ignored when + sending a request. + + All required parameters must be populated in order to send to Azure. + + :ivar id: The resource identifier. + :vartype id: str + :ivar name: The resource name. + :vartype name: str + :ivar type: The resource type. + :vartype type: str + :ivar etag: Etag identifies change in the resource. + :vartype etag: str + :param properties: Required. Properties of the trigger. + :type properties: ~azure.mgmt.datafactory.models.Trigger + """ + + _validation = { + 'id': {'readonly': True}, + 'name': {'readonly': True}, + 'type': {'readonly': True}, + 'etag': {'readonly': True}, + 'properties': {'required': True}, + } + + _attribute_map = { + 'id': {'key': 'id', 'type': 'str'}, + 'name': {'key': 'name', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'str'}, + 'etag': {'key': 'etag', 'type': 'str'}, + 'properties': {'key': 'properties', 'type': 'Trigger'}, + } + + def __init__(self, *, properties, **kwargs) -> None: + super(TriggerResource, self).__init__(, **kwargs) + self.properties = properties diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger_run.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger_run.py index 2fb74b9dcbab..9fad7bbfd9fa 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger_run.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger_run.py @@ -65,9 +65,9 @@ class TriggerRun(Model): 'triggered_pipelines': {'key': 'triggeredPipelines', 'type': '{str}'}, } - def __init__(self, additional_properties=None): - super(TriggerRun, self).__init__() - self.additional_properties = additional_properties + def __init__(self, **kwargs): + super(TriggerRun, self).__init__(**kwargs) + self.additional_properties = kwargs.get('additional_properties', None) self.trigger_run_id = None self.trigger_name = None self.trigger_type = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger_run_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger_run_py3.py new file mode 100644 index 000000000000..5a9fe50f6894 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger_run_py3.py @@ -0,0 +1,78 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class TriggerRun(Model): + """Trigger runs. + + Variables are only populated by the server, and will be ignored when + sending a request. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :ivar trigger_run_id: Trigger run id. + :vartype trigger_run_id: str + :ivar trigger_name: Trigger name. + :vartype trigger_name: str + :ivar trigger_type: Trigger type. + :vartype trigger_type: str + :ivar trigger_run_timestamp: Trigger run start time. + :vartype trigger_run_timestamp: datetime + :ivar status: Trigger run status. Possible values include: 'Succeeded', + 'Failed', 'Inprogress' + :vartype status: str or ~azure.mgmt.datafactory.models.TriggerRunStatus + :ivar message: Trigger error message. + :vartype message: str + :ivar properties: List of property name and value related to trigger run. + Name, value pair depends on type of trigger. + :vartype properties: dict[str, str] + :ivar triggered_pipelines: List of pipeline name and run Id triggered by + the trigger run. + :vartype triggered_pipelines: dict[str, str] + """ + + _validation = { + 'trigger_run_id': {'readonly': True}, + 'trigger_name': {'readonly': True}, + 'trigger_type': {'readonly': True}, + 'trigger_run_timestamp': {'readonly': True}, + 'status': {'readonly': True}, + 'message': {'readonly': True}, + 'properties': {'readonly': True}, + 'triggered_pipelines': {'readonly': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'trigger_run_id': {'key': 'triggerRunId', 'type': 'str'}, + 'trigger_name': {'key': 'triggerName', 'type': 'str'}, + 'trigger_type': {'key': 'triggerType', 'type': 'str'}, + 'trigger_run_timestamp': {'key': 'triggerRunTimestamp', 'type': 'iso-8601'}, + 'status': {'key': 'status', 'type': 'str'}, + 'message': {'key': 'message', 'type': 'str'}, + 'properties': {'key': 'properties', 'type': '{str}'}, + 'triggered_pipelines': {'key': 'triggeredPipelines', 'type': '{str}'}, + } + + def __init__(self, *, additional_properties=None, **kwargs) -> None: + super(TriggerRun, self).__init__(**kwargs) + self.additional_properties = additional_properties + self.trigger_run_id = None + self.trigger_name = None + self.trigger_type = None + self.trigger_run_timestamp = None + self.status = None + self.message = None + self.properties = None + self.triggered_pipelines = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/tumbling_window_trigger.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/tumbling_window_trigger.py index 22f7693bed1f..ec6d09496677 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/tumbling_window_trigger.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/tumbling_window_trigger.py @@ -20,6 +20,8 @@ class TumblingWindowTrigger(Trigger): Variables are only populated by the server, and will be ignored when sending a request. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -30,21 +32,21 @@ class TumblingWindowTrigger(Trigger): 'Started', 'Stopped', 'Disabled' :vartype runtime_state: str or ~azure.mgmt.datafactory.models.TriggerRuntimeState - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param pipeline: Pipeline for which runs are created when an event is - fired for trigger window that is ready. + :param pipeline: Required. Pipeline for which runs are created when an + event is fired for trigger window that is ready. :type pipeline: ~azure.mgmt.datafactory.models.TriggerPipelineReference - :param frequency: The frequency of the time windows. Possible values - include: 'Minute', 'Hour' + :param frequency: Required. The frequency of the time windows. Possible + values include: 'Minute', 'Hour' :type frequency: str or ~azure.mgmt.datafactory.models.TumblingWindowFrequency - :param interval: The interval of the time windows. The minimum interval - allowed is 15 Minutes. + :param interval: Required. The interval of the time windows. The minimum + interval allowed is 15 Minutes. :type interval: int - :param start_time: The start time for the time period for the trigger - during which events are fired for windows that are ready. Only UTC time is - currently supported. + :param start_time: Required. The start time for the time period for the + trigger during which events are fired for windows that are ready. Only UTC + time is currently supported. :type start_time: datetime :param end_time: The end time for the time period for the trigger during which events are fired for windows that are ready. Only UTC time is @@ -55,8 +57,8 @@ class TumblingWindowTrigger(Trigger): default is 0. Type: string (or Expression with resultType string), pattern: ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type delay: object - :param max_concurrency: The max number of parallel time windows (ready for - execution) for which a new run is triggered. + :param max_concurrency: Required. The max number of parallel time windows + (ready for execution) for which a new run is triggered. :type max_concurrency: int :param retry_policy: Retry policy that will be applied for failed pipeline runs. @@ -88,14 +90,14 @@ class TumblingWindowTrigger(Trigger): 'retry_policy': {'key': 'typeProperties.retryPolicy', 'type': 'RetryPolicy'}, } - def __init__(self, pipeline, frequency, interval, start_time, max_concurrency, additional_properties=None, description=None, end_time=None, delay=None, retry_policy=None): - super(TumblingWindowTrigger, self).__init__(additional_properties=additional_properties, description=description) - self.pipeline = pipeline - self.frequency = frequency - self.interval = interval - self.start_time = start_time - self.end_time = end_time - self.delay = delay - self.max_concurrency = max_concurrency - self.retry_policy = retry_policy + def __init__(self, **kwargs): + super(TumblingWindowTrigger, self).__init__(**kwargs) + self.pipeline = kwargs.get('pipeline', None) + self.frequency = kwargs.get('frequency', None) + self.interval = kwargs.get('interval', None) + self.start_time = kwargs.get('start_time', None) + self.end_time = kwargs.get('end_time', None) + self.delay = kwargs.get('delay', None) + self.max_concurrency = kwargs.get('max_concurrency', None) + self.retry_policy = kwargs.get('retry_policy', None) self.type = 'TumblingWindowTrigger' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/tumbling_window_trigger_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/tumbling_window_trigger_py3.py new file mode 100644 index 000000000000..a436cc6fa2b4 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/tumbling_window_trigger_py3.py @@ -0,0 +1,103 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .trigger import Trigger + + +class TumblingWindowTrigger(Trigger): + """Trigger that schedules pipeline runs for all fixed time interval windows + from a start time without gaps and also supports backfill scenarios (when + start time is in the past). + + Variables are only populated by the server, and will be ignored when + sending a request. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param description: Trigger description. + :type description: str + :ivar runtime_state: Indicates if trigger is running or not. Updated when + Start/Stop APIs are called on the Trigger. Possible values include: + 'Started', 'Stopped', 'Disabled' + :vartype runtime_state: str or + ~azure.mgmt.datafactory.models.TriggerRuntimeState + :param type: Required. Constant filled by server. + :type type: str + :param pipeline: Required. Pipeline for which runs are created when an + event is fired for trigger window that is ready. + :type pipeline: ~azure.mgmt.datafactory.models.TriggerPipelineReference + :param frequency: Required. The frequency of the time windows. Possible + values include: 'Minute', 'Hour' + :type frequency: str or + ~azure.mgmt.datafactory.models.TumblingWindowFrequency + :param interval: Required. The interval of the time windows. The minimum + interval allowed is 15 Minutes. + :type interval: int + :param start_time: Required. The start time for the time period for the + trigger during which events are fired for windows that are ready. Only UTC + time is currently supported. + :type start_time: datetime + :param end_time: The end time for the time period for the trigger during + which events are fired for windows that are ready. Only UTC time is + currently supported. + :type end_time: datetime + :param delay: Specifies how long the trigger waits past due time before + triggering new run. It doesn't alter window start and end time. The + default is 0. Type: string (or Expression with resultType string), + pattern: ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type delay: object + :param max_concurrency: Required. The max number of parallel time windows + (ready for execution) for which a new run is triggered. + :type max_concurrency: int + :param retry_policy: Retry policy that will be applied for failed pipeline + runs. + :type retry_policy: ~azure.mgmt.datafactory.models.RetryPolicy + """ + + _validation = { + 'runtime_state': {'readonly': True}, + 'type': {'required': True}, + 'pipeline': {'required': True}, + 'frequency': {'required': True}, + 'interval': {'required': True}, + 'start_time': {'required': True}, + 'max_concurrency': {'required': True, 'maximum': 50, 'minimum': 1}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'runtime_state': {'key': 'runtimeState', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'str'}, + 'pipeline': {'key': 'pipeline', 'type': 'TriggerPipelineReference'}, + 'frequency': {'key': 'typeProperties.frequency', 'type': 'str'}, + 'interval': {'key': 'typeProperties.interval', 'type': 'int'}, + 'start_time': {'key': 'typeProperties.startTime', 'type': 'iso-8601'}, + 'end_time': {'key': 'typeProperties.endTime', 'type': 'iso-8601'}, + 'delay': {'key': 'typeProperties.delay', 'type': 'object'}, + 'max_concurrency': {'key': 'typeProperties.maxConcurrency', 'type': 'int'}, + 'retry_policy': {'key': 'typeProperties.retryPolicy', 'type': 'RetryPolicy'}, + } + + def __init__(self, *, pipeline, frequency, interval: int, start_time, max_concurrency: int, additional_properties=None, description: str=None, end_time=None, delay=None, retry_policy=None, **kwargs) -> None: + super(TumblingWindowTrigger, self).__init__(additional_properties=additional_properties, description=description, **kwargs) + self.pipeline = pipeline + self.frequency = frequency + self.interval = interval + self.start_time = start_time + self.end_time = end_time + self.delay = delay + self.max_concurrency = max_concurrency + self.retry_policy = retry_policy + self.type = 'TumblingWindowTrigger' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/until_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/until_activity.py index 590d4303e712..7cf4fc07594b 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/until_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/until_activity.py @@ -16,19 +16,21 @@ class UntilActivity(ControlActivity): """This activity executes inner activities until the specified boolean expression results to true or timeout is reached, whichever is earlier. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] - :param name: Activity name. + :param name: Required. Activity name. :type name: str :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param expression: An expression that would evaluate to Boolean. The loop - will continue until this expression evaluates to true + :param expression: Required. An expression that would evaluate to Boolean. + The loop will continue until this expression evaluates to true :type expression: ~azure.mgmt.datafactory.models.Expression :param timeout: Specifies the timeout for the activity to run. If there is no value specified, it takes the value of TimeSpan.FromDays(7) which is 1 @@ -37,7 +39,7 @@ class UntilActivity(ControlActivity): string (or Expression with resultType string), pattern: ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type timeout: object - :param activities: List of activities to execute. + :param activities: Required. List of activities to execute. :type activities: list[~azure.mgmt.datafactory.models.Activity] """ @@ -59,9 +61,9 @@ class UntilActivity(ControlActivity): 'activities': {'key': 'typeProperties.activities', 'type': '[Activity]'}, } - def __init__(self, name, expression, activities, additional_properties=None, description=None, depends_on=None, timeout=None): - super(UntilActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on) - self.expression = expression - self.timeout = timeout - self.activities = activities + def __init__(self, **kwargs): + super(UntilActivity, self).__init__(**kwargs) + self.expression = kwargs.get('expression', None) + self.timeout = kwargs.get('timeout', None) + self.activities = kwargs.get('activities', None) self.type = 'Until' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/until_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/until_activity_py3.py new file mode 100644 index 000000000000..5d3c2739feef --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/until_activity_py3.py @@ -0,0 +1,69 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .control_activity import ControlActivity + + +class UntilActivity(ControlActivity): + """This activity executes inner activities until the specified boolean + expression results to true or timeout is reached, whichever is earlier. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param name: Required. Activity name. + :type name: str + :param description: Activity description. + :type description: str + :param depends_on: Activity depends on condition. + :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param type: Required. Constant filled by server. + :type type: str + :param expression: Required. An expression that would evaluate to Boolean. + The loop will continue until this expression evaluates to true + :type expression: ~azure.mgmt.datafactory.models.Expression + :param timeout: Specifies the timeout for the activity to run. If there is + no value specified, it takes the value of TimeSpan.FromDays(7) which is 1 + week as default. Type: string (or Expression with resultType string), + pattern: ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). Type: + string (or Expression with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type timeout: object + :param activities: Required. List of activities to execute. + :type activities: list[~azure.mgmt.datafactory.models.Activity] + """ + + _validation = { + 'name': {'required': True}, + 'type': {'required': True}, + 'expression': {'required': True}, + 'activities': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'name': {'key': 'name', 'type': 'str'}, + 'description': {'key': 'description', 'type': 'str'}, + 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'expression': {'key': 'typeProperties.expression', 'type': 'Expression'}, + 'timeout': {'key': 'typeProperties.timeout', 'type': 'object'}, + 'activities': {'key': 'typeProperties.activities', 'type': '[Activity]'}, + } + + def __init__(self, *, name: str, expression, activities, additional_properties=None, description: str=None, depends_on=None, timeout=None, **kwargs) -> None: + super(UntilActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, **kwargs) + self.expression = expression + self.timeout = timeout + self.activities = activities + self.type = 'Until' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/update_integration_runtime_node_request.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/update_integration_runtime_node_request.py index 6bb70551971d..c6460310225a 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/update_integration_runtime_node_request.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/update_integration_runtime_node_request.py @@ -29,6 +29,6 @@ class UpdateIntegrationRuntimeNodeRequest(Model): 'concurrent_jobs_limit': {'key': 'concurrentJobsLimit', 'type': 'int'}, } - def __init__(self, concurrent_jobs_limit=None): - super(UpdateIntegrationRuntimeNodeRequest, self).__init__() - self.concurrent_jobs_limit = concurrent_jobs_limit + def __init__(self, **kwargs): + super(UpdateIntegrationRuntimeNodeRequest, self).__init__(**kwargs) + self.concurrent_jobs_limit = kwargs.get('concurrent_jobs_limit', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/update_integration_runtime_node_request_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/update_integration_runtime_node_request_py3.py new file mode 100644 index 000000000000..de1605885139 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/update_integration_runtime_node_request_py3.py @@ -0,0 +1,34 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class UpdateIntegrationRuntimeNodeRequest(Model): + """Update integration runtime node request. + + :param concurrent_jobs_limit: The number of concurrent jobs permitted to + run on the integration runtime node. Values between 1 and + maxConcurrentJobs(inclusive) are allowed. + :type concurrent_jobs_limit: int + """ + + _validation = { + 'concurrent_jobs_limit': {'minimum': 1}, + } + + _attribute_map = { + 'concurrent_jobs_limit': {'key': 'concurrentJobsLimit', 'type': 'int'}, + } + + def __init__(self, *, concurrent_jobs_limit: int=None, **kwargs) -> None: + super(UpdateIntegrationRuntimeNodeRequest, self).__init__(**kwargs) + self.concurrent_jobs_limit = concurrent_jobs_limit diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/update_integration_runtime_request.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/update_integration_runtime_request.py index f4e02900146a..bd5e332b50f5 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/update_integration_runtime_request.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/update_integration_runtime_request.py @@ -32,7 +32,7 @@ class UpdateIntegrationRuntimeRequest(Model): 'update_delay_offset': {'key': 'updateDelayOffset', 'type': 'str'}, } - def __init__(self, auto_update=None, update_delay_offset=None): - super(UpdateIntegrationRuntimeRequest, self).__init__() - self.auto_update = auto_update - self.update_delay_offset = update_delay_offset + def __init__(self, **kwargs): + super(UpdateIntegrationRuntimeRequest, self).__init__(**kwargs) + self.auto_update = kwargs.get('auto_update', None) + self.update_delay_offset = kwargs.get('update_delay_offset', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/update_integration_runtime_request_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/update_integration_runtime_request_py3.py new file mode 100644 index 000000000000..731cb942b472 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/update_integration_runtime_request_py3.py @@ -0,0 +1,38 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class UpdateIntegrationRuntimeRequest(Model): + """Update integration runtime request. + + :param auto_update: Enables or disables the auto-update feature of the + self-hosted integration runtime. See + https://go.microsoft.com/fwlink/?linkid=854189. Possible values include: + 'On', 'Off' + :type auto_update: str or + ~azure.mgmt.datafactory.models.IntegrationRuntimeAutoUpdate + :param update_delay_offset: The time offset (in hours) in the day, e.g., + PT03H is 3 hours. The integration runtime auto update will happen on that + time. + :type update_delay_offset: str + """ + + _attribute_map = { + 'auto_update': {'key': 'autoUpdate', 'type': 'str'}, + 'update_delay_offset': {'key': 'updateDelayOffset', 'type': 'str'}, + } + + def __init__(self, *, auto_update=None, update_delay_offset: str=None, **kwargs) -> None: + super(UpdateIntegrationRuntimeRequest, self).__init__(**kwargs) + self.auto_update = auto_update + self.update_delay_offset = update_delay_offset diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/vertica_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/vertica_linked_service.py index 8310c2f1b2dc..b6bb2c2d389d 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/vertica_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/vertica_linked_service.py @@ -15,6 +15,8 @@ class VerticaLinkedService(LinkedService): """Vertica linked service. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -29,7 +31,7 @@ class VerticaLinkedService(LinkedService): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param connection_string: An ODBC connection string. :type connection_string: ~azure.mgmt.datafactory.models.SecretBase @@ -54,8 +56,8 @@ class VerticaLinkedService(LinkedService): 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, additional_properties=None, connect_via=None, description=None, parameters=None, annotations=None, connection_string=None, encrypted_credential=None): - super(VerticaLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations) - self.connection_string = connection_string - self.encrypted_credential = encrypted_credential + def __init__(self, **kwargs): + super(VerticaLinkedService, self).__init__(**kwargs) + self.connection_string = kwargs.get('connection_string', None) + self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'Vertica' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/vertica_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/vertica_linked_service_py3.py new file mode 100644 index 000000000000..8162656a5e20 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/vertica_linked_service_py3.py @@ -0,0 +1,63 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class VerticaLinkedService(LinkedService): + """Vertica linked service. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param connection_string: An ODBC connection string. + :type connection_string: ~azure.mgmt.datafactory.models.SecretBase + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'SecretBase'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, *, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, connection_string=None, encrypted_credential=None, **kwargs) -> None: + super(VerticaLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.connection_string = connection_string + self.encrypted_credential = encrypted_credential + self.type = 'Vertica' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/vertica_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/vertica_source.py index 081b51fb0d99..1670c0e9fc49 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/vertica_source.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/vertica_source.py @@ -15,6 +15,8 @@ class VerticaSource(CopySource): """A copy activity Vertica source. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -25,7 +27,7 @@ class VerticaSource(CopySource): with resultType string), pattern: ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type source_retry_wait: object - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). @@ -44,7 +46,7 @@ class VerticaSource(CopySource): 'query': {'key': 'query', 'type': 'object'}, } - def __init__(self, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None): - super(VerticaSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait) - self.query = query + def __init__(self, **kwargs): + super(VerticaSource, self).__init__(**kwargs) + self.query = kwargs.get('query', None) self.type = 'VerticaSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/vertica_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/vertica_source_py3.py new file mode 100644 index 000000000000..25eb02039b81 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/vertica_source_py3.py @@ -0,0 +1,52 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class VerticaSource(CopySource): + """A copy activity Vertica source. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Required. Constant filled by server. + :type type: str + :param query: A query to retrieve data from source. Type: string (or + Expression with resultType string). + :type query: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'query': {'key': 'query', 'type': 'object'}, + } + + def __init__(self, *, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None, **kwargs) -> None: + super(VerticaSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, **kwargs) + self.query = query + self.type = 'VerticaSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/vertica_table_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/vertica_table_dataset.py index 04b6c209f19f..98a70f28cee9 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/vertica_table_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/vertica_table_dataset.py @@ -15,6 +15,8 @@ class VerticaTableDataset(Dataset): """Vertica dataset. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -23,7 +25,7 @@ class VerticaTableDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object - :param linked_service_name: Linked service reference. + :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference :param parameters: Parameters for dataset. @@ -32,7 +34,7 @@ class VerticaTableDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str """ @@ -41,6 +43,16 @@ class VerticaTableDataset(Dataset): 'type': {'required': True}, } - def __init__(self, linked_service_name, additional_properties=None, description=None, structure=None, parameters=None, annotations=None): - super(VerticaTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations) + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, **kwargs): + super(VerticaTableDataset, self).__init__(**kwargs) self.type = 'VerticaTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/vertica_table_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/vertica_table_dataset_py3.py new file mode 100644 index 000000000000..6681fca04f33 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/vertica_table_dataset_py3.py @@ -0,0 +1,58 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class VerticaTableDataset(Dataset): + """Vertica dataset. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Required. Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: + super(VerticaTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + self.type = 'VerticaTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/wait_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/wait_activity.py index 188289735a34..0ed8536f0afc 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/wait_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/wait_activity.py @@ -15,18 +15,20 @@ class WaitActivity(ControlActivity): """This activity suspends pipeline execution for the specified interval. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] - :param name: Activity name. + :param name: Required. Activity name. :type name: str :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param wait_time_in_seconds: Duration in seconds. + :param wait_time_in_seconds: Required. Duration in seconds. :type wait_time_in_seconds: int """ @@ -45,7 +47,7 @@ class WaitActivity(ControlActivity): 'wait_time_in_seconds': {'key': 'typeProperties.waitTimeInSeconds', 'type': 'int'}, } - def __init__(self, name, wait_time_in_seconds, additional_properties=None, description=None, depends_on=None): - super(WaitActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on) - self.wait_time_in_seconds = wait_time_in_seconds + def __init__(self, **kwargs): + super(WaitActivity, self).__init__(**kwargs) + self.wait_time_in_seconds = kwargs.get('wait_time_in_seconds', None) self.type = 'Wait' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/wait_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/wait_activity_py3.py new file mode 100644 index 000000000000..7c7dc1d1f30f --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/wait_activity_py3.py @@ -0,0 +1,53 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .control_activity import ControlActivity + + +class WaitActivity(ControlActivity): + """This activity suspends pipeline execution for the specified interval. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param name: Required. Activity name. + :type name: str + :param description: Activity description. + :type description: str + :param depends_on: Activity depends on condition. + :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param type: Required. Constant filled by server. + :type type: str + :param wait_time_in_seconds: Required. Duration in seconds. + :type wait_time_in_seconds: int + """ + + _validation = { + 'name': {'required': True}, + 'type': {'required': True}, + 'wait_time_in_seconds': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'name': {'key': 'name', 'type': 'str'}, + 'description': {'key': 'description', 'type': 'str'}, + 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'wait_time_in_seconds': {'key': 'typeProperties.waitTimeInSeconds', 'type': 'int'}, + } + + def __init__(self, *, name: str, wait_time_in_seconds: int, additional_properties=None, description: str=None, depends_on=None, **kwargs) -> None: + super(WaitActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, **kwargs) + self.wait_time_in_seconds = wait_time_in_seconds + self.type = 'Wait' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_activity.py index 8774e991fb03..65831d837ae9 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_activity.py @@ -15,27 +15,29 @@ class WebActivity(ExecutionActivity): """Web activity. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] - :param name: Activity name. + :param name: Required. Activity name. :type name: str :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference :param policy: Activity policy. :type policy: ~azure.mgmt.datafactory.models.ActivityPolicy - :param method: Rest API method for target endpoint. Possible values - include: 'GET', 'POST', 'PUT', 'DELETE' + :param method: Required. Rest API method for target endpoint. Possible + values include: 'GET', 'POST', 'PUT', 'DELETE' :type method: str or ~azure.mgmt.datafactory.models.WebActivityMethod - :param url: Web activity target endpoint and path. Type: string (or - Expression with resultType string). + :param url: Required. Web activity target endpoint and path. Type: string + (or Expression with resultType string). :type url: object :param headers: Represents the headers that will be sent to the request. For example, to set the language and type on a request: "headers" : { @@ -81,13 +83,13 @@ class WebActivity(ExecutionActivity): 'linked_services': {'key': 'typeProperties.linkedServices', 'type': '[LinkedServiceReference]'}, } - def __init__(self, name, method, url, additional_properties=None, description=None, depends_on=None, linked_service_name=None, policy=None, headers=None, body=None, authentication=None, datasets=None, linked_services=None): - super(WebActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, linked_service_name=linked_service_name, policy=policy) - self.method = method - self.url = url - self.headers = headers - self.body = body - self.authentication = authentication - self.datasets = datasets - self.linked_services = linked_services + def __init__(self, **kwargs): + super(WebActivity, self).__init__(**kwargs) + self.method = kwargs.get('method', None) + self.url = kwargs.get('url', None) + self.headers = kwargs.get('headers', None) + self.body = kwargs.get('body', None) + self.authentication = kwargs.get('authentication', None) + self.datasets = kwargs.get('datasets', None) + self.linked_services = kwargs.get('linked_services', None) self.type = 'WebActivity' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_activity_authentication.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_activity_authentication.py index 3f2d35b9b62e..6ebb193ae5e9 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_activity_authentication.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_activity_authentication.py @@ -15,7 +15,10 @@ class WebActivityAuthentication(Model): """Web activity authentication properties. - :param type: Web activity authentication (Basic/ClientCertificate/MSI) + All required parameters must be populated in order to send to Azure. + + :param type: Required. Web activity authentication + (Basic/ClientCertificate/MSI) :type type: str :param pfx: Base64-encoded contents of a PFX file. :type pfx: ~azure.mgmt.datafactory.models.SecureString @@ -41,10 +44,10 @@ class WebActivityAuthentication(Model): 'resource': {'key': 'resource', 'type': 'str'}, } - def __init__(self, type, pfx=None, username=None, password=None, resource=None): - super(WebActivityAuthentication, self).__init__() - self.type = type - self.pfx = pfx - self.username = username - self.password = password - self.resource = resource + def __init__(self, **kwargs): + super(WebActivityAuthentication, self).__init__(**kwargs) + self.type = kwargs.get('type', None) + self.pfx = kwargs.get('pfx', None) + self.username = kwargs.get('username', None) + self.password = kwargs.get('password', None) + self.resource = kwargs.get('resource', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_activity_authentication_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_activity_authentication_py3.py new file mode 100644 index 000000000000..4c2b68ba7161 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_activity_authentication_py3.py @@ -0,0 +1,53 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class WebActivityAuthentication(Model): + """Web activity authentication properties. + + All required parameters must be populated in order to send to Azure. + + :param type: Required. Web activity authentication + (Basic/ClientCertificate/MSI) + :type type: str + :param pfx: Base64-encoded contents of a PFX file. + :type pfx: ~azure.mgmt.datafactory.models.SecureString + :param username: Web activity authentication user name for basic + authentication. + :type username: str + :param password: Password for the PFX file or basic authentication. + :type password: ~azure.mgmt.datafactory.models.SecureString + :param resource: Resource for which Azure Auth token will be requested + when using MSI Authentication. + :type resource: str + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'type': {'key': 'type', 'type': 'str'}, + 'pfx': {'key': 'pfx', 'type': 'SecureString'}, + 'username': {'key': 'username', 'type': 'str'}, + 'password': {'key': 'password', 'type': 'SecureString'}, + 'resource': {'key': 'resource', 'type': 'str'}, + } + + def __init__(self, *, type: str, pfx=None, username: str=None, password=None, resource: str=None, **kwargs) -> None: + super(WebActivityAuthentication, self).__init__(**kwargs) + self.type = type + self.pfx = pfx + self.username = username + self.password = password + self.resource = resource diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_activity_py3.py new file mode 100644 index 000000000000..5b3220509c34 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_activity_py3.py @@ -0,0 +1,95 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .execution_activity import ExecutionActivity + + +class WebActivity(ExecutionActivity): + """Web activity. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param name: Required. Activity name. + :type name: str + :param description: Activity description. + :type description: str + :param depends_on: Activity depends on condition. + :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param type: Required. Constant filled by server. + :type type: str + :param linked_service_name: Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param policy: Activity policy. + :type policy: ~azure.mgmt.datafactory.models.ActivityPolicy + :param method: Required. Rest API method for target endpoint. Possible + values include: 'GET', 'POST', 'PUT', 'DELETE' + :type method: str or ~azure.mgmt.datafactory.models.WebActivityMethod + :param url: Required. Web activity target endpoint and path. Type: string + (or Expression with resultType string). + :type url: object + :param headers: Represents the headers that will be sent to the request. + For example, to set the language and type on a request: "headers" : { + "Accept-Language": "en-us", "Content-Type": "application/json" }. Type: + string (or Expression with resultType string). + :type headers: object + :param body: Represents the payload that will be sent to the endpoint. + Required for POST/PUT method, not allowed for GET method Type: string (or + Expression with resultType string). + :type body: object + :param authentication: Authentication method used for calling the + endpoint. + :type authentication: + ~azure.mgmt.datafactory.models.WebActivityAuthentication + :param datasets: List of datasets passed to web endpoint. + :type datasets: list[~azure.mgmt.datafactory.models.DatasetReference] + :param linked_services: List of linked services passed to web endpoint. + :type linked_services: + list[~azure.mgmt.datafactory.models.LinkedServiceReference] + """ + + _validation = { + 'name': {'required': True}, + 'type': {'required': True}, + 'method': {'required': True}, + 'url': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'name': {'key': 'name', 'type': 'str'}, + 'description': {'key': 'description', 'type': 'str'}, + 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, + 'method': {'key': 'typeProperties.method', 'type': 'str'}, + 'url': {'key': 'typeProperties.url', 'type': 'object'}, + 'headers': {'key': 'typeProperties.headers', 'type': 'object'}, + 'body': {'key': 'typeProperties.body', 'type': 'object'}, + 'authentication': {'key': 'typeProperties.authentication', 'type': 'WebActivityAuthentication'}, + 'datasets': {'key': 'typeProperties.datasets', 'type': '[DatasetReference]'}, + 'linked_services': {'key': 'typeProperties.linkedServices', 'type': '[LinkedServiceReference]'}, + } + + def __init__(self, *, name: str, method, url, additional_properties=None, description: str=None, depends_on=None, linked_service_name=None, policy=None, headers=None, body=None, authentication=None, datasets=None, linked_services=None, **kwargs) -> None: + super(WebActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, linked_service_name=linked_service_name, policy=policy, **kwargs) + self.method = method + self.url = url + self.headers = headers + self.body = body + self.authentication = authentication + self.datasets = datasets + self.linked_services = linked_services + self.type = 'WebActivity' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_anonymous_authentication.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_anonymous_authentication.py index 4b95d5b488ff..d3bd2f2594ab 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_anonymous_authentication.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_anonymous_authentication.py @@ -16,11 +16,13 @@ class WebAnonymousAuthentication(WebLinkedServiceTypeProperties): """A WebLinkedService that uses anonymous authentication to communicate with an HTTP endpoint. - :param url: The URL of the web service endpoint, e.g. + All required parameters must be populated in order to send to Azure. + + :param url: Required. The URL of the web service endpoint, e.g. http://www.microsoft.com . Type: string (or Expression with resultType string). :type url: object - :param authentication_type: Constant filled by server. + :param authentication_type: Required. Constant filled by server. :type authentication_type: str """ @@ -29,6 +31,11 @@ class WebAnonymousAuthentication(WebLinkedServiceTypeProperties): 'authentication_type': {'required': True}, } - def __init__(self, url): - super(WebAnonymousAuthentication, self).__init__(url=url) + _attribute_map = { + 'url': {'key': 'url', 'type': 'object'}, + 'authentication_type': {'key': 'authenticationType', 'type': 'str'}, + } + + def __init__(self, **kwargs): + super(WebAnonymousAuthentication, self).__init__(**kwargs) self.authentication_type = 'Anonymous' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_anonymous_authentication_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_anonymous_authentication_py3.py new file mode 100644 index 000000000000..c889c410c585 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_anonymous_authentication_py3.py @@ -0,0 +1,41 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .web_linked_service_type_properties import WebLinkedServiceTypeProperties + + +class WebAnonymousAuthentication(WebLinkedServiceTypeProperties): + """A WebLinkedService that uses anonymous authentication to communicate with + an HTTP endpoint. + + All required parameters must be populated in order to send to Azure. + + :param url: Required. The URL of the web service endpoint, e.g. + http://www.microsoft.com . Type: string (or Expression with resultType + string). + :type url: object + :param authentication_type: Required. Constant filled by server. + :type authentication_type: str + """ + + _validation = { + 'url': {'required': True}, + 'authentication_type': {'required': True}, + } + + _attribute_map = { + 'url': {'key': 'url', 'type': 'object'}, + 'authentication_type': {'key': 'authenticationType', 'type': 'str'}, + } + + def __init__(self, *, url, **kwargs) -> None: + super(WebAnonymousAuthentication, self).__init__(url=url, **kwargs) + self.authentication_type = 'Anonymous' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_basic_authentication.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_basic_authentication.py index f32f30499a14..90050f7dae28 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_basic_authentication.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_basic_authentication.py @@ -16,16 +16,18 @@ class WebBasicAuthentication(WebLinkedServiceTypeProperties): """A WebLinkedService that uses basic authentication to communicate with an HTTP endpoint. - :param url: The URL of the web service endpoint, e.g. + All required parameters must be populated in order to send to Azure. + + :param url: Required. The URL of the web service endpoint, e.g. http://www.microsoft.com . Type: string (or Expression with resultType string). :type url: object - :param authentication_type: Constant filled by server. + :param authentication_type: Required. Constant filled by server. :type authentication_type: str - :param username: User name for Basic authentication. Type: string (or - Expression with resultType string). + :param username: Required. User name for Basic authentication. Type: + string (or Expression with resultType string). :type username: object - :param password: The password for Basic authentication. + :param password: Required. The password for Basic authentication. :type password: ~azure.mgmt.datafactory.models.SecretBase """ @@ -43,8 +45,8 @@ class WebBasicAuthentication(WebLinkedServiceTypeProperties): 'password': {'key': 'password', 'type': 'SecretBase'}, } - def __init__(self, url, username, password): - super(WebBasicAuthentication, self).__init__(url=url) - self.username = username - self.password = password + def __init__(self, **kwargs): + super(WebBasicAuthentication, self).__init__(**kwargs) + self.username = kwargs.get('username', None) + self.password = kwargs.get('password', None) self.authentication_type = 'Basic' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_basic_authentication_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_basic_authentication_py3.py new file mode 100644 index 000000000000..a6ba3cc384cf --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_basic_authentication_py3.py @@ -0,0 +1,52 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .web_linked_service_type_properties import WebLinkedServiceTypeProperties + + +class WebBasicAuthentication(WebLinkedServiceTypeProperties): + """A WebLinkedService that uses basic authentication to communicate with an + HTTP endpoint. + + All required parameters must be populated in order to send to Azure. + + :param url: Required. The URL of the web service endpoint, e.g. + http://www.microsoft.com . Type: string (or Expression with resultType + string). + :type url: object + :param authentication_type: Required. Constant filled by server. + :type authentication_type: str + :param username: Required. User name for Basic authentication. Type: + string (or Expression with resultType string). + :type username: object + :param password: Required. The password for Basic authentication. + :type password: ~azure.mgmt.datafactory.models.SecretBase + """ + + _validation = { + 'url': {'required': True}, + 'authentication_type': {'required': True}, + 'username': {'required': True}, + 'password': {'required': True}, + } + + _attribute_map = { + 'url': {'key': 'url', 'type': 'object'}, + 'authentication_type': {'key': 'authenticationType', 'type': 'str'}, + 'username': {'key': 'username', 'type': 'object'}, + 'password': {'key': 'password', 'type': 'SecretBase'}, + } + + def __init__(self, *, url, username, password, **kwargs) -> None: + super(WebBasicAuthentication, self).__init__(url=url, **kwargs) + self.username = username + self.password = password + self.authentication_type = 'Basic' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_client_certificate_authentication.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_client_certificate_authentication.py index 2f4103a772ca..671808ca85d1 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_client_certificate_authentication.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_client_certificate_authentication.py @@ -18,15 +18,17 @@ class WebClientCertificateAuthentication(WebLinkedServiceTypeProperties): authentication; the server must also provide valid credentials to the client. - :param url: The URL of the web service endpoint, e.g. + All required parameters must be populated in order to send to Azure. + + :param url: Required. The URL of the web service endpoint, e.g. http://www.microsoft.com . Type: string (or Expression with resultType string). :type url: object - :param authentication_type: Constant filled by server. + :param authentication_type: Required. Constant filled by server. :type authentication_type: str - :param pfx: Base64-encoded contents of a PFX file. + :param pfx: Required. Base64-encoded contents of a PFX file. :type pfx: ~azure.mgmt.datafactory.models.SecretBase - :param password: Password for the PFX file. + :param password: Required. Password for the PFX file. :type password: ~azure.mgmt.datafactory.models.SecretBase """ @@ -44,8 +46,8 @@ class WebClientCertificateAuthentication(WebLinkedServiceTypeProperties): 'password': {'key': 'password', 'type': 'SecretBase'}, } - def __init__(self, url, pfx, password): - super(WebClientCertificateAuthentication, self).__init__(url=url) - self.pfx = pfx - self.password = password + def __init__(self, **kwargs): + super(WebClientCertificateAuthentication, self).__init__(**kwargs) + self.pfx = kwargs.get('pfx', None) + self.password = kwargs.get('password', None) self.authentication_type = 'ClientCertificate' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_client_certificate_authentication_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_client_certificate_authentication_py3.py new file mode 100644 index 000000000000..16d41490e2c2 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_client_certificate_authentication_py3.py @@ -0,0 +1,53 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .web_linked_service_type_properties import WebLinkedServiceTypeProperties + + +class WebClientCertificateAuthentication(WebLinkedServiceTypeProperties): + """A WebLinkedService that uses client certificate based authentication to + communicate with an HTTP endpoint. This scheme follows mutual + authentication; the server must also provide valid credentials to the + client. + + All required parameters must be populated in order to send to Azure. + + :param url: Required. The URL of the web service endpoint, e.g. + http://www.microsoft.com . Type: string (or Expression with resultType + string). + :type url: object + :param authentication_type: Required. Constant filled by server. + :type authentication_type: str + :param pfx: Required. Base64-encoded contents of a PFX file. + :type pfx: ~azure.mgmt.datafactory.models.SecretBase + :param password: Required. Password for the PFX file. + :type password: ~azure.mgmt.datafactory.models.SecretBase + """ + + _validation = { + 'url': {'required': True}, + 'authentication_type': {'required': True}, + 'pfx': {'required': True}, + 'password': {'required': True}, + } + + _attribute_map = { + 'url': {'key': 'url', 'type': 'object'}, + 'authentication_type': {'key': 'authenticationType', 'type': 'str'}, + 'pfx': {'key': 'pfx', 'type': 'SecretBase'}, + 'password': {'key': 'password', 'type': 'SecretBase'}, + } + + def __init__(self, *, url, pfx, password, **kwargs) -> None: + super(WebClientCertificateAuthentication, self).__init__(url=url, **kwargs) + self.pfx = pfx + self.password = password + self.authentication_type = 'ClientCertificate' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_linked_service.py index a4eb3f8a16a5..cee3bd37409c 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_linked_service.py @@ -15,6 +15,8 @@ class WebLinkedService(LinkedService): """Web linked service. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -29,9 +31,9 @@ class WebLinkedService(LinkedService): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param type_properties: Web linked service properties. + :param type_properties: Required. Web linked service properties. :type type_properties: ~azure.mgmt.datafactory.models.WebLinkedServiceTypeProperties """ @@ -51,7 +53,7 @@ class WebLinkedService(LinkedService): 'type_properties': {'key': 'typeProperties', 'type': 'WebLinkedServiceTypeProperties'}, } - def __init__(self, type_properties, additional_properties=None, connect_via=None, description=None, parameters=None, annotations=None): - super(WebLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations) - self.type_properties = type_properties + def __init__(self, **kwargs): + super(WebLinkedService, self).__init__(**kwargs) + self.type_properties = kwargs.get('type_properties', None) self.type = 'Web' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_linked_service_py3.py new file mode 100644 index 000000000000..0e49f955d436 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_linked_service_py3.py @@ -0,0 +1,59 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class WebLinkedService(LinkedService): + """Web linked service. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param type_properties: Required. Web linked service properties. + :type type_properties: + ~azure.mgmt.datafactory.models.WebLinkedServiceTypeProperties + """ + + _validation = { + 'type': {'required': True}, + 'type_properties': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'type_properties': {'key': 'typeProperties', 'type': 'WebLinkedServiceTypeProperties'}, + } + + def __init__(self, *, type_properties, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, **kwargs) -> None: + super(WebLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.type_properties = type_properties + self.type = 'Web' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_linked_service_type_properties.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_linked_service_type_properties.py index 684401273413..22290e80b19f 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_linked_service_type_properties.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_linked_service_type_properties.py @@ -20,11 +20,13 @@ class WebLinkedServiceTypeProperties(Model): sub-classes are: WebClientCertificateAuthentication, WebBasicAuthentication, WebAnonymousAuthentication - :param url: The URL of the web service endpoint, e.g. + All required parameters must be populated in order to send to Azure. + + :param url: Required. The URL of the web service endpoint, e.g. http://www.microsoft.com . Type: string (or Expression with resultType string). :type url: object - :param authentication_type: Constant filled by server. + :param authentication_type: Required. Constant filled by server. :type authentication_type: str """ @@ -42,7 +44,7 @@ class WebLinkedServiceTypeProperties(Model): 'authentication_type': {'ClientCertificate': 'WebClientCertificateAuthentication', 'Basic': 'WebBasicAuthentication', 'Anonymous': 'WebAnonymousAuthentication'} } - def __init__(self, url): - super(WebLinkedServiceTypeProperties, self).__init__() - self.url = url + def __init__(self, **kwargs): + super(WebLinkedServiceTypeProperties, self).__init__(**kwargs) + self.url = kwargs.get('url', None) self.authentication_type = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_linked_service_type_properties_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_linked_service_type_properties_py3.py new file mode 100644 index 000000000000..1c162c2f1004 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_linked_service_type_properties_py3.py @@ -0,0 +1,50 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class WebLinkedServiceTypeProperties(Model): + """Base definition of WebLinkedServiceTypeProperties, this typeProperties is + polymorphic based on authenticationType, so not flattened in SDK models. + + You probably want to use the sub-classes and not this class directly. Known + sub-classes are: WebClientCertificateAuthentication, + WebBasicAuthentication, WebAnonymousAuthentication + + All required parameters must be populated in order to send to Azure. + + :param url: Required. The URL of the web service endpoint, e.g. + http://www.microsoft.com . Type: string (or Expression with resultType + string). + :type url: object + :param authentication_type: Required. Constant filled by server. + :type authentication_type: str + """ + + _validation = { + 'url': {'required': True}, + 'authentication_type': {'required': True}, + } + + _attribute_map = { + 'url': {'key': 'url', 'type': 'object'}, + 'authentication_type': {'key': 'authenticationType', 'type': 'str'}, + } + + _subtype_map = { + 'authentication_type': {'ClientCertificate': 'WebClientCertificateAuthentication', 'Basic': 'WebBasicAuthentication', 'Anonymous': 'WebAnonymousAuthentication'} + } + + def __init__(self, *, url, **kwargs) -> None: + super(WebLinkedServiceTypeProperties, self).__init__(**kwargs) + self.url = url + self.authentication_type = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_source.py index 69117b9b01aa..13bcbfbb62d7 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_source.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_source.py @@ -15,6 +15,8 @@ class WebSource(CopySource): """A copy activity source for web page table. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -25,7 +27,7 @@ class WebSource(CopySource): with resultType string), pattern: ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type source_retry_wait: object - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str """ @@ -33,6 +35,13 @@ class WebSource(CopySource): 'type': {'required': True}, } - def __init__(self, additional_properties=None, source_retry_count=None, source_retry_wait=None): - super(WebSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait) + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, **kwargs): + super(WebSource, self).__init__(**kwargs) self.type = 'WebSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_source_py3.py new file mode 100644 index 000000000000..97571eb096f8 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_source_py3.py @@ -0,0 +1,47 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class WebSource(CopySource): + """A copy activity source for web page table. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Required. Constant filled by server. + :type type: str + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, *, additional_properties=None, source_retry_count=None, source_retry_wait=None, **kwargs) -> None: + super(WebSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, **kwargs) + self.type = 'WebSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_table_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_table_dataset.py index aba15558471a..671773cfa70f 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_table_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_table_dataset.py @@ -15,6 +15,8 @@ class WebTableDataset(Dataset): """The dataset points to a HTML table in the web page. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -23,7 +25,7 @@ class WebTableDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object - :param linked_service_name: Linked service reference. + :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference :param parameters: Parameters for dataset. @@ -32,10 +34,10 @@ class WebTableDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param index: The zero-based index of the table in the web page. Type: - integer (or Expression with resultType integer), minimum: 0. + :param index: Required. The zero-based index of the table in the web page. + Type: integer (or Expression with resultType integer), minimum: 0. :type index: object :param path: The relative URL to the web page from the linked service URL. Type: string (or Expression with resultType string). @@ -60,8 +62,8 @@ class WebTableDataset(Dataset): 'path': {'key': 'typeProperties.path', 'type': 'object'}, } - def __init__(self, linked_service_name, index, additional_properties=None, description=None, structure=None, parameters=None, annotations=None, path=None): - super(WebTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations) - self.index = index - self.path = path + def __init__(self, **kwargs): + super(WebTableDataset, self).__init__(**kwargs) + self.index = kwargs.get('index', None) + self.path = kwargs.get('path', None) self.type = 'WebTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_table_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_table_dataset_py3.py new file mode 100644 index 000000000000..a5780a7e480a --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_table_dataset_py3.py @@ -0,0 +1,69 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class WebTableDataset(Dataset): + """The dataset points to a HTML table in the web page. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Required. Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param index: Required. The zero-based index of the table in the web page. + Type: integer (or Expression with resultType integer), minimum: 0. + :type index: object + :param path: The relative URL to the web page from the linked service URL. + Type: string (or Expression with resultType string). + :type path: object + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + 'index': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'index': {'key': 'typeProperties.index', 'type': 'object'}, + 'path': {'key': 'typeProperties.path', 'type': 'object'}, + } + + def __init__(self, *, linked_service_name, index, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, path=None, **kwargs) -> None: + super(WebTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + self.index = index + self.path = path + self.type = 'WebTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_linked_service.py index 0451a3b175f7..84129d6e38d1 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_linked_service.py @@ -15,6 +15,8 @@ class XeroLinkedService(LinkedService): """Xero Serivce linked service. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -29,9 +31,10 @@ class XeroLinkedService(LinkedService): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param host: The endpoint of the Xero server. (i.e. api.xero.com) + :param host: Required. The endpoint of the Xero server. (i.e. + api.xero.com) :type host: object :param consumer_key: The consumer key associated with the Xero application. @@ -78,13 +81,13 @@ class XeroLinkedService(LinkedService): 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, host, additional_properties=None, connect_via=None, description=None, parameters=None, annotations=None, consumer_key=None, private_key=None, use_encrypted_endpoints=None, use_host_verification=None, use_peer_verification=None, encrypted_credential=None): - super(XeroLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations) - self.host = host - self.consumer_key = consumer_key - self.private_key = private_key - self.use_encrypted_endpoints = use_encrypted_endpoints - self.use_host_verification = use_host_verification - self.use_peer_verification = use_peer_verification - self.encrypted_credential = encrypted_credential + def __init__(self, **kwargs): + super(XeroLinkedService, self).__init__(**kwargs) + self.host = kwargs.get('host', None) + self.consumer_key = kwargs.get('consumer_key', None) + self.private_key = kwargs.get('private_key', None) + self.use_encrypted_endpoints = kwargs.get('use_encrypted_endpoints', None) + self.use_host_verification = kwargs.get('use_host_verification', None) + self.use_peer_verification = kwargs.get('use_peer_verification', None) + self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'Xero' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_linked_service_py3.py new file mode 100644 index 000000000000..104af987bddc --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_linked_service_py3.py @@ -0,0 +1,93 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class XeroLinkedService(LinkedService): + """Xero Serivce linked service. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param host: Required. The endpoint of the Xero server. (i.e. + api.xero.com) + :type host: object + :param consumer_key: The consumer key associated with the Xero + application. + :type consumer_key: ~azure.mgmt.datafactory.models.SecretBase + :param private_key: The private key from the .pem file that was generated + for your Xero private application. You must include all the text from the + .pem file, including the Unix line endings( + ). + :type private_key: ~azure.mgmt.datafactory.models.SecretBase + :param use_encrypted_endpoints: Specifies whether the data source + endpoints are encrypted using HTTPS. The default value is true. + :type use_encrypted_endpoints: object + :param use_host_verification: Specifies whether to require the host name + in the server's certificate to match the host name of the server when + connecting over SSL. The default value is true. + :type use_host_verification: object + :param use_peer_verification: Specifies whether to verify the identity of + the server when connecting over SSL. The default value is true. + :type use_peer_verification: object + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'host': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'host': {'key': 'typeProperties.host', 'type': 'object'}, + 'consumer_key': {'key': 'typeProperties.consumerKey', 'type': 'SecretBase'}, + 'private_key': {'key': 'typeProperties.privateKey', 'type': 'SecretBase'}, + 'use_encrypted_endpoints': {'key': 'typeProperties.useEncryptedEndpoints', 'type': 'object'}, + 'use_host_verification': {'key': 'typeProperties.useHostVerification', 'type': 'object'}, + 'use_peer_verification': {'key': 'typeProperties.usePeerVerification', 'type': 'object'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, *, host, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, consumer_key=None, private_key=None, use_encrypted_endpoints=None, use_host_verification=None, use_peer_verification=None, encrypted_credential=None, **kwargs) -> None: + super(XeroLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.host = host + self.consumer_key = consumer_key + self.private_key = private_key + self.use_encrypted_endpoints = use_encrypted_endpoints + self.use_host_verification = use_host_verification + self.use_peer_verification = use_peer_verification + self.encrypted_credential = encrypted_credential + self.type = 'Xero' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_object_dataset.py index 128e7e34e5b0..6ef72bc09878 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_object_dataset.py @@ -15,6 +15,8 @@ class XeroObjectDataset(Dataset): """Xero Serivce dataset. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -23,7 +25,7 @@ class XeroObjectDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object - :param linked_service_name: Linked service reference. + :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference :param parameters: Parameters for dataset. @@ -32,7 +34,7 @@ class XeroObjectDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str """ @@ -41,6 +43,16 @@ class XeroObjectDataset(Dataset): 'type': {'required': True}, } - def __init__(self, linked_service_name, additional_properties=None, description=None, structure=None, parameters=None, annotations=None): - super(XeroObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations) + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, **kwargs): + super(XeroObjectDataset, self).__init__(**kwargs) self.type = 'XeroObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_object_dataset_py3.py new file mode 100644 index 000000000000..80a74d953589 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_object_dataset_py3.py @@ -0,0 +1,58 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class XeroObjectDataset(Dataset): + """Xero Serivce dataset. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Required. Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: + super(XeroObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + self.type = 'XeroObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_source.py index e5b2e2e2bca1..c9780ecf4d67 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_source.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_source.py @@ -15,6 +15,8 @@ class XeroSource(CopySource): """A copy activity Xero Serivce source. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -25,7 +27,7 @@ class XeroSource(CopySource): with resultType string), pattern: ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type source_retry_wait: object - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). @@ -44,7 +46,7 @@ class XeroSource(CopySource): 'query': {'key': 'query', 'type': 'object'}, } - def __init__(self, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None): - super(XeroSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait) - self.query = query + def __init__(self, **kwargs): + super(XeroSource, self).__init__(**kwargs) + self.query = kwargs.get('query', None) self.type = 'XeroSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_source_py3.py new file mode 100644 index 000000000000..4a1c2d2bdbc3 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_source_py3.py @@ -0,0 +1,52 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class XeroSource(CopySource): + """A copy activity Xero Serivce source. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Required. Constant filled by server. + :type type: str + :param query: A query to retrieve data from source. Type: string (or + Expression with resultType string). + :type query: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'query': {'key': 'query', 'type': 'object'}, + } + + def __init__(self, *, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None, **kwargs) -> None: + super(XeroSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, **kwargs) + self.query = query + self.type = 'XeroSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/zoho_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/zoho_linked_service.py index 56a265798762..997efb5fc242 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/zoho_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/zoho_linked_service.py @@ -15,6 +15,8 @@ class ZohoLinkedService(LinkedService): """Zoho server linked service. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -29,9 +31,9 @@ class ZohoLinkedService(LinkedService): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str - :param endpoint: The endpoint of the Zoho server. (i.e. + :param endpoint: Required. The endpoint of the Zoho server. (i.e. crm.zoho.com/crm/private) :type endpoint: object :param access_token: The access token for Zoho authentication. @@ -72,12 +74,12 @@ class ZohoLinkedService(LinkedService): 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, endpoint, additional_properties=None, connect_via=None, description=None, parameters=None, annotations=None, access_token=None, use_encrypted_endpoints=None, use_host_verification=None, use_peer_verification=None, encrypted_credential=None): - super(ZohoLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations) - self.endpoint = endpoint - self.access_token = access_token - self.use_encrypted_endpoints = use_encrypted_endpoints - self.use_host_verification = use_host_verification - self.use_peer_verification = use_peer_verification - self.encrypted_credential = encrypted_credential + def __init__(self, **kwargs): + super(ZohoLinkedService, self).__init__(**kwargs) + self.endpoint = kwargs.get('endpoint', None) + self.access_token = kwargs.get('access_token', None) + self.use_encrypted_endpoints = kwargs.get('use_encrypted_endpoints', None) + self.use_host_verification = kwargs.get('use_host_verification', None) + self.use_peer_verification = kwargs.get('use_peer_verification', None) + self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'Zoho' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/zoho_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/zoho_linked_service_py3.py new file mode 100644 index 000000000000..4e5d0f5f31f3 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/zoho_linked_service_py3.py @@ -0,0 +1,85 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class ZohoLinkedService(LinkedService): + """Zoho server linked service. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param endpoint: Required. The endpoint of the Zoho server. (i.e. + crm.zoho.com/crm/private) + :type endpoint: object + :param access_token: The access token for Zoho authentication. + :type access_token: ~azure.mgmt.datafactory.models.SecretBase + :param use_encrypted_endpoints: Specifies whether the data source + endpoints are encrypted using HTTPS. The default value is true. + :type use_encrypted_endpoints: object + :param use_host_verification: Specifies whether to require the host name + in the server's certificate to match the host name of the server when + connecting over SSL. The default value is true. + :type use_host_verification: object + :param use_peer_verification: Specifies whether to verify the identity of + the server when connecting over SSL. The default value is true. + :type use_peer_verification: object + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'endpoint': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'endpoint': {'key': 'typeProperties.endpoint', 'type': 'object'}, + 'access_token': {'key': 'typeProperties.accessToken', 'type': 'SecretBase'}, + 'use_encrypted_endpoints': {'key': 'typeProperties.useEncryptedEndpoints', 'type': 'object'}, + 'use_host_verification': {'key': 'typeProperties.useHostVerification', 'type': 'object'}, + 'use_peer_verification': {'key': 'typeProperties.usePeerVerification', 'type': 'object'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, *, endpoint, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, access_token=None, use_encrypted_endpoints=None, use_host_verification=None, use_peer_verification=None, encrypted_credential=None, **kwargs) -> None: + super(ZohoLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.endpoint = endpoint + self.access_token = access_token + self.use_encrypted_endpoints = use_encrypted_endpoints + self.use_host_verification = use_host_verification + self.use_peer_verification = use_peer_verification + self.encrypted_credential = encrypted_credential + self.type = 'Zoho' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/zoho_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/zoho_object_dataset.py index b742050b6948..26ab79ab4301 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/zoho_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/zoho_object_dataset.py @@ -15,6 +15,8 @@ class ZohoObjectDataset(Dataset): """Zoho server dataset. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -23,7 +25,7 @@ class ZohoObjectDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object - :param linked_service_name: Linked service reference. + :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference :param parameters: Parameters for dataset. @@ -32,7 +34,7 @@ class ZohoObjectDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str """ @@ -41,6 +43,16 @@ class ZohoObjectDataset(Dataset): 'type': {'required': True}, } - def __init__(self, linked_service_name, additional_properties=None, description=None, structure=None, parameters=None, annotations=None): - super(ZohoObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations) + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, **kwargs): + super(ZohoObjectDataset, self).__init__(**kwargs) self.type = 'ZohoObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/zoho_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/zoho_object_dataset_py3.py new file mode 100644 index 000000000000..112d3bbd3952 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/zoho_object_dataset_py3.py @@ -0,0 +1,58 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class ZohoObjectDataset(Dataset): + """Zoho server dataset. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Required. Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: + super(ZohoObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + self.type = 'ZohoObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/zoho_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/zoho_source.py index 058c10ee5f59..248d50d55297 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/zoho_source.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/zoho_source.py @@ -15,6 +15,8 @@ class ZohoSource(CopySource): """A copy activity Zoho server source. + All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -25,7 +27,7 @@ class ZohoSource(CopySource): with resultType string), pattern: ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type source_retry_wait: object - :param type: Constant filled by server. + :param type: Required. Constant filled by server. :type type: str :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). @@ -44,7 +46,7 @@ class ZohoSource(CopySource): 'query': {'key': 'query', 'type': 'object'}, } - def __init__(self, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None): - super(ZohoSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait) - self.query = query + def __init__(self, **kwargs): + super(ZohoSource, self).__init__(**kwargs) + self.query = kwargs.get('query', None) self.type = 'ZohoSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/zoho_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/zoho_source_py3.py new file mode 100644 index 000000000000..8533b239ae13 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/zoho_source_py3.py @@ -0,0 +1,52 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class ZohoSource(CopySource): + """A copy activity Zoho server source. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Required. Constant filled by server. + :type type: str + :param query: A query to retrieve data from source. Type: string (or + Expression with resultType string). + :type query: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'query': {'key': 'query', 'type': 'object'}, + } + + def __init__(self, *, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None, **kwargs) -> None: + super(ZohoSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, **kwargs) + self.query = query + self.type = 'ZohoSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/integration_runtimes_operations.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/integration_runtimes_operations.py index 3cc1ef61e744..6790452bfc6f 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/integration_runtimes_operations.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/integration_runtimes_operations.py @@ -11,8 +11,8 @@ import uuid from msrest.pipeline import ClientRawResponse -from msrest.exceptions import DeserializationError -from msrestazure.azure_operation import AzureOperationPoller +from msrest.polling import LROPoller, NoPolling +from msrestazure.polling.arm_polling import ARMPolling from .. import models @@ -705,7 +705,7 @@ def _start_initial( return deserialized def start( - self, resource_group_name, factory_name, integration_runtime_name, custom_headers=None, raw=False, **operation_config): + self, resource_group_name, factory_name, integration_runtime_name, custom_headers=None, raw=False, polling=True, **operation_config): """Starts a ManagedReserved type integration runtime. :param resource_group_name: The resource group name. @@ -715,13 +715,17 @@ def start( :param integration_runtime_name: The integration runtime name. :type integration_runtime_name: str :param dict custom_headers: headers that will be added to the request - :param bool raw: returns the direct response alongside the - deserialized response - :return: An instance of AzureOperationPoller that returns - IntegrationRuntimeStatusResponse or ClientRawResponse if raw=true + :param bool raw: The poller return type is ClientRawResponse, the + direct response alongside the deserialized response + :param polling: True for ARMPolling, False for no polling, or a + polling object for personal polling strategy + :return: An instance of LROPoller that returns + IntegrationRuntimeStatusResponse or + ClientRawResponse if raw==True :rtype: ~msrestazure.azure_operation.AzureOperationPoller[~azure.mgmt.datafactory.models.IntegrationRuntimeStatusResponse] - or ~msrest.pipeline.ClientRawResponse + or + ~msrestazure.azure_operation.AzureOperationPoller[~msrest.pipeline.ClientRawResponse[~azure.mgmt.datafactory.models.IntegrationRuntimeStatusResponse]] :raises: :class:`ErrorResponseException` """ @@ -733,28 +737,8 @@ def start( raw=True, **operation_config ) - if raw: - return raw_result - - # Construct and send request - def long_running_send(): - return raw_result.response - - def get_long_running_status(status_link, headers=None): - - request = self._client.get(status_link) - if headers: - request.headers.update(headers) - header_parameters = {} - header_parameters['x-ms-client-request-id'] = raw_result.response.request.headers['x-ms-client-request-id'] - return self._client.send( - request, header_parameters, stream=False, **operation_config) def get_long_running_output(response): - - if response.status_code not in [200, 202]: - raise models.ErrorResponseException(self._deserialize, response) - deserialized = self._deserialize('IntegrationRuntimeStatusResponse', response) if raw: @@ -763,12 +747,13 @@ def get_long_running_output(response): return deserialized - long_running_operation_timeout = operation_config.get( + lro_delay = operation_config.get( 'long_running_operation_timeout', self.config.long_running_operation_timeout) - return AzureOperationPoller( - long_running_send, get_long_running_output, - get_long_running_status, long_running_operation_timeout) + if polling is True: polling_method = ARMPolling(lro_delay, **operation_config) + elif polling is False: polling_method = NoPolling() + else: polling_method = polling + return LROPoller(self._client, raw_result, get_long_running_output, polling_method) start.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/integrationRuntimes/{integrationRuntimeName}/start'} @@ -810,7 +795,7 @@ def _stop_initial( return client_raw_response def stop( - self, resource_group_name, factory_name, integration_runtime_name, custom_headers=None, raw=False, **operation_config): + self, resource_group_name, factory_name, integration_runtime_name, custom_headers=None, raw=False, polling=True, **operation_config): """Stops a ManagedReserved type integration runtime. :param resource_group_name: The resource group name. @@ -820,12 +805,14 @@ def stop( :param integration_runtime_name: The integration runtime name. :type integration_runtime_name: str :param dict custom_headers: headers that will be added to the request - :param bool raw: returns the direct response alongside the - deserialized response - :return: An instance of AzureOperationPoller that returns None or - ClientRawResponse if raw=true + :param bool raw: The poller return type is ClientRawResponse, the + direct response alongside the deserialized response + :param polling: True for ARMPolling, False for no polling, or a + polling object for personal polling strategy + :return: An instance of LROPoller that returns None or + ClientRawResponse if raw==True :rtype: ~msrestazure.azure_operation.AzureOperationPoller[None] or - ~msrest.pipeline.ClientRawResponse + ~msrestazure.azure_operation.AzureOperationPoller[~msrest.pipeline.ClientRawResponse[None]] :raises: :class:`ErrorResponseException` """ @@ -837,38 +824,19 @@ def stop( raw=True, **operation_config ) - if raw: - return raw_result - - # Construct and send request - def long_running_send(): - return raw_result.response - - def get_long_running_status(status_link, headers=None): - - request = self._client.get(status_link) - if headers: - request.headers.update(headers) - header_parameters = {} - header_parameters['x-ms-client-request-id'] = raw_result.response.request.headers['x-ms-client-request-id'] - return self._client.send( - request, header_parameters, stream=False, **operation_config) def get_long_running_output(response): - - if response.status_code not in [200, 202]: - raise models.ErrorResponseException(self._deserialize, response) - if raw: client_raw_response = ClientRawResponse(None, response) return client_raw_response - long_running_operation_timeout = operation_config.get( + lro_delay = operation_config.get( 'long_running_operation_timeout', self.config.long_running_operation_timeout) - return AzureOperationPoller( - long_running_send, get_long_running_output, - get_long_running_status, long_running_operation_timeout) + if polling is True: polling_method = ARMPolling(lro_delay, **operation_config) + elif polling is False: polling_method = NoPolling() + else: polling_method = polling + return LROPoller(self._client, raw_result, get_long_running_output, polling_method) stop.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/integrationRuntimes/{integrationRuntimeName}/stop'} def remove_node( diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/triggers_operations.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/triggers_operations.py index 60482647dd80..e54de00202ea 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/triggers_operations.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/triggers_operations.py @@ -11,8 +11,8 @@ import uuid from msrest.pipeline import ClientRawResponse -from msrest.exceptions import DeserializationError -from msrestazure.azure_operation import AzureOperationPoller +from msrest.polling import LROPoller, NoPolling +from msrestazure.polling.arm_polling import ARMPolling from .. import models @@ -345,7 +345,7 @@ def _start_initial( return client_raw_response def start( - self, resource_group_name, factory_name, trigger_name, custom_headers=None, raw=False, **operation_config): + self, resource_group_name, factory_name, trigger_name, custom_headers=None, raw=False, polling=True, **operation_config): """Starts a trigger. :param resource_group_name: The resource group name. @@ -355,12 +355,14 @@ def start( :param trigger_name: The trigger name. :type trigger_name: str :param dict custom_headers: headers that will be added to the request - :param bool raw: returns the direct response alongside the - deserialized response - :return: An instance of AzureOperationPoller that returns None or - ClientRawResponse if raw=true + :param bool raw: The poller return type is ClientRawResponse, the + direct response alongside the deserialized response + :param polling: True for ARMPolling, False for no polling, or a + polling object for personal polling strategy + :return: An instance of LROPoller that returns None or + ClientRawResponse if raw==True :rtype: ~msrestazure.azure_operation.AzureOperationPoller[None] or - ~msrest.pipeline.ClientRawResponse + ~msrestazure.azure_operation.AzureOperationPoller[~msrest.pipeline.ClientRawResponse[None]] :raises: :class:`ErrorResponseException` """ @@ -372,38 +374,19 @@ def start( raw=True, **operation_config ) - if raw: - return raw_result - - # Construct and send request - def long_running_send(): - return raw_result.response - - def get_long_running_status(status_link, headers=None): - - request = self._client.get(status_link) - if headers: - request.headers.update(headers) - header_parameters = {} - header_parameters['x-ms-client-request-id'] = raw_result.response.request.headers['x-ms-client-request-id'] - return self._client.send( - request, header_parameters, stream=False, **operation_config) def get_long_running_output(response): - - if response.status_code not in [200]: - raise models.ErrorResponseException(self._deserialize, response) - if raw: client_raw_response = ClientRawResponse(None, response) return client_raw_response - long_running_operation_timeout = operation_config.get( + lro_delay = operation_config.get( 'long_running_operation_timeout', self.config.long_running_operation_timeout) - return AzureOperationPoller( - long_running_send, get_long_running_output, - get_long_running_status, long_running_operation_timeout) + if polling is True: polling_method = ARMPolling(lro_delay, **operation_config) + elif polling is False: polling_method = NoPolling() + else: polling_method = polling + return LROPoller(self._client, raw_result, get_long_running_output, polling_method) start.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/triggers/{triggerName}/start'} @@ -445,7 +428,7 @@ def _stop_initial( return client_raw_response def stop( - self, resource_group_name, factory_name, trigger_name, custom_headers=None, raw=False, **operation_config): + self, resource_group_name, factory_name, trigger_name, custom_headers=None, raw=False, polling=True, **operation_config): """Stops a trigger. :param resource_group_name: The resource group name. @@ -455,12 +438,14 @@ def stop( :param trigger_name: The trigger name. :type trigger_name: str :param dict custom_headers: headers that will be added to the request - :param bool raw: returns the direct response alongside the - deserialized response - :return: An instance of AzureOperationPoller that returns None or - ClientRawResponse if raw=true + :param bool raw: The poller return type is ClientRawResponse, the + direct response alongside the deserialized response + :param polling: True for ARMPolling, False for no polling, or a + polling object for personal polling strategy + :return: An instance of LROPoller that returns None or + ClientRawResponse if raw==True :rtype: ~msrestazure.azure_operation.AzureOperationPoller[None] or - ~msrest.pipeline.ClientRawResponse + ~msrestazure.azure_operation.AzureOperationPoller[~msrest.pipeline.ClientRawResponse[None]] :raises: :class:`ErrorResponseException` """ @@ -472,38 +457,19 @@ def stop( raw=True, **operation_config ) - if raw: - return raw_result - - # Construct and send request - def long_running_send(): - return raw_result.response - - def get_long_running_status(status_link, headers=None): - - request = self._client.get(status_link) - if headers: - request.headers.update(headers) - header_parameters = {} - header_parameters['x-ms-client-request-id'] = raw_result.response.request.headers['x-ms-client-request-id'] - return self._client.send( - request, header_parameters, stream=False, **operation_config) def get_long_running_output(response): - - if response.status_code not in [200]: - raise models.ErrorResponseException(self._deserialize, response) - if raw: client_raw_response = ClientRawResponse(None, response) return client_raw_response - long_running_operation_timeout = operation_config.get( + lro_delay = operation_config.get( 'long_running_operation_timeout', self.config.long_running_operation_timeout) - return AzureOperationPoller( - long_running_send, get_long_running_output, - get_long_running_status, long_running_operation_timeout) + if polling is True: polling_method = ARMPolling(lro_delay, **operation_config) + elif polling is False: polling_method = NoPolling() + else: polling_method = polling + return LROPoller(self._client, raw_result, get_long_running_output, polling_method) stop.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/triggers/{triggerName}/stop'} def list_runs( From cdd265d69bfeed236abb3e1c0358ad1b85776aeb Mon Sep 17 00:00:00 2001 From: Azure SDK for Python bot Date: Fri, 6 Apr 2018 14:17:59 -0700 Subject: [PATCH 02/54] Generated from bc92a82bf90045bafefbf1a584aad824a4748b47 (#2359) update description 2 --- .../azure/mgmt/datafactory/models/__init__.py | 15 ++- .../mgmt/datafactory/models/copy_source.py | 27 +++--- .../datafactory/models/copy_source_py3.py | 27 +++--- .../azure/mgmt/datafactory/models/dataset.py | 27 +++--- .../mgmt/datafactory/models/dataset_py3.py | 27 +++--- .../models/execute_ssis_package_activity.py | 33 ++++++- .../execute_ssis_package_activity_py3.py | 35 ++++++- .../mgmt/datafactory/models/linked_service.py | 4 +- .../datafactory/models/linked_service_py3.py | 4 +- .../models/responsys_linked_service.py | 94 ------------------- .../models/responsys_linked_service_py3.py | 94 ------------------- .../models/responsys_object_dataset.py | 58 ------------ .../models/responsys_object_dataset_py3.py | 58 ------------ .../datafactory/models/responsys_source.py | 52 ---------- .../models/responsys_source_py3.py | 52 ---------- .../models/ssis_execution_parameter.py | 35 +++++++ .../models/ssis_execution_parameter_py3.py | 35 +++++++ .../models/ssis_property_override.py | 40 ++++++++ .../models/ssis_property_override_py3.py | 40 ++++++++ .../azure/mgmt/datafactory/version.py | 2 +- 20 files changed, 276 insertions(+), 483 deletions(-) delete mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_linked_service.py delete mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_linked_service_py3.py delete mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_object_dataset.py delete mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_object_dataset_py3.py delete mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_source.py delete mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_source_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/ssis_execution_parameter.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/ssis_execution_parameter_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/ssis_property_override.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/ssis_property_override_py3.py diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py index 23607a9e47f1..f25f49f6cb4a 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py @@ -58,7 +58,6 @@ from .operation_service_specification_py3 import OperationServiceSpecification from .operation_py3 import Operation from .operation_list_response_py3 import OperationListResponse - from .responsys_linked_service_py3 import ResponsysLinkedService from .azure_databricks_linked_service_py3 import AzureDatabricksLinkedService from .azure_data_lake_analytics_linked_service_py3 import AzureDataLakeAnalyticsLinkedService from .hd_insight_on_demand_linked_service_py3 import HDInsightOnDemandLinkedService @@ -132,7 +131,6 @@ from .sql_server_linked_service_py3 import SqlServerLinkedService from .azure_sql_dw_linked_service_py3 import AzureSqlDWLinkedService from .azure_storage_linked_service_py3 import AzureStorageLinkedService - from .responsys_object_dataset_py3 import ResponsysObjectDataset from .salesforce_marketing_cloud_object_dataset_py3 import SalesforceMarketingCloudObjectDataset from .vertica_table_dataset_py3 import VerticaTableDataset from .netezza_table_dataset_py3 import NetezzaTableDataset @@ -215,7 +213,6 @@ from .web_activity_py3 import WebActivity from .redshift_unload_settings_py3 import RedshiftUnloadSettings from .amazon_redshift_source_py3 import AmazonRedshiftSource - from .responsys_source_py3 import ResponsysSource from .salesforce_marketing_cloud_source_py3 import SalesforceMarketingCloudSource from .vertica_source_py3 import VerticaSource from .netezza_source_py3 import NetezzaSource @@ -271,6 +268,8 @@ from .sql_server_stored_procedure_activity_py3 import SqlServerStoredProcedureActivity from .custom_activity_reference_object_py3 import CustomActivityReferenceObject from .custom_activity_py3 import CustomActivity + from .ssis_property_override_py3 import SSISPropertyOverride + from .ssis_execution_parameter_py3 import SSISExecutionParameter from .ssis_package_location_py3 import SSISPackageLocation from .execute_ssis_package_activity_py3 import ExecuteSSISPackageActivity from .hd_insight_spark_activity_py3 import HDInsightSparkActivity @@ -380,7 +379,6 @@ from .operation_service_specification import OperationServiceSpecification from .operation import Operation from .operation_list_response import OperationListResponse - from .responsys_linked_service import ResponsysLinkedService from .azure_databricks_linked_service import AzureDatabricksLinkedService from .azure_data_lake_analytics_linked_service import AzureDataLakeAnalyticsLinkedService from .hd_insight_on_demand_linked_service import HDInsightOnDemandLinkedService @@ -454,7 +452,6 @@ from .sql_server_linked_service import SqlServerLinkedService from .azure_sql_dw_linked_service import AzureSqlDWLinkedService from .azure_storage_linked_service import AzureStorageLinkedService - from .responsys_object_dataset import ResponsysObjectDataset from .salesforce_marketing_cloud_object_dataset import SalesforceMarketingCloudObjectDataset from .vertica_table_dataset import VerticaTableDataset from .netezza_table_dataset import NetezzaTableDataset @@ -537,7 +534,6 @@ from .web_activity import WebActivity from .redshift_unload_settings import RedshiftUnloadSettings from .amazon_redshift_source import AmazonRedshiftSource - from .responsys_source import ResponsysSource from .salesforce_marketing_cloud_source import SalesforceMarketingCloudSource from .vertica_source import VerticaSource from .netezza_source import NetezzaSource @@ -593,6 +589,8 @@ from .sql_server_stored_procedure_activity import SqlServerStoredProcedureActivity from .custom_activity_reference_object import CustomActivityReferenceObject from .custom_activity import CustomActivity + from .ssis_property_override import SSISPropertyOverride + from .ssis_execution_parameter import SSISExecutionParameter from .ssis_package_location import SSISPackageLocation from .execute_ssis_package_activity import ExecuteSSISPackageActivity from .hd_insight_spark_activity import HDInsightSparkActivity @@ -770,7 +768,6 @@ 'OperationServiceSpecification', 'Operation', 'OperationListResponse', - 'ResponsysLinkedService', 'AzureDatabricksLinkedService', 'AzureDataLakeAnalyticsLinkedService', 'HDInsightOnDemandLinkedService', @@ -844,7 +841,6 @@ 'SqlServerLinkedService', 'AzureSqlDWLinkedService', 'AzureStorageLinkedService', - 'ResponsysObjectDataset', 'SalesforceMarketingCloudObjectDataset', 'VerticaTableDataset', 'NetezzaTableDataset', @@ -927,7 +923,6 @@ 'WebActivity', 'RedshiftUnloadSettings', 'AmazonRedshiftSource', - 'ResponsysSource', 'SalesforceMarketingCloudSource', 'VerticaSource', 'NetezzaSource', @@ -983,6 +978,8 @@ 'SqlServerStoredProcedureActivity', 'CustomActivityReferenceObject', 'CustomActivity', + 'SSISPropertyOverride', + 'SSISExecutionParameter', 'SSISPackageLocation', 'ExecuteSSISPackageActivity', 'HDInsightSparkActivity', diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_source.py index 9a11107fc8e8..e7c55fbe9e9c 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_source.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_source.py @@ -16,19 +16,18 @@ class CopySource(Model): """A copy activity source. You probably want to use the sub-classes and not this class directly. Known - sub-classes are: AmazonRedshiftSource, ResponsysSource, - SalesforceMarketingCloudSource, VerticaSource, NetezzaSource, ZohoSource, - XeroSource, SquareSource, SparkSource, ShopifySource, ServiceNowSource, - QuickBooksSource, PrestoSource, PhoenixSource, PaypalSource, MarketoSource, - MariaDBSource, MagentoSource, JiraSource, ImpalaSource, HubspotSource, - HiveSource, HBaseSource, GreenplumSource, GoogleBigQuerySource, - EloquaSource, DrillSource, CouchbaseSource, ConcurSource, - AzurePostgreSqlSource, AmazonMWSSource, HttpSource, - AzureDataLakeStoreSource, MongoDbSource, CassandraSource, WebSource, - OracleSource, AzureMySqlSource, HdfsSource, FileSystemSource, SqlDWSource, - SqlSource, SapEccSource, SapCloudForCustomerSource, SalesforceSource, - RelationalSource, DynamicsSource, DocumentDbCollectionSource, BlobSource, - AzureTableSource + sub-classes are: AmazonRedshiftSource, SalesforceMarketingCloudSource, + VerticaSource, NetezzaSource, ZohoSource, XeroSource, SquareSource, + SparkSource, ShopifySource, ServiceNowSource, QuickBooksSource, + PrestoSource, PhoenixSource, PaypalSource, MarketoSource, MariaDBSource, + MagentoSource, JiraSource, ImpalaSource, HubspotSource, HiveSource, + HBaseSource, GreenplumSource, GoogleBigQuerySource, EloquaSource, + DrillSource, CouchbaseSource, ConcurSource, AzurePostgreSqlSource, + AmazonMWSSource, HttpSource, AzureDataLakeStoreSource, MongoDbSource, + CassandraSource, WebSource, OracleSource, AzureMySqlSource, HdfsSource, + FileSystemSource, SqlDWSource, SqlSource, SapEccSource, + SapCloudForCustomerSource, SalesforceSource, RelationalSource, + DynamicsSource, DocumentDbCollectionSource, BlobSource, AzureTableSource All required parameters must be populated in order to send to Azure. @@ -58,7 +57,7 @@ class CopySource(Model): } _subtype_map = { - 'type': {'AmazonRedshiftSource': 'AmazonRedshiftSource', 'ResponsysSource': 'ResponsysSource', 'SalesforceMarketingCloudSource': 'SalesforceMarketingCloudSource', 'VerticaSource': 'VerticaSource', 'NetezzaSource': 'NetezzaSource', 'ZohoSource': 'ZohoSource', 'XeroSource': 'XeroSource', 'SquareSource': 'SquareSource', 'SparkSource': 'SparkSource', 'ShopifySource': 'ShopifySource', 'ServiceNowSource': 'ServiceNowSource', 'QuickBooksSource': 'QuickBooksSource', 'PrestoSource': 'PrestoSource', 'PhoenixSource': 'PhoenixSource', 'PaypalSource': 'PaypalSource', 'MarketoSource': 'MarketoSource', 'MariaDBSource': 'MariaDBSource', 'MagentoSource': 'MagentoSource', 'JiraSource': 'JiraSource', 'ImpalaSource': 'ImpalaSource', 'HubspotSource': 'HubspotSource', 'HiveSource': 'HiveSource', 'HBaseSource': 'HBaseSource', 'GreenplumSource': 'GreenplumSource', 'GoogleBigQuerySource': 'GoogleBigQuerySource', 'EloquaSource': 'EloquaSource', 'DrillSource': 'DrillSource', 'CouchbaseSource': 'CouchbaseSource', 'ConcurSource': 'ConcurSource', 'AzurePostgreSqlSource': 'AzurePostgreSqlSource', 'AmazonMWSSource': 'AmazonMWSSource', 'HttpSource': 'HttpSource', 'AzureDataLakeStoreSource': 'AzureDataLakeStoreSource', 'MongoDbSource': 'MongoDbSource', 'CassandraSource': 'CassandraSource', 'WebSource': 'WebSource', 'OracleSource': 'OracleSource', 'AzureMySqlSource': 'AzureMySqlSource', 'HdfsSource': 'HdfsSource', 'FileSystemSource': 'FileSystemSource', 'SqlDWSource': 'SqlDWSource', 'SqlSource': 'SqlSource', 'SapEccSource': 'SapEccSource', 'SapCloudForCustomerSource': 'SapCloudForCustomerSource', 'SalesforceSource': 'SalesforceSource', 'RelationalSource': 'RelationalSource', 'DynamicsSource': 'DynamicsSource', 'DocumentDbCollectionSource': 'DocumentDbCollectionSource', 'BlobSource': 'BlobSource', 'AzureTableSource': 'AzureTableSource'} + 'type': {'AmazonRedshiftSource': 'AmazonRedshiftSource', 'SalesforceMarketingCloudSource': 'SalesforceMarketingCloudSource', 'VerticaSource': 'VerticaSource', 'NetezzaSource': 'NetezzaSource', 'ZohoSource': 'ZohoSource', 'XeroSource': 'XeroSource', 'SquareSource': 'SquareSource', 'SparkSource': 'SparkSource', 'ShopifySource': 'ShopifySource', 'ServiceNowSource': 'ServiceNowSource', 'QuickBooksSource': 'QuickBooksSource', 'PrestoSource': 'PrestoSource', 'PhoenixSource': 'PhoenixSource', 'PaypalSource': 'PaypalSource', 'MarketoSource': 'MarketoSource', 'MariaDBSource': 'MariaDBSource', 'MagentoSource': 'MagentoSource', 'JiraSource': 'JiraSource', 'ImpalaSource': 'ImpalaSource', 'HubspotSource': 'HubspotSource', 'HiveSource': 'HiveSource', 'HBaseSource': 'HBaseSource', 'GreenplumSource': 'GreenplumSource', 'GoogleBigQuerySource': 'GoogleBigQuerySource', 'EloquaSource': 'EloquaSource', 'DrillSource': 'DrillSource', 'CouchbaseSource': 'CouchbaseSource', 'ConcurSource': 'ConcurSource', 'AzurePostgreSqlSource': 'AzurePostgreSqlSource', 'AmazonMWSSource': 'AmazonMWSSource', 'HttpSource': 'HttpSource', 'AzureDataLakeStoreSource': 'AzureDataLakeStoreSource', 'MongoDbSource': 'MongoDbSource', 'CassandraSource': 'CassandraSource', 'WebSource': 'WebSource', 'OracleSource': 'OracleSource', 'AzureMySqlSource': 'AzureMySqlSource', 'HdfsSource': 'HdfsSource', 'FileSystemSource': 'FileSystemSource', 'SqlDWSource': 'SqlDWSource', 'SqlSource': 'SqlSource', 'SapEccSource': 'SapEccSource', 'SapCloudForCustomerSource': 'SapCloudForCustomerSource', 'SalesforceSource': 'SalesforceSource', 'RelationalSource': 'RelationalSource', 'DynamicsSource': 'DynamicsSource', 'DocumentDbCollectionSource': 'DocumentDbCollectionSource', 'BlobSource': 'BlobSource', 'AzureTableSource': 'AzureTableSource'} } def __init__(self, **kwargs): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_source_py3.py index 7c1a96b2897a..f117bfbe7c3b 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_source_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_source_py3.py @@ -16,19 +16,18 @@ class CopySource(Model): """A copy activity source. You probably want to use the sub-classes and not this class directly. Known - sub-classes are: AmazonRedshiftSource, ResponsysSource, - SalesforceMarketingCloudSource, VerticaSource, NetezzaSource, ZohoSource, - XeroSource, SquareSource, SparkSource, ShopifySource, ServiceNowSource, - QuickBooksSource, PrestoSource, PhoenixSource, PaypalSource, MarketoSource, - MariaDBSource, MagentoSource, JiraSource, ImpalaSource, HubspotSource, - HiveSource, HBaseSource, GreenplumSource, GoogleBigQuerySource, - EloquaSource, DrillSource, CouchbaseSource, ConcurSource, - AzurePostgreSqlSource, AmazonMWSSource, HttpSource, - AzureDataLakeStoreSource, MongoDbSource, CassandraSource, WebSource, - OracleSource, AzureMySqlSource, HdfsSource, FileSystemSource, SqlDWSource, - SqlSource, SapEccSource, SapCloudForCustomerSource, SalesforceSource, - RelationalSource, DynamicsSource, DocumentDbCollectionSource, BlobSource, - AzureTableSource + sub-classes are: AmazonRedshiftSource, SalesforceMarketingCloudSource, + VerticaSource, NetezzaSource, ZohoSource, XeroSource, SquareSource, + SparkSource, ShopifySource, ServiceNowSource, QuickBooksSource, + PrestoSource, PhoenixSource, PaypalSource, MarketoSource, MariaDBSource, + MagentoSource, JiraSource, ImpalaSource, HubspotSource, HiveSource, + HBaseSource, GreenplumSource, GoogleBigQuerySource, EloquaSource, + DrillSource, CouchbaseSource, ConcurSource, AzurePostgreSqlSource, + AmazonMWSSource, HttpSource, AzureDataLakeStoreSource, MongoDbSource, + CassandraSource, WebSource, OracleSource, AzureMySqlSource, HdfsSource, + FileSystemSource, SqlDWSource, SqlSource, SapEccSource, + SapCloudForCustomerSource, SalesforceSource, RelationalSource, + DynamicsSource, DocumentDbCollectionSource, BlobSource, AzureTableSource All required parameters must be populated in order to send to Azure. @@ -58,7 +57,7 @@ class CopySource(Model): } _subtype_map = { - 'type': {'AmazonRedshiftSource': 'AmazonRedshiftSource', 'ResponsysSource': 'ResponsysSource', 'SalesforceMarketingCloudSource': 'SalesforceMarketingCloudSource', 'VerticaSource': 'VerticaSource', 'NetezzaSource': 'NetezzaSource', 'ZohoSource': 'ZohoSource', 'XeroSource': 'XeroSource', 'SquareSource': 'SquareSource', 'SparkSource': 'SparkSource', 'ShopifySource': 'ShopifySource', 'ServiceNowSource': 'ServiceNowSource', 'QuickBooksSource': 'QuickBooksSource', 'PrestoSource': 'PrestoSource', 'PhoenixSource': 'PhoenixSource', 'PaypalSource': 'PaypalSource', 'MarketoSource': 'MarketoSource', 'MariaDBSource': 'MariaDBSource', 'MagentoSource': 'MagentoSource', 'JiraSource': 'JiraSource', 'ImpalaSource': 'ImpalaSource', 'HubspotSource': 'HubspotSource', 'HiveSource': 'HiveSource', 'HBaseSource': 'HBaseSource', 'GreenplumSource': 'GreenplumSource', 'GoogleBigQuerySource': 'GoogleBigQuerySource', 'EloquaSource': 'EloquaSource', 'DrillSource': 'DrillSource', 'CouchbaseSource': 'CouchbaseSource', 'ConcurSource': 'ConcurSource', 'AzurePostgreSqlSource': 'AzurePostgreSqlSource', 'AmazonMWSSource': 'AmazonMWSSource', 'HttpSource': 'HttpSource', 'AzureDataLakeStoreSource': 'AzureDataLakeStoreSource', 'MongoDbSource': 'MongoDbSource', 'CassandraSource': 'CassandraSource', 'WebSource': 'WebSource', 'OracleSource': 'OracleSource', 'AzureMySqlSource': 'AzureMySqlSource', 'HdfsSource': 'HdfsSource', 'FileSystemSource': 'FileSystemSource', 'SqlDWSource': 'SqlDWSource', 'SqlSource': 'SqlSource', 'SapEccSource': 'SapEccSource', 'SapCloudForCustomerSource': 'SapCloudForCustomerSource', 'SalesforceSource': 'SalesforceSource', 'RelationalSource': 'RelationalSource', 'DynamicsSource': 'DynamicsSource', 'DocumentDbCollectionSource': 'DocumentDbCollectionSource', 'BlobSource': 'BlobSource', 'AzureTableSource': 'AzureTableSource'} + 'type': {'AmazonRedshiftSource': 'AmazonRedshiftSource', 'SalesforceMarketingCloudSource': 'SalesforceMarketingCloudSource', 'VerticaSource': 'VerticaSource', 'NetezzaSource': 'NetezzaSource', 'ZohoSource': 'ZohoSource', 'XeroSource': 'XeroSource', 'SquareSource': 'SquareSource', 'SparkSource': 'SparkSource', 'ShopifySource': 'ShopifySource', 'ServiceNowSource': 'ServiceNowSource', 'QuickBooksSource': 'QuickBooksSource', 'PrestoSource': 'PrestoSource', 'PhoenixSource': 'PhoenixSource', 'PaypalSource': 'PaypalSource', 'MarketoSource': 'MarketoSource', 'MariaDBSource': 'MariaDBSource', 'MagentoSource': 'MagentoSource', 'JiraSource': 'JiraSource', 'ImpalaSource': 'ImpalaSource', 'HubspotSource': 'HubspotSource', 'HiveSource': 'HiveSource', 'HBaseSource': 'HBaseSource', 'GreenplumSource': 'GreenplumSource', 'GoogleBigQuerySource': 'GoogleBigQuerySource', 'EloquaSource': 'EloquaSource', 'DrillSource': 'DrillSource', 'CouchbaseSource': 'CouchbaseSource', 'ConcurSource': 'ConcurSource', 'AzurePostgreSqlSource': 'AzurePostgreSqlSource', 'AmazonMWSSource': 'AmazonMWSSource', 'HttpSource': 'HttpSource', 'AzureDataLakeStoreSource': 'AzureDataLakeStoreSource', 'MongoDbSource': 'MongoDbSource', 'CassandraSource': 'CassandraSource', 'WebSource': 'WebSource', 'OracleSource': 'OracleSource', 'AzureMySqlSource': 'AzureMySqlSource', 'HdfsSource': 'HdfsSource', 'FileSystemSource': 'FileSystemSource', 'SqlDWSource': 'SqlDWSource', 'SqlSource': 'SqlSource', 'SapEccSource': 'SapEccSource', 'SapCloudForCustomerSource': 'SapCloudForCustomerSource', 'SalesforceSource': 'SalesforceSource', 'RelationalSource': 'RelationalSource', 'DynamicsSource': 'DynamicsSource', 'DocumentDbCollectionSource': 'DocumentDbCollectionSource', 'BlobSource': 'BlobSource', 'AzureTableSource': 'AzureTableSource'} } def __init__(self, *, additional_properties=None, source_retry_count=None, source_retry_wait=None, **kwargs) -> None: diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset.py index 4ebebafdda73..9fc6e4a433f6 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset.py @@ -17,19 +17,18 @@ class Dataset(Model): data stores, such as tables, files, folders, and documents. You probably want to use the sub-classes and not this class directly. Known - sub-classes are: ResponsysObjectDataset, - SalesforceMarketingCloudObjectDataset, VerticaTableDataset, - NetezzaTableDataset, ZohoObjectDataset, XeroObjectDataset, - SquareObjectDataset, SparkObjectDataset, ShopifyObjectDataset, - ServiceNowObjectDataset, QuickBooksObjectDataset, PrestoObjectDataset, - PhoenixObjectDataset, PaypalObjectDataset, MarketoObjectDataset, - MariaDBTableDataset, MagentoObjectDataset, JiraObjectDataset, - ImpalaObjectDataset, HubspotObjectDataset, HiveObjectDataset, - HBaseObjectDataset, GreenplumTableDataset, GoogleBigQueryObjectDataset, - EloquaObjectDataset, DrillTableDataset, CouchbaseTableDataset, - ConcurObjectDataset, AzurePostgreSqlTableDataset, AmazonMWSObjectDataset, - HttpDataset, AzureSearchIndexDataset, WebTableDataset, - SqlServerTableDataset, SapEccResourceDataset, + sub-classes are: SalesforceMarketingCloudObjectDataset, + VerticaTableDataset, NetezzaTableDataset, ZohoObjectDataset, + XeroObjectDataset, SquareObjectDataset, SparkObjectDataset, + ShopifyObjectDataset, ServiceNowObjectDataset, QuickBooksObjectDataset, + PrestoObjectDataset, PhoenixObjectDataset, PaypalObjectDataset, + MarketoObjectDataset, MariaDBTableDataset, MagentoObjectDataset, + JiraObjectDataset, ImpalaObjectDataset, HubspotObjectDataset, + HiveObjectDataset, HBaseObjectDataset, GreenplumTableDataset, + GoogleBigQueryObjectDataset, EloquaObjectDataset, DrillTableDataset, + CouchbaseTableDataset, ConcurObjectDataset, AzurePostgreSqlTableDataset, + AmazonMWSObjectDataset, HttpDataset, AzureSearchIndexDataset, + WebTableDataset, SqlServerTableDataset, SapEccResourceDataset, SapCloudForCustomerResourceDataset, SalesforceObjectDataset, RelationalTableDataset, AzureMySqlTableDataset, OracleTableDataset, ODataResourceDataset, MongoDbCollectionDataset, FileShareDataset, @@ -77,7 +76,7 @@ class Dataset(Model): } _subtype_map = { - 'type': {'ResponsysObject': 'ResponsysObjectDataset', 'SalesforceMarketingCloudObject': 'SalesforceMarketingCloudObjectDataset', 'VerticaTable': 'VerticaTableDataset', 'NetezzaTable': 'NetezzaTableDataset', 'ZohoObject': 'ZohoObjectDataset', 'XeroObject': 'XeroObjectDataset', 'SquareObject': 'SquareObjectDataset', 'SparkObject': 'SparkObjectDataset', 'ShopifyObject': 'ShopifyObjectDataset', 'ServiceNowObject': 'ServiceNowObjectDataset', 'QuickBooksObject': 'QuickBooksObjectDataset', 'PrestoObject': 'PrestoObjectDataset', 'PhoenixObject': 'PhoenixObjectDataset', 'PaypalObject': 'PaypalObjectDataset', 'MarketoObject': 'MarketoObjectDataset', 'MariaDBTable': 'MariaDBTableDataset', 'MagentoObject': 'MagentoObjectDataset', 'JiraObject': 'JiraObjectDataset', 'ImpalaObject': 'ImpalaObjectDataset', 'HubspotObject': 'HubspotObjectDataset', 'HiveObject': 'HiveObjectDataset', 'HBaseObject': 'HBaseObjectDataset', 'GreenplumTable': 'GreenplumTableDataset', 'GoogleBigQueryObject': 'GoogleBigQueryObjectDataset', 'EloquaObject': 'EloquaObjectDataset', 'DrillTable': 'DrillTableDataset', 'CouchbaseTable': 'CouchbaseTableDataset', 'ConcurObject': 'ConcurObjectDataset', 'AzurePostgreSqlTable': 'AzurePostgreSqlTableDataset', 'AmazonMWSObject': 'AmazonMWSObjectDataset', 'HttpFile': 'HttpDataset', 'AzureSearchIndex': 'AzureSearchIndexDataset', 'WebTable': 'WebTableDataset', 'SqlServerTable': 'SqlServerTableDataset', 'SapEccResource': 'SapEccResourceDataset', 'SapCloudForCustomerResource': 'SapCloudForCustomerResourceDataset', 'SalesforceObject': 'SalesforceObjectDataset', 'RelationalTable': 'RelationalTableDataset', 'AzureMySqlTable': 'AzureMySqlTableDataset', 'OracleTable': 'OracleTableDataset', 'ODataResource': 'ODataResourceDataset', 'MongoDbCollection': 'MongoDbCollectionDataset', 'FileShare': 'FileShareDataset', 'AzureDataLakeStoreFile': 'AzureDataLakeStoreDataset', 'DynamicsEntity': 'DynamicsEntityDataset', 'DocumentDbCollection': 'DocumentDbCollectionDataset', 'CustomDataset': 'CustomDataset', 'CassandraTable': 'CassandraTableDataset', 'AzureSqlDWTable': 'AzureSqlDWTableDataset', 'AzureSqlTable': 'AzureSqlTableDataset', 'AzureTable': 'AzureTableDataset', 'AzureBlob': 'AzureBlobDataset', 'AmazonS3Object': 'AmazonS3Dataset'} + 'type': {'SalesforceMarketingCloudObject': 'SalesforceMarketingCloudObjectDataset', 'VerticaTable': 'VerticaTableDataset', 'NetezzaTable': 'NetezzaTableDataset', 'ZohoObject': 'ZohoObjectDataset', 'XeroObject': 'XeroObjectDataset', 'SquareObject': 'SquareObjectDataset', 'SparkObject': 'SparkObjectDataset', 'ShopifyObject': 'ShopifyObjectDataset', 'ServiceNowObject': 'ServiceNowObjectDataset', 'QuickBooksObject': 'QuickBooksObjectDataset', 'PrestoObject': 'PrestoObjectDataset', 'PhoenixObject': 'PhoenixObjectDataset', 'PaypalObject': 'PaypalObjectDataset', 'MarketoObject': 'MarketoObjectDataset', 'MariaDBTable': 'MariaDBTableDataset', 'MagentoObject': 'MagentoObjectDataset', 'JiraObject': 'JiraObjectDataset', 'ImpalaObject': 'ImpalaObjectDataset', 'HubspotObject': 'HubspotObjectDataset', 'HiveObject': 'HiveObjectDataset', 'HBaseObject': 'HBaseObjectDataset', 'GreenplumTable': 'GreenplumTableDataset', 'GoogleBigQueryObject': 'GoogleBigQueryObjectDataset', 'EloquaObject': 'EloquaObjectDataset', 'DrillTable': 'DrillTableDataset', 'CouchbaseTable': 'CouchbaseTableDataset', 'ConcurObject': 'ConcurObjectDataset', 'AzurePostgreSqlTable': 'AzurePostgreSqlTableDataset', 'AmazonMWSObject': 'AmazonMWSObjectDataset', 'HttpFile': 'HttpDataset', 'AzureSearchIndex': 'AzureSearchIndexDataset', 'WebTable': 'WebTableDataset', 'SqlServerTable': 'SqlServerTableDataset', 'SapEccResource': 'SapEccResourceDataset', 'SapCloudForCustomerResource': 'SapCloudForCustomerResourceDataset', 'SalesforceObject': 'SalesforceObjectDataset', 'RelationalTable': 'RelationalTableDataset', 'AzureMySqlTable': 'AzureMySqlTableDataset', 'OracleTable': 'OracleTableDataset', 'ODataResource': 'ODataResourceDataset', 'MongoDbCollection': 'MongoDbCollectionDataset', 'FileShare': 'FileShareDataset', 'AzureDataLakeStoreFile': 'AzureDataLakeStoreDataset', 'DynamicsEntity': 'DynamicsEntityDataset', 'DocumentDbCollection': 'DocumentDbCollectionDataset', 'CustomDataset': 'CustomDataset', 'CassandraTable': 'CassandraTableDataset', 'AzureSqlDWTable': 'AzureSqlDWTableDataset', 'AzureSqlTable': 'AzureSqlTableDataset', 'AzureTable': 'AzureTableDataset', 'AzureBlob': 'AzureBlobDataset', 'AmazonS3Object': 'AmazonS3Dataset'} } def __init__(self, **kwargs): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_py3.py index 65f43811d7a7..18d3a0d74068 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_py3.py @@ -17,19 +17,18 @@ class Dataset(Model): data stores, such as tables, files, folders, and documents. You probably want to use the sub-classes and not this class directly. Known - sub-classes are: ResponsysObjectDataset, - SalesforceMarketingCloudObjectDataset, VerticaTableDataset, - NetezzaTableDataset, ZohoObjectDataset, XeroObjectDataset, - SquareObjectDataset, SparkObjectDataset, ShopifyObjectDataset, - ServiceNowObjectDataset, QuickBooksObjectDataset, PrestoObjectDataset, - PhoenixObjectDataset, PaypalObjectDataset, MarketoObjectDataset, - MariaDBTableDataset, MagentoObjectDataset, JiraObjectDataset, - ImpalaObjectDataset, HubspotObjectDataset, HiveObjectDataset, - HBaseObjectDataset, GreenplumTableDataset, GoogleBigQueryObjectDataset, - EloquaObjectDataset, DrillTableDataset, CouchbaseTableDataset, - ConcurObjectDataset, AzurePostgreSqlTableDataset, AmazonMWSObjectDataset, - HttpDataset, AzureSearchIndexDataset, WebTableDataset, - SqlServerTableDataset, SapEccResourceDataset, + sub-classes are: SalesforceMarketingCloudObjectDataset, + VerticaTableDataset, NetezzaTableDataset, ZohoObjectDataset, + XeroObjectDataset, SquareObjectDataset, SparkObjectDataset, + ShopifyObjectDataset, ServiceNowObjectDataset, QuickBooksObjectDataset, + PrestoObjectDataset, PhoenixObjectDataset, PaypalObjectDataset, + MarketoObjectDataset, MariaDBTableDataset, MagentoObjectDataset, + JiraObjectDataset, ImpalaObjectDataset, HubspotObjectDataset, + HiveObjectDataset, HBaseObjectDataset, GreenplumTableDataset, + GoogleBigQueryObjectDataset, EloquaObjectDataset, DrillTableDataset, + CouchbaseTableDataset, ConcurObjectDataset, AzurePostgreSqlTableDataset, + AmazonMWSObjectDataset, HttpDataset, AzureSearchIndexDataset, + WebTableDataset, SqlServerTableDataset, SapEccResourceDataset, SapCloudForCustomerResourceDataset, SalesforceObjectDataset, RelationalTableDataset, AzureMySqlTableDataset, OracleTableDataset, ODataResourceDataset, MongoDbCollectionDataset, FileShareDataset, @@ -77,7 +76,7 @@ class Dataset(Model): } _subtype_map = { - 'type': {'ResponsysObject': 'ResponsysObjectDataset', 'SalesforceMarketingCloudObject': 'SalesforceMarketingCloudObjectDataset', 'VerticaTable': 'VerticaTableDataset', 'NetezzaTable': 'NetezzaTableDataset', 'ZohoObject': 'ZohoObjectDataset', 'XeroObject': 'XeroObjectDataset', 'SquareObject': 'SquareObjectDataset', 'SparkObject': 'SparkObjectDataset', 'ShopifyObject': 'ShopifyObjectDataset', 'ServiceNowObject': 'ServiceNowObjectDataset', 'QuickBooksObject': 'QuickBooksObjectDataset', 'PrestoObject': 'PrestoObjectDataset', 'PhoenixObject': 'PhoenixObjectDataset', 'PaypalObject': 'PaypalObjectDataset', 'MarketoObject': 'MarketoObjectDataset', 'MariaDBTable': 'MariaDBTableDataset', 'MagentoObject': 'MagentoObjectDataset', 'JiraObject': 'JiraObjectDataset', 'ImpalaObject': 'ImpalaObjectDataset', 'HubspotObject': 'HubspotObjectDataset', 'HiveObject': 'HiveObjectDataset', 'HBaseObject': 'HBaseObjectDataset', 'GreenplumTable': 'GreenplumTableDataset', 'GoogleBigQueryObject': 'GoogleBigQueryObjectDataset', 'EloquaObject': 'EloquaObjectDataset', 'DrillTable': 'DrillTableDataset', 'CouchbaseTable': 'CouchbaseTableDataset', 'ConcurObject': 'ConcurObjectDataset', 'AzurePostgreSqlTable': 'AzurePostgreSqlTableDataset', 'AmazonMWSObject': 'AmazonMWSObjectDataset', 'HttpFile': 'HttpDataset', 'AzureSearchIndex': 'AzureSearchIndexDataset', 'WebTable': 'WebTableDataset', 'SqlServerTable': 'SqlServerTableDataset', 'SapEccResource': 'SapEccResourceDataset', 'SapCloudForCustomerResource': 'SapCloudForCustomerResourceDataset', 'SalesforceObject': 'SalesforceObjectDataset', 'RelationalTable': 'RelationalTableDataset', 'AzureMySqlTable': 'AzureMySqlTableDataset', 'OracleTable': 'OracleTableDataset', 'ODataResource': 'ODataResourceDataset', 'MongoDbCollection': 'MongoDbCollectionDataset', 'FileShare': 'FileShareDataset', 'AzureDataLakeStoreFile': 'AzureDataLakeStoreDataset', 'DynamicsEntity': 'DynamicsEntityDataset', 'DocumentDbCollection': 'DocumentDbCollectionDataset', 'CustomDataset': 'CustomDataset', 'CassandraTable': 'CassandraTableDataset', 'AzureSqlDWTable': 'AzureSqlDWTableDataset', 'AzureSqlTable': 'AzureSqlTableDataset', 'AzureTable': 'AzureTableDataset', 'AzureBlob': 'AzureBlobDataset', 'AmazonS3Object': 'AmazonS3Dataset'} + 'type': {'SalesforceMarketingCloudObject': 'SalesforceMarketingCloudObjectDataset', 'VerticaTable': 'VerticaTableDataset', 'NetezzaTable': 'NetezzaTableDataset', 'ZohoObject': 'ZohoObjectDataset', 'XeroObject': 'XeroObjectDataset', 'SquareObject': 'SquareObjectDataset', 'SparkObject': 'SparkObjectDataset', 'ShopifyObject': 'ShopifyObjectDataset', 'ServiceNowObject': 'ServiceNowObjectDataset', 'QuickBooksObject': 'QuickBooksObjectDataset', 'PrestoObject': 'PrestoObjectDataset', 'PhoenixObject': 'PhoenixObjectDataset', 'PaypalObject': 'PaypalObjectDataset', 'MarketoObject': 'MarketoObjectDataset', 'MariaDBTable': 'MariaDBTableDataset', 'MagentoObject': 'MagentoObjectDataset', 'JiraObject': 'JiraObjectDataset', 'ImpalaObject': 'ImpalaObjectDataset', 'HubspotObject': 'HubspotObjectDataset', 'HiveObject': 'HiveObjectDataset', 'HBaseObject': 'HBaseObjectDataset', 'GreenplumTable': 'GreenplumTableDataset', 'GoogleBigQueryObject': 'GoogleBigQueryObjectDataset', 'EloquaObject': 'EloquaObjectDataset', 'DrillTable': 'DrillTableDataset', 'CouchbaseTable': 'CouchbaseTableDataset', 'ConcurObject': 'ConcurObjectDataset', 'AzurePostgreSqlTable': 'AzurePostgreSqlTableDataset', 'AmazonMWSObject': 'AmazonMWSObjectDataset', 'HttpFile': 'HttpDataset', 'AzureSearchIndex': 'AzureSearchIndexDataset', 'WebTable': 'WebTableDataset', 'SqlServerTable': 'SqlServerTableDataset', 'SapEccResource': 'SapEccResourceDataset', 'SapCloudForCustomerResource': 'SapCloudForCustomerResourceDataset', 'SalesforceObject': 'SalesforceObjectDataset', 'RelationalTable': 'RelationalTableDataset', 'AzureMySqlTable': 'AzureMySqlTableDataset', 'OracleTable': 'OracleTableDataset', 'ODataResource': 'ODataResourceDataset', 'MongoDbCollection': 'MongoDbCollectionDataset', 'FileShare': 'FileShareDataset', 'AzureDataLakeStoreFile': 'AzureDataLakeStoreDataset', 'DynamicsEntity': 'DynamicsEntityDataset', 'DocumentDbCollection': 'DocumentDbCollectionDataset', 'CustomDataset': 'CustomDataset', 'CassandraTable': 'CassandraTableDataset', 'AzureSqlDWTable': 'AzureSqlDWTableDataset', 'AzureSqlTable': 'AzureSqlTableDataset', 'AzureTable': 'AzureTableDataset', 'AzureBlob': 'AzureBlobDataset', 'AmazonS3Object': 'AmazonS3Dataset'} } def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_ssis_package_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_ssis_package_activity.py index a72ca4bbafdb..f835b78bec55 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_ssis_package_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_ssis_package_activity.py @@ -40,12 +40,31 @@ class ExecuteSSISPackageActivity(ExecutionActivity): :type runtime: str or ~azure.mgmt.datafactory.models.SSISExecutionRuntime :param logging_level: The logging level of SSIS package execution. :type logging_level: str - :param environment_path: The environment path to execution the SSIS - package. + :param environment_path: The environment path to execute the SSIS package. :type environment_path: str :param connect_via: Required. The integration runtime reference. :type connect_via: ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param project_parameters: The project level parameters to execute the + SSIS package. + :type project_parameters: dict[str, + ~azure.mgmt.datafactory.models.SSISExecutionParameter] + :param package_parameters: The package level parameters to execute the + SSIS package. + :type package_parameters: dict[str, + ~azure.mgmt.datafactory.models.SSISExecutionParameter] + :param project_connection_managers: The project level connection managers + to execute the SSIS package. + :type project_connection_managers: dict[str, dict[str, + ~azure.mgmt.datafactory.models.SSISExecutionParameter]] + :param package_connection_managers: The package level connection managers + to execute the SSIS package. + :type package_connection_managers: dict[str, dict[str, + ~azure.mgmt.datafactory.models.SSISExecutionParameter]] + :param property_overrides: The property overrides to execute the SSIS + package. + :type property_overrides: dict[str, + ~azure.mgmt.datafactory.models.SSISPropertyOverride] """ _validation = { @@ -68,6 +87,11 @@ class ExecuteSSISPackageActivity(ExecutionActivity): 'logging_level': {'key': 'typeProperties.loggingLevel', 'type': 'str'}, 'environment_path': {'key': 'typeProperties.environmentPath', 'type': 'str'}, 'connect_via': {'key': 'typeProperties.connectVia', 'type': 'IntegrationRuntimeReference'}, + 'project_parameters': {'key': 'typeProperties.projectParameters', 'type': '{SSISExecutionParameter}'}, + 'package_parameters': {'key': 'typeProperties.packageParameters', 'type': '{SSISExecutionParameter}'}, + 'project_connection_managers': {'key': 'typeProperties.projectConnectionManagers', 'type': '{{SSISExecutionParameter}}'}, + 'package_connection_managers': {'key': 'typeProperties.packageConnectionManagers', 'type': '{{SSISExecutionParameter}}'}, + 'property_overrides': {'key': 'typeProperties.propertyOverrides', 'type': '{SSISPropertyOverride}'}, } def __init__(self, **kwargs): @@ -77,4 +101,9 @@ def __init__(self, **kwargs): self.logging_level = kwargs.get('logging_level', None) self.environment_path = kwargs.get('environment_path', None) self.connect_via = kwargs.get('connect_via', None) + self.project_parameters = kwargs.get('project_parameters', None) + self.package_parameters = kwargs.get('package_parameters', None) + self.project_connection_managers = kwargs.get('project_connection_managers', None) + self.package_connection_managers = kwargs.get('package_connection_managers', None) + self.property_overrides = kwargs.get('property_overrides', None) self.type = 'ExecuteSSISPackage' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_ssis_package_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_ssis_package_activity_py3.py index df053457dd16..9c1de33e7433 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_ssis_package_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_ssis_package_activity_py3.py @@ -40,12 +40,31 @@ class ExecuteSSISPackageActivity(ExecutionActivity): :type runtime: str or ~azure.mgmt.datafactory.models.SSISExecutionRuntime :param logging_level: The logging level of SSIS package execution. :type logging_level: str - :param environment_path: The environment path to execution the SSIS - package. + :param environment_path: The environment path to execute the SSIS package. :type environment_path: str :param connect_via: Required. The integration runtime reference. :type connect_via: ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param project_parameters: The project level parameters to execute the + SSIS package. + :type project_parameters: dict[str, + ~azure.mgmt.datafactory.models.SSISExecutionParameter] + :param package_parameters: The package level parameters to execute the + SSIS package. + :type package_parameters: dict[str, + ~azure.mgmt.datafactory.models.SSISExecutionParameter] + :param project_connection_managers: The project level connection managers + to execute the SSIS package. + :type project_connection_managers: dict[str, dict[str, + ~azure.mgmt.datafactory.models.SSISExecutionParameter]] + :param package_connection_managers: The package level connection managers + to execute the SSIS package. + :type package_connection_managers: dict[str, dict[str, + ~azure.mgmt.datafactory.models.SSISExecutionParameter]] + :param property_overrides: The property overrides to execute the SSIS + package. + :type property_overrides: dict[str, + ~azure.mgmt.datafactory.models.SSISPropertyOverride] """ _validation = { @@ -68,13 +87,23 @@ class ExecuteSSISPackageActivity(ExecutionActivity): 'logging_level': {'key': 'typeProperties.loggingLevel', 'type': 'str'}, 'environment_path': {'key': 'typeProperties.environmentPath', 'type': 'str'}, 'connect_via': {'key': 'typeProperties.connectVia', 'type': 'IntegrationRuntimeReference'}, + 'project_parameters': {'key': 'typeProperties.projectParameters', 'type': '{SSISExecutionParameter}'}, + 'package_parameters': {'key': 'typeProperties.packageParameters', 'type': '{SSISExecutionParameter}'}, + 'project_connection_managers': {'key': 'typeProperties.projectConnectionManagers', 'type': '{{SSISExecutionParameter}}'}, + 'package_connection_managers': {'key': 'typeProperties.packageConnectionManagers', 'type': '{{SSISExecutionParameter}}'}, + 'property_overrides': {'key': 'typeProperties.propertyOverrides', 'type': '{SSISPropertyOverride}'}, } - def __init__(self, *, name: str, package_location, connect_via, additional_properties=None, description: str=None, depends_on=None, linked_service_name=None, policy=None, runtime=None, logging_level: str=None, environment_path: str=None, **kwargs) -> None: + def __init__(self, *, name: str, package_location, connect_via, additional_properties=None, description: str=None, depends_on=None, linked_service_name=None, policy=None, runtime=None, logging_level: str=None, environment_path: str=None, project_parameters=None, package_parameters=None, project_connection_managers=None, package_connection_managers=None, property_overrides=None, **kwargs) -> None: super(ExecuteSSISPackageActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, linked_service_name=linked_service_name, policy=policy, **kwargs) self.package_location = package_location self.runtime = runtime self.logging_level = logging_level self.environment_path = environment_path self.connect_via = connect_via + self.project_parameters = project_parameters + self.package_parameters = package_parameters + self.project_connection_managers = project_connection_managers + self.package_connection_managers = package_connection_managers + self.property_overrides = property_overrides self.type = 'ExecuteSSISPackage' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service.py index ecb6b42e145a..efb89ee0eacd 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service.py @@ -18,7 +18,7 @@ class LinkedService(Model): resource. You probably want to use the sub-classes and not this class directly. Known - sub-classes are: ResponsysLinkedService, AzureDatabricksLinkedService, + sub-classes are: AzureDatabricksLinkedService, AzureDataLakeAnalyticsLinkedService, HDInsightOnDemandLinkedService, SalesforceMarketingCloudLinkedService, NetezzaLinkedService, VerticaLinkedService, ZohoLinkedService, XeroLinkedService, @@ -79,7 +79,7 @@ class LinkedService(Model): } _subtype_map = { - 'type': {'Responsys': 'ResponsysLinkedService', 'AzureDatabricks': 'AzureDatabricksLinkedService', 'AzureDataLakeAnalytics': 'AzureDataLakeAnalyticsLinkedService', 'HDInsightOnDemand': 'HDInsightOnDemandLinkedService', 'SalesforceMarketingCloud': 'SalesforceMarketingCloudLinkedService', 'Netezza': 'NetezzaLinkedService', 'Vertica': 'VerticaLinkedService', 'Zoho': 'ZohoLinkedService', 'Xero': 'XeroLinkedService', 'Square': 'SquareLinkedService', 'Spark': 'SparkLinkedService', 'Shopify': 'ShopifyLinkedService', 'ServiceNow': 'ServiceNowLinkedService', 'QuickBooks': 'QuickBooksLinkedService', 'Presto': 'PrestoLinkedService', 'Phoenix': 'PhoenixLinkedService', 'Paypal': 'PaypalLinkedService', 'Marketo': 'MarketoLinkedService', 'MariaDB': 'MariaDBLinkedService', 'Magento': 'MagentoLinkedService', 'Jira': 'JiraLinkedService', 'Impala': 'ImpalaLinkedService', 'Hubspot': 'HubspotLinkedService', 'Hive': 'HiveLinkedService', 'HBase': 'HBaseLinkedService', 'Greenplum': 'GreenplumLinkedService', 'GoogleBigQuery': 'GoogleBigQueryLinkedService', 'Eloqua': 'EloquaLinkedService', 'Drill': 'DrillLinkedService', 'Couchbase': 'CouchbaseLinkedService', 'Concur': 'ConcurLinkedService', 'AzurePostgreSql': 'AzurePostgreSqlLinkedService', 'AmazonMWS': 'AmazonMWSLinkedService', 'SapHana': 'SapHanaLinkedService', 'SapBW': 'SapBWLinkedService', 'Sftp': 'SftpServerLinkedService', 'FtpServer': 'FtpServerLinkedService', 'HttpServer': 'HttpLinkedService', 'AzureSearch': 'AzureSearchLinkedService', 'CustomDataSource': 'CustomDataSourceLinkedService', 'AmazonRedshift': 'AmazonRedshiftLinkedService', 'AmazonS3': 'AmazonS3LinkedService', 'SapEcc': 'SapEccLinkedService', 'SapCloudForCustomer': 'SapCloudForCustomerLinkedService', 'Salesforce': 'SalesforceLinkedService', 'AzureDataLakeStore': 'AzureDataLakeStoreLinkedService', 'MongoDb': 'MongoDbLinkedService', 'Cassandra': 'CassandraLinkedService', 'Web': 'WebLinkedService', 'OData': 'ODataLinkedService', 'Hdfs': 'HdfsLinkedService', 'Odbc': 'OdbcLinkedService', 'AzureML': 'AzureMLLinkedService', 'Teradata': 'TeradataLinkedService', 'Db2': 'Db2LinkedService', 'Sybase': 'SybaseLinkedService', 'PostgreSql': 'PostgreSqlLinkedService', 'MySql': 'MySqlLinkedService', 'AzureMySql': 'AzureMySqlLinkedService', 'Oracle': 'OracleLinkedService', 'FileServer': 'FileServerLinkedService', 'HDInsight': 'HDInsightLinkedService', 'Dynamics': 'DynamicsLinkedService', 'CosmosDb': 'CosmosDbLinkedService', 'AzureKeyVault': 'AzureKeyVaultLinkedService', 'AzureBatch': 'AzureBatchLinkedService', 'AzureSqlDatabase': 'AzureSqlDatabaseLinkedService', 'SqlServer': 'SqlServerLinkedService', 'AzureSqlDW': 'AzureSqlDWLinkedService', 'AzureStorage': 'AzureStorageLinkedService'} + 'type': {'AzureDatabricks': 'AzureDatabricksLinkedService', 'AzureDataLakeAnalytics': 'AzureDataLakeAnalyticsLinkedService', 'HDInsightOnDemand': 'HDInsightOnDemandLinkedService', 'SalesforceMarketingCloud': 'SalesforceMarketingCloudLinkedService', 'Netezza': 'NetezzaLinkedService', 'Vertica': 'VerticaLinkedService', 'Zoho': 'ZohoLinkedService', 'Xero': 'XeroLinkedService', 'Square': 'SquareLinkedService', 'Spark': 'SparkLinkedService', 'Shopify': 'ShopifyLinkedService', 'ServiceNow': 'ServiceNowLinkedService', 'QuickBooks': 'QuickBooksLinkedService', 'Presto': 'PrestoLinkedService', 'Phoenix': 'PhoenixLinkedService', 'Paypal': 'PaypalLinkedService', 'Marketo': 'MarketoLinkedService', 'MariaDB': 'MariaDBLinkedService', 'Magento': 'MagentoLinkedService', 'Jira': 'JiraLinkedService', 'Impala': 'ImpalaLinkedService', 'Hubspot': 'HubspotLinkedService', 'Hive': 'HiveLinkedService', 'HBase': 'HBaseLinkedService', 'Greenplum': 'GreenplumLinkedService', 'GoogleBigQuery': 'GoogleBigQueryLinkedService', 'Eloqua': 'EloquaLinkedService', 'Drill': 'DrillLinkedService', 'Couchbase': 'CouchbaseLinkedService', 'Concur': 'ConcurLinkedService', 'AzurePostgreSql': 'AzurePostgreSqlLinkedService', 'AmazonMWS': 'AmazonMWSLinkedService', 'SapHana': 'SapHanaLinkedService', 'SapBW': 'SapBWLinkedService', 'Sftp': 'SftpServerLinkedService', 'FtpServer': 'FtpServerLinkedService', 'HttpServer': 'HttpLinkedService', 'AzureSearch': 'AzureSearchLinkedService', 'CustomDataSource': 'CustomDataSourceLinkedService', 'AmazonRedshift': 'AmazonRedshiftLinkedService', 'AmazonS3': 'AmazonS3LinkedService', 'SapEcc': 'SapEccLinkedService', 'SapCloudForCustomer': 'SapCloudForCustomerLinkedService', 'Salesforce': 'SalesforceLinkedService', 'AzureDataLakeStore': 'AzureDataLakeStoreLinkedService', 'MongoDb': 'MongoDbLinkedService', 'Cassandra': 'CassandraLinkedService', 'Web': 'WebLinkedService', 'OData': 'ODataLinkedService', 'Hdfs': 'HdfsLinkedService', 'Odbc': 'OdbcLinkedService', 'AzureML': 'AzureMLLinkedService', 'Teradata': 'TeradataLinkedService', 'Db2': 'Db2LinkedService', 'Sybase': 'SybaseLinkedService', 'PostgreSql': 'PostgreSqlLinkedService', 'MySql': 'MySqlLinkedService', 'AzureMySql': 'AzureMySqlLinkedService', 'Oracle': 'OracleLinkedService', 'FileServer': 'FileServerLinkedService', 'HDInsight': 'HDInsightLinkedService', 'Dynamics': 'DynamicsLinkedService', 'CosmosDb': 'CosmosDbLinkedService', 'AzureKeyVault': 'AzureKeyVaultLinkedService', 'AzureBatch': 'AzureBatchLinkedService', 'AzureSqlDatabase': 'AzureSqlDatabaseLinkedService', 'SqlServer': 'SqlServerLinkedService', 'AzureSqlDW': 'AzureSqlDWLinkedService', 'AzureStorage': 'AzureStorageLinkedService'} } def __init__(self, **kwargs): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service_py3.py index 39119635cba5..202165ec4982 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service_py3.py @@ -18,7 +18,7 @@ class LinkedService(Model): resource. You probably want to use the sub-classes and not this class directly. Known - sub-classes are: ResponsysLinkedService, AzureDatabricksLinkedService, + sub-classes are: AzureDatabricksLinkedService, AzureDataLakeAnalyticsLinkedService, HDInsightOnDemandLinkedService, SalesforceMarketingCloudLinkedService, NetezzaLinkedService, VerticaLinkedService, ZohoLinkedService, XeroLinkedService, @@ -79,7 +79,7 @@ class LinkedService(Model): } _subtype_map = { - 'type': {'Responsys': 'ResponsysLinkedService', 'AzureDatabricks': 'AzureDatabricksLinkedService', 'AzureDataLakeAnalytics': 'AzureDataLakeAnalyticsLinkedService', 'HDInsightOnDemand': 'HDInsightOnDemandLinkedService', 'SalesforceMarketingCloud': 'SalesforceMarketingCloudLinkedService', 'Netezza': 'NetezzaLinkedService', 'Vertica': 'VerticaLinkedService', 'Zoho': 'ZohoLinkedService', 'Xero': 'XeroLinkedService', 'Square': 'SquareLinkedService', 'Spark': 'SparkLinkedService', 'Shopify': 'ShopifyLinkedService', 'ServiceNow': 'ServiceNowLinkedService', 'QuickBooks': 'QuickBooksLinkedService', 'Presto': 'PrestoLinkedService', 'Phoenix': 'PhoenixLinkedService', 'Paypal': 'PaypalLinkedService', 'Marketo': 'MarketoLinkedService', 'MariaDB': 'MariaDBLinkedService', 'Magento': 'MagentoLinkedService', 'Jira': 'JiraLinkedService', 'Impala': 'ImpalaLinkedService', 'Hubspot': 'HubspotLinkedService', 'Hive': 'HiveLinkedService', 'HBase': 'HBaseLinkedService', 'Greenplum': 'GreenplumLinkedService', 'GoogleBigQuery': 'GoogleBigQueryLinkedService', 'Eloqua': 'EloquaLinkedService', 'Drill': 'DrillLinkedService', 'Couchbase': 'CouchbaseLinkedService', 'Concur': 'ConcurLinkedService', 'AzurePostgreSql': 'AzurePostgreSqlLinkedService', 'AmazonMWS': 'AmazonMWSLinkedService', 'SapHana': 'SapHanaLinkedService', 'SapBW': 'SapBWLinkedService', 'Sftp': 'SftpServerLinkedService', 'FtpServer': 'FtpServerLinkedService', 'HttpServer': 'HttpLinkedService', 'AzureSearch': 'AzureSearchLinkedService', 'CustomDataSource': 'CustomDataSourceLinkedService', 'AmazonRedshift': 'AmazonRedshiftLinkedService', 'AmazonS3': 'AmazonS3LinkedService', 'SapEcc': 'SapEccLinkedService', 'SapCloudForCustomer': 'SapCloudForCustomerLinkedService', 'Salesforce': 'SalesforceLinkedService', 'AzureDataLakeStore': 'AzureDataLakeStoreLinkedService', 'MongoDb': 'MongoDbLinkedService', 'Cassandra': 'CassandraLinkedService', 'Web': 'WebLinkedService', 'OData': 'ODataLinkedService', 'Hdfs': 'HdfsLinkedService', 'Odbc': 'OdbcLinkedService', 'AzureML': 'AzureMLLinkedService', 'Teradata': 'TeradataLinkedService', 'Db2': 'Db2LinkedService', 'Sybase': 'SybaseLinkedService', 'PostgreSql': 'PostgreSqlLinkedService', 'MySql': 'MySqlLinkedService', 'AzureMySql': 'AzureMySqlLinkedService', 'Oracle': 'OracleLinkedService', 'FileServer': 'FileServerLinkedService', 'HDInsight': 'HDInsightLinkedService', 'Dynamics': 'DynamicsLinkedService', 'CosmosDb': 'CosmosDbLinkedService', 'AzureKeyVault': 'AzureKeyVaultLinkedService', 'AzureBatch': 'AzureBatchLinkedService', 'AzureSqlDatabase': 'AzureSqlDatabaseLinkedService', 'SqlServer': 'SqlServerLinkedService', 'AzureSqlDW': 'AzureSqlDWLinkedService', 'AzureStorage': 'AzureStorageLinkedService'} + 'type': {'AzureDatabricks': 'AzureDatabricksLinkedService', 'AzureDataLakeAnalytics': 'AzureDataLakeAnalyticsLinkedService', 'HDInsightOnDemand': 'HDInsightOnDemandLinkedService', 'SalesforceMarketingCloud': 'SalesforceMarketingCloudLinkedService', 'Netezza': 'NetezzaLinkedService', 'Vertica': 'VerticaLinkedService', 'Zoho': 'ZohoLinkedService', 'Xero': 'XeroLinkedService', 'Square': 'SquareLinkedService', 'Spark': 'SparkLinkedService', 'Shopify': 'ShopifyLinkedService', 'ServiceNow': 'ServiceNowLinkedService', 'QuickBooks': 'QuickBooksLinkedService', 'Presto': 'PrestoLinkedService', 'Phoenix': 'PhoenixLinkedService', 'Paypal': 'PaypalLinkedService', 'Marketo': 'MarketoLinkedService', 'MariaDB': 'MariaDBLinkedService', 'Magento': 'MagentoLinkedService', 'Jira': 'JiraLinkedService', 'Impala': 'ImpalaLinkedService', 'Hubspot': 'HubspotLinkedService', 'Hive': 'HiveLinkedService', 'HBase': 'HBaseLinkedService', 'Greenplum': 'GreenplumLinkedService', 'GoogleBigQuery': 'GoogleBigQueryLinkedService', 'Eloqua': 'EloquaLinkedService', 'Drill': 'DrillLinkedService', 'Couchbase': 'CouchbaseLinkedService', 'Concur': 'ConcurLinkedService', 'AzurePostgreSql': 'AzurePostgreSqlLinkedService', 'AmazonMWS': 'AmazonMWSLinkedService', 'SapHana': 'SapHanaLinkedService', 'SapBW': 'SapBWLinkedService', 'Sftp': 'SftpServerLinkedService', 'FtpServer': 'FtpServerLinkedService', 'HttpServer': 'HttpLinkedService', 'AzureSearch': 'AzureSearchLinkedService', 'CustomDataSource': 'CustomDataSourceLinkedService', 'AmazonRedshift': 'AmazonRedshiftLinkedService', 'AmazonS3': 'AmazonS3LinkedService', 'SapEcc': 'SapEccLinkedService', 'SapCloudForCustomer': 'SapCloudForCustomerLinkedService', 'Salesforce': 'SalesforceLinkedService', 'AzureDataLakeStore': 'AzureDataLakeStoreLinkedService', 'MongoDb': 'MongoDbLinkedService', 'Cassandra': 'CassandraLinkedService', 'Web': 'WebLinkedService', 'OData': 'ODataLinkedService', 'Hdfs': 'HdfsLinkedService', 'Odbc': 'OdbcLinkedService', 'AzureML': 'AzureMLLinkedService', 'Teradata': 'TeradataLinkedService', 'Db2': 'Db2LinkedService', 'Sybase': 'SybaseLinkedService', 'PostgreSql': 'PostgreSqlLinkedService', 'MySql': 'MySqlLinkedService', 'AzureMySql': 'AzureMySqlLinkedService', 'Oracle': 'OracleLinkedService', 'FileServer': 'FileServerLinkedService', 'HDInsight': 'HDInsightLinkedService', 'Dynamics': 'DynamicsLinkedService', 'CosmosDb': 'CosmosDbLinkedService', 'AzureKeyVault': 'AzureKeyVaultLinkedService', 'AzureBatch': 'AzureBatchLinkedService', 'AzureSqlDatabase': 'AzureSqlDatabaseLinkedService', 'SqlServer': 'SqlServerLinkedService', 'AzureSqlDW': 'AzureSqlDWLinkedService', 'AzureStorage': 'AzureStorageLinkedService'} } def __init__(self, *, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, **kwargs) -> None: diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_linked_service.py deleted file mode 100644 index 9c1b8e4c3cbd..000000000000 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_linked_service.py +++ /dev/null @@ -1,94 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for -# license information. -# -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is -# regenerated. -# -------------------------------------------------------------------------- - -from .linked_service import LinkedService - - -class ResponsysLinkedService(LinkedService): - """Responsys linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] - :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.mgmt.datafactory.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.mgmt.datafactory.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the - Dataset. - :type annotations: list[object] - :param type: Required. Constant filled by server. - :type type: str - :param endpoint: Required. The endpoint of the Responsys server. - :type endpoint: object - :param client_id: Required. The client ID associated with the Responsys - application. Type: string (or Expression with resultType string). - :type client_id: object - :param client_secret: The client secret associated with the Responsys - application. Type: string (or Expression with resultType string). - :type client_secret: ~azure.mgmt.datafactory.models.SecretBase - :param use_encrypted_endpoints: Specifies whether the data source - endpoints are encrypted using HTTPS. The default value is true. Type: - boolean (or Expression with resultType boolean). - :type use_encrypted_endpoints: object - :param use_host_verification: Specifies whether to require the host name - in the server's certificate to match the host name of the server when - connecting over SSL. The default value is true. Type: boolean (or - Expression with resultType boolean). - :type use_host_verification: object - :param use_peer_verification: Specifies whether to verify the identity of - the server when connecting over SSL. The default value is true. Type: - boolean (or Expression with resultType boolean). - :type use_peer_verification: object - :param encrypted_credential: The encrypted credential used for - authentication. Credentials are encrypted using the integration runtime - credential manager. Type: string (or Expression with resultType string). - :type encrypted_credential: object - """ - - _validation = { - 'type': {'required': True}, - 'endpoint': {'required': True}, - 'client_id': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'type': {'key': 'type', 'type': 'str'}, - 'endpoint': {'key': 'typeProperties.endpoint', 'type': 'object'}, - 'client_id': {'key': 'typeProperties.clientId', 'type': 'object'}, - 'client_secret': {'key': 'typeProperties.clientSecret', 'type': 'SecretBase'}, - 'use_encrypted_endpoints': {'key': 'typeProperties.useEncryptedEndpoints', 'type': 'object'}, - 'use_host_verification': {'key': 'typeProperties.useHostVerification', 'type': 'object'}, - 'use_peer_verification': {'key': 'typeProperties.usePeerVerification', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__(self, **kwargs): - super(ResponsysLinkedService, self).__init__(**kwargs) - self.endpoint = kwargs.get('endpoint', None) - self.client_id = kwargs.get('client_id', None) - self.client_secret = kwargs.get('client_secret', None) - self.use_encrypted_endpoints = kwargs.get('use_encrypted_endpoints', None) - self.use_host_verification = kwargs.get('use_host_verification', None) - self.use_peer_verification = kwargs.get('use_peer_verification', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - self.type = 'Responsys' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_linked_service_py3.py deleted file mode 100644 index efad0984a99f..000000000000 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_linked_service_py3.py +++ /dev/null @@ -1,94 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for -# license information. -# -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is -# regenerated. -# -------------------------------------------------------------------------- - -from .linked_service import LinkedService - - -class ResponsysLinkedService(LinkedService): - """Responsys linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] - :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.mgmt.datafactory.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.mgmt.datafactory.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the - Dataset. - :type annotations: list[object] - :param type: Required. Constant filled by server. - :type type: str - :param endpoint: Required. The endpoint of the Responsys server. - :type endpoint: object - :param client_id: Required. The client ID associated with the Responsys - application. Type: string (or Expression with resultType string). - :type client_id: object - :param client_secret: The client secret associated with the Responsys - application. Type: string (or Expression with resultType string). - :type client_secret: ~azure.mgmt.datafactory.models.SecretBase - :param use_encrypted_endpoints: Specifies whether the data source - endpoints are encrypted using HTTPS. The default value is true. Type: - boolean (or Expression with resultType boolean). - :type use_encrypted_endpoints: object - :param use_host_verification: Specifies whether to require the host name - in the server's certificate to match the host name of the server when - connecting over SSL. The default value is true. Type: boolean (or - Expression with resultType boolean). - :type use_host_verification: object - :param use_peer_verification: Specifies whether to verify the identity of - the server when connecting over SSL. The default value is true. Type: - boolean (or Expression with resultType boolean). - :type use_peer_verification: object - :param encrypted_credential: The encrypted credential used for - authentication. Credentials are encrypted using the integration runtime - credential manager. Type: string (or Expression with resultType string). - :type encrypted_credential: object - """ - - _validation = { - 'type': {'required': True}, - 'endpoint': {'required': True}, - 'client_id': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'type': {'key': 'type', 'type': 'str'}, - 'endpoint': {'key': 'typeProperties.endpoint', 'type': 'object'}, - 'client_id': {'key': 'typeProperties.clientId', 'type': 'object'}, - 'client_secret': {'key': 'typeProperties.clientSecret', 'type': 'SecretBase'}, - 'use_encrypted_endpoints': {'key': 'typeProperties.useEncryptedEndpoints', 'type': 'object'}, - 'use_host_verification': {'key': 'typeProperties.useHostVerification', 'type': 'object'}, - 'use_peer_verification': {'key': 'typeProperties.usePeerVerification', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__(self, *, endpoint, client_id, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, client_secret=None, use_encrypted_endpoints=None, use_host_verification=None, use_peer_verification=None, encrypted_credential=None, **kwargs) -> None: - super(ResponsysLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.endpoint = endpoint - self.client_id = client_id - self.client_secret = client_secret - self.use_encrypted_endpoints = use_encrypted_endpoints - self.use_host_verification = use_host_verification - self.use_peer_verification = use_peer_verification - self.encrypted_credential = encrypted_credential - self.type = 'Responsys' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_object_dataset.py deleted file mode 100644 index eef47fbf10a8..000000000000 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_object_dataset.py +++ /dev/null @@ -1,58 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for -# license information. -# -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is -# regenerated. -# -------------------------------------------------------------------------- - -from .dataset import Dataset - - -class ResponsysObjectDataset(Dataset): - """Responsys dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: - array (or Expression with resultType array), itemType: DatasetDataElement. - :type structure: object - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.mgmt.datafactory.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.mgmt.datafactory.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the - Dataset. - :type annotations: list[object] - :param type: Required. Constant filled by server. - :type type: str - """ - - _validation = { - 'linked_service_name': {'required': True}, - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'type': {'key': 'type', 'type': 'str'}, - } - - def __init__(self, **kwargs): - super(ResponsysObjectDataset, self).__init__(**kwargs) - self.type = 'ResponsysObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_object_dataset_py3.py deleted file mode 100644 index e7b88bd52cec..000000000000 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_object_dataset_py3.py +++ /dev/null @@ -1,58 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for -# license information. -# -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is -# regenerated. -# -------------------------------------------------------------------------- - -from .dataset import Dataset - - -class ResponsysObjectDataset(Dataset): - """Responsys dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: - array (or Expression with resultType array), itemType: DatasetDataElement. - :type structure: object - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.mgmt.datafactory.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.mgmt.datafactory.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the - Dataset. - :type annotations: list[object] - :param type: Required. Constant filled by server. - :type type: str - """ - - _validation = { - 'linked_service_name': {'required': True}, - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'type': {'key': 'type', 'type': 'str'}, - } - - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: - super(ResponsysObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'ResponsysObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_source.py deleted file mode 100644 index 1e1a9397a6ba..000000000000 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_source.py +++ /dev/null @@ -1,52 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for -# license information. -# -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is -# regenerated. -# -------------------------------------------------------------------------- - -from .copy_source import CopySource - - -class ResponsysSource(CopySource): - """A copy activity Responsys source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] - :param source_retry_count: Source retry count. Type: integer (or - Expression with resultType integer). - :type source_retry_count: object - :param source_retry_wait: Source retry wait. Type: string (or Expression - with resultType string), pattern: - ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: object - :param type: Required. Constant filled by server. - :type type: str - :param query: A query to retrieve data from source. Type: string (or - Expression with resultType string). - :type query: object - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'type': {'key': 'type', 'type': 'str'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__(self, **kwargs): - super(ResponsysSource, self).__init__(**kwargs) - self.query = kwargs.get('query', None) - self.type = 'ResponsysSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_source_py3.py deleted file mode 100644 index 1934f830d515..000000000000 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_source_py3.py +++ /dev/null @@ -1,52 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for -# license information. -# -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is -# regenerated. -# -------------------------------------------------------------------------- - -from .copy_source import CopySource - - -class ResponsysSource(CopySource): - """A copy activity Responsys source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] - :param source_retry_count: Source retry count. Type: integer (or - Expression with resultType integer). - :type source_retry_count: object - :param source_retry_wait: Source retry wait. Type: string (or Expression - with resultType string), pattern: - ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: object - :param type: Required. Constant filled by server. - :type type: str - :param query: A query to retrieve data from source. Type: string (or - Expression with resultType string). - :type query: object - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'type': {'key': 'type', 'type': 'str'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__(self, *, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None, **kwargs) -> None: - super(ResponsysSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, **kwargs) - self.query = query - self.type = 'ResponsysSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/ssis_execution_parameter.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/ssis_execution_parameter.py new file mode 100644 index 000000000000..36f295c5a4aa --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/ssis_execution_parameter.py @@ -0,0 +1,35 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class SSISExecutionParameter(Model): + """SSIS execution parameter. + + All required parameters must be populated in order to send to Azure. + + :param value: Required. SSIS package execution parameter value. Type: + string (or Expression with resultType string). + :type value: object + """ + + _validation = { + 'value': {'required': True}, + } + + _attribute_map = { + 'value': {'key': 'value', 'type': 'object'}, + } + + def __init__(self, **kwargs): + super(SSISExecutionParameter, self).__init__(**kwargs) + self.value = kwargs.get('value', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/ssis_execution_parameter_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/ssis_execution_parameter_py3.py new file mode 100644 index 000000000000..cd10dd457a42 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/ssis_execution_parameter_py3.py @@ -0,0 +1,35 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class SSISExecutionParameter(Model): + """SSIS execution parameter. + + All required parameters must be populated in order to send to Azure. + + :param value: Required. SSIS package execution parameter value. Type: + string (or Expression with resultType string). + :type value: object + """ + + _validation = { + 'value': {'required': True}, + } + + _attribute_map = { + 'value': {'key': 'value', 'type': 'object'}, + } + + def __init__(self, *, value, **kwargs) -> None: + super(SSISExecutionParameter, self).__init__(**kwargs) + self.value = value diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/ssis_property_override.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/ssis_property_override.py new file mode 100644 index 000000000000..30b78594e6ab --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/ssis_property_override.py @@ -0,0 +1,40 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class SSISPropertyOverride(Model): + """SSIS property override. + + All required parameters must be populated in order to send to Azure. + + :param value: Required. SSIS package property override value. Type: string + (or Expression with resultType string). + :type value: object + :param is_sensitive: Whether SSIS package property override value is + sensitive data. Value will be encrypted in SSISDB if it is true + :type is_sensitive: bool + """ + + _validation = { + 'value': {'required': True}, + } + + _attribute_map = { + 'value': {'key': 'value', 'type': 'object'}, + 'is_sensitive': {'key': 'isSensitive', 'type': 'bool'}, + } + + def __init__(self, **kwargs): + super(SSISPropertyOverride, self).__init__(**kwargs) + self.value = kwargs.get('value', None) + self.is_sensitive = kwargs.get('is_sensitive', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/ssis_property_override_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/ssis_property_override_py3.py new file mode 100644 index 000000000000..b425a19adc7e --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/ssis_property_override_py3.py @@ -0,0 +1,40 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class SSISPropertyOverride(Model): + """SSIS property override. + + All required parameters must be populated in order to send to Azure. + + :param value: Required. SSIS package property override value. Type: string + (or Expression with resultType string). + :type value: object + :param is_sensitive: Whether SSIS package property override value is + sensitive data. Value will be encrypted in SSISDB if it is true + :type is_sensitive: bool + """ + + _validation = { + 'value': {'required': True}, + } + + _attribute_map = { + 'value': {'key': 'value', 'type': 'object'}, + 'is_sensitive': {'key': 'isSensitive', 'type': 'bool'}, + } + + def __init__(self, *, value, is_sensitive: bool=None, **kwargs) -> None: + super(SSISPropertyOverride, self).__init__(**kwargs) + self.value = value + self.is_sensitive = is_sensitive diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/version.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/version.py index 5a7feab42d26..a410d4adb270 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/version.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/version.py @@ -9,5 +9,5 @@ # regenerated. # -------------------------------------------------------------------------- -VERSION = "0.6.0" +VERSION = "2017-09-01-preview" From 9c3ebef3814709ee1cd3f13a0510340e458b21f3 Mon Sep 17 00:00:00 2001 From: Azure SDK for Python bot Date: Fri, 20 Apr 2018 10:13:13 -0700 Subject: [PATCH 03/54] =?UTF-8?q?[AutoPR=20datafactory/resource-manager]?= =?UTF-8?q?=20Change=20connection-string=20to=20be=20object=20and=20suppor?= =?UTF-8?q?t=20making=20severname,=20d=E2=80=A6=20(#2363)?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit * Generated from 5a4a7d356b6a096b96e0e20a050c20d22ab32150 Change connection-string to be object and support making severname, databasename visible to customers * Generated from a3865a4230e1cd415b7970b09d82083432fc2b9f Update description * Generated from a3865a4230e1cd415b7970b09d82083432fc2b9f Update description --- .../data_factory_management_client.py | 6 +- .../azure/mgmt/datafactory/models/__init__.py | 9 ++ .../azure_key_vault_secret_reference_py3.py | 2 +- .../models/azure_my_sql_linked_service.py | 7 +- .../models/azure_my_sql_linked_service_py3.py | 7 +- .../azure_postgre_sql_linked_service.py | 7 +- .../azure_postgre_sql_linked_service_py3.py | 7 +- .../azure_sql_database_linked_service.py | 7 +- .../azure_sql_database_linked_service_py3.py | 7 +- .../models/azure_sql_dw_linked_service.py | 8 +- .../models/azure_sql_dw_linked_service_py3.py | 8 +- .../models/azure_storage_linked_service.py | 7 +- .../azure_storage_linked_service_py3.py | 7 +- .../mgmt/datafactory/models/copy_source.py | 27 +++--- .../datafactory/models/copy_source_py3.py | 27 +++--- .../models/cosmos_db_linked_service.py | 7 +- .../models/cosmos_db_linked_service_py3.py | 7 +- .../models/couchbase_linked_service.py | 7 +- .../models/couchbase_linked_service_py3.py | 7 +- .../azure/mgmt/datafactory/models/dataset.py | 27 +++--- .../mgmt/datafactory/models/dataset_py3.py | 27 +++--- .../models/dataset_resource_py3.py | 2 +- .../models/drill_linked_service.py | 7 +- .../models/drill_linked_service_py3.py | 7 +- .../models/greenplum_linked_service.py | 7 +- .../models/greenplum_linked_service_py3.py | 7 +- .../integration_runtime_resource_py3.py | 2 +- .../linked_integration_runtime_key_py3.py | 2 +- .../linked_integration_runtime_rbac_py3.py | 2 +- .../mgmt/datafactory/models/linked_service.py | 4 +- .../datafactory/models/linked_service_py3.py | 4 +- .../models/linked_service_resource_py3.py | 2 +- .../models/maria_db_linked_service.py | 7 +- .../models/maria_db_linked_service_py3.py | 7 +- .../models/netezza_linked_service.py | 7 +- .../models/netezza_linked_service_py3.py | 7 +- .../datafactory/models/odbc_linked_service.py | 7 +- .../models/odbc_linked_service_py3.py | 7 +- .../models/oracle_linked_service.py | 7 +- .../models/oracle_linked_service_py3.py | 7 +- .../models/pipeline_resource_py3.py | 2 +- .../models/responsys_linked_service.py | 94 +++++++++++++++++++ .../models/responsys_linked_service_py3.py | 94 +++++++++++++++++++ .../models/responsys_object_dataset.py | 58 ++++++++++++ .../models/responsys_object_dataset_py3.py | 58 ++++++++++++ .../datafactory/models/responsys_source.py | 52 ++++++++++ .../models/responsys_source_py3.py | 52 ++++++++++ .../datafactory/models/secure_string_py3.py | 2 +- .../models/sql_server_linked_service.py | 7 +- .../models/sql_server_linked_service_py3.py | 7 +- .../models/trigger_resource_py3.py | 2 +- .../models/vertica_linked_service.py | 7 +- .../models/vertica_linked_service_py3.py | 7 +- .../azure/mgmt/datafactory/version.py | 2 +- 54 files changed, 612 insertions(+), 159 deletions(-) create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_linked_service.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_linked_service_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_object_dataset.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_object_dataset_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_source.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_source_py3.py diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/data_factory_management_client.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/data_factory_management_client.py index 97b5b70d4587..86680dcf55dc 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/data_factory_management_client.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/data_factory_management_client.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from msrest.service_client import ServiceClient +from msrest.service_client import SDKClient from msrest import Serializer, Deserializer from msrestazure import AzureConfiguration from .version import VERSION @@ -58,7 +58,7 @@ def __init__( self.subscription_id = subscription_id -class DataFactoryManagementClient(object): +class DataFactoryManagementClient(SDKClient): """The Azure Data Factory V2 management API provides a RESTful set of web services that interact with Azure Data Factory V2 services. :ivar config: Configuration for client. @@ -97,7 +97,7 @@ def __init__( self, credentials, subscription_id, base_url=None): self.config = DataFactoryManagementClientConfiguration(credentials, subscription_id, base_url) - self._client = ServiceClient(self.config.credentials, self.config) + super(DataFactoryManagementClient, self).__init__(self.config.credentials, self.config) client_models = {k: v for k, v in models.__dict__.items() if isinstance(v, type)} self.api_version = '2017-09-01-preview' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py index f25f49f6cb4a..21a24faf8a96 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py @@ -58,6 +58,7 @@ from .operation_service_specification_py3 import OperationServiceSpecification from .operation_py3 import Operation from .operation_list_response_py3 import OperationListResponse + from .responsys_linked_service_py3 import ResponsysLinkedService from .azure_databricks_linked_service_py3 import AzureDatabricksLinkedService from .azure_data_lake_analytics_linked_service_py3 import AzureDataLakeAnalyticsLinkedService from .hd_insight_on_demand_linked_service_py3 import HDInsightOnDemandLinkedService @@ -131,6 +132,7 @@ from .sql_server_linked_service_py3 import SqlServerLinkedService from .azure_sql_dw_linked_service_py3 import AzureSqlDWLinkedService from .azure_storage_linked_service_py3 import AzureStorageLinkedService + from .responsys_object_dataset_py3 import ResponsysObjectDataset from .salesforce_marketing_cloud_object_dataset_py3 import SalesforceMarketingCloudObjectDataset from .vertica_table_dataset_py3 import VerticaTableDataset from .netezza_table_dataset_py3 import NetezzaTableDataset @@ -213,6 +215,7 @@ from .web_activity_py3 import WebActivity from .redshift_unload_settings_py3 import RedshiftUnloadSettings from .amazon_redshift_source_py3 import AmazonRedshiftSource + from .responsys_source_py3 import ResponsysSource from .salesforce_marketing_cloud_source_py3 import SalesforceMarketingCloudSource from .vertica_source_py3 import VerticaSource from .netezza_source_py3 import NetezzaSource @@ -379,6 +382,7 @@ from .operation_service_specification import OperationServiceSpecification from .operation import Operation from .operation_list_response import OperationListResponse + from .responsys_linked_service import ResponsysLinkedService from .azure_databricks_linked_service import AzureDatabricksLinkedService from .azure_data_lake_analytics_linked_service import AzureDataLakeAnalyticsLinkedService from .hd_insight_on_demand_linked_service import HDInsightOnDemandLinkedService @@ -452,6 +456,7 @@ from .sql_server_linked_service import SqlServerLinkedService from .azure_sql_dw_linked_service import AzureSqlDWLinkedService from .azure_storage_linked_service import AzureStorageLinkedService + from .responsys_object_dataset import ResponsysObjectDataset from .salesforce_marketing_cloud_object_dataset import SalesforceMarketingCloudObjectDataset from .vertica_table_dataset import VerticaTableDataset from .netezza_table_dataset import NetezzaTableDataset @@ -534,6 +539,7 @@ from .web_activity import WebActivity from .redshift_unload_settings import RedshiftUnloadSettings from .amazon_redshift_source import AmazonRedshiftSource + from .responsys_source import ResponsysSource from .salesforce_marketing_cloud_source import SalesforceMarketingCloudSource from .vertica_source import VerticaSource from .netezza_source import NetezzaSource @@ -768,6 +774,7 @@ 'OperationServiceSpecification', 'Operation', 'OperationListResponse', + 'ResponsysLinkedService', 'AzureDatabricksLinkedService', 'AzureDataLakeAnalyticsLinkedService', 'HDInsightOnDemandLinkedService', @@ -841,6 +848,7 @@ 'SqlServerLinkedService', 'AzureSqlDWLinkedService', 'AzureStorageLinkedService', + 'ResponsysObjectDataset', 'SalesforceMarketingCloudObjectDataset', 'VerticaTableDataset', 'NetezzaTableDataset', @@ -923,6 +931,7 @@ 'WebActivity', 'RedshiftUnloadSettings', 'AmazonRedshiftSource', + 'ResponsysSource', 'SalesforceMarketingCloudSource', 'VerticaSource', 'NetezzaSource', diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_key_vault_secret_reference_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_key_vault_secret_reference_py3.py index 4ce7e7c32313..af03d495f3fd 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_key_vault_secret_reference_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_key_vault_secret_reference_py3.py @@ -44,7 +44,7 @@ class AzureKeyVaultSecretReference(SecretBase): } def __init__(self, *, store, secret_name, secret_version=None, **kwargs) -> None: - super(AzureKeyVaultSecretReference, self).__init__(, **kwargs) + super(AzureKeyVaultSecretReference, self).__init__(**kwargs) self.store = store self.secret_name = secret_name self.secret_version = secret_version diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_linked_service.py index 471ecb2e21f0..aa8f74d8ecda 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_linked_service.py @@ -33,8 +33,9 @@ class AzureMySqlLinkedService(LinkedService): :type annotations: list[object] :param type: Required. Constant filled by server. :type type: str - :param connection_string: Required. The connection string. - :type connection_string: ~azure.mgmt.datafactory.models.SecretBase + :param connection_string: Required. The connection string. Type: string, + SecureString or AzureKeyVaultSecretReference. + :type connection_string: object :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -53,7 +54,7 @@ class AzureMySqlLinkedService(LinkedService): 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, 'type': {'key': 'type', 'type': 'str'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'SecretBase'}, + 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_linked_service_py3.py index f6dde454fb27..f63266d2211e 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_linked_service_py3.py @@ -33,8 +33,9 @@ class AzureMySqlLinkedService(LinkedService): :type annotations: list[object] :param type: Required. Constant filled by server. :type type: str - :param connection_string: Required. The connection string. - :type connection_string: ~azure.mgmt.datafactory.models.SecretBase + :param connection_string: Required. The connection string. Type: string, + SecureString or AzureKeyVaultSecretReference. + :type connection_string: object :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -53,7 +54,7 @@ class AzureMySqlLinkedService(LinkedService): 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, 'type': {'key': 'type', 'type': 'str'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'SecretBase'}, + 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_linked_service.py index 62824fc83fcd..168ba5f589d7 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_linked_service.py @@ -33,8 +33,9 @@ class AzurePostgreSqlLinkedService(LinkedService): :type annotations: list[object] :param type: Required. Constant filled by server. :type type: str - :param connection_string: An ODBC connection string. - :type connection_string: ~azure.mgmt.datafactory.models.SecretBase + :param connection_string: An ODBC connection string. Type: string, + SecureString or AzureKeyVaultSecretReference. + :type connection_string: object :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -52,7 +53,7 @@ class AzurePostgreSqlLinkedService(LinkedService): 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, 'type': {'key': 'type', 'type': 'str'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'SecretBase'}, + 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_linked_service_py3.py index fe496511710a..85ccf186006a 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_linked_service_py3.py @@ -33,8 +33,9 @@ class AzurePostgreSqlLinkedService(LinkedService): :type annotations: list[object] :param type: Required. Constant filled by server. :type type: str - :param connection_string: An ODBC connection string. - :type connection_string: ~azure.mgmt.datafactory.models.SecretBase + :param connection_string: An ODBC connection string. Type: string, + SecureString or AzureKeyVaultSecretReference. + :type connection_string: object :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -52,7 +53,7 @@ class AzurePostgreSqlLinkedService(LinkedService): 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, 'type': {'key': 'type', 'type': 'str'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'SecretBase'}, + 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_database_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_database_linked_service.py index 250b004f00be..c1ecc189d194 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_database_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_database_linked_service.py @@ -33,8 +33,9 @@ class AzureSqlDatabaseLinkedService(LinkedService): :type annotations: list[object] :param type: Required. Constant filled by server. :type type: str - :param connection_string: Required. The connection string. - :type connection_string: ~azure.mgmt.datafactory.models.SecretBase + :param connection_string: Required. The connection string. Type: string, + SecureString or AzureKeyVaultSecretReference. + :type connection_string: object :param service_principal_id: The ID of the service principal used to authenticate against Azure SQL Database. Type: string (or Expression with resultType string). @@ -63,7 +64,7 @@ class AzureSqlDatabaseLinkedService(LinkedService): 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, 'type': {'key': 'type', 'type': 'str'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'SecretBase'}, + 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, 'service_principal_id': {'key': 'typeProperties.servicePrincipalId', 'type': 'object'}, 'service_principal_key': {'key': 'typeProperties.servicePrincipalKey', 'type': 'SecretBase'}, 'tenant': {'key': 'typeProperties.tenant', 'type': 'object'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_database_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_database_linked_service_py3.py index ffa86ded4b4e..f816df691d6d 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_database_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_database_linked_service_py3.py @@ -33,8 +33,9 @@ class AzureSqlDatabaseLinkedService(LinkedService): :type annotations: list[object] :param type: Required. Constant filled by server. :type type: str - :param connection_string: Required. The connection string. - :type connection_string: ~azure.mgmt.datafactory.models.SecretBase + :param connection_string: Required. The connection string. Type: string, + SecureString or AzureKeyVaultSecretReference. + :type connection_string: object :param service_principal_id: The ID of the service principal used to authenticate against Azure SQL Database. Type: string (or Expression with resultType string). @@ -63,7 +64,7 @@ class AzureSqlDatabaseLinkedService(LinkedService): 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, 'type': {'key': 'type', 'type': 'str'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'SecretBase'}, + 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, 'service_principal_id': {'key': 'typeProperties.servicePrincipalId', 'type': 'object'}, 'service_principal_key': {'key': 'typeProperties.servicePrincipalKey', 'type': 'SecretBase'}, 'tenant': {'key': 'typeProperties.tenant', 'type': 'object'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_dw_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_dw_linked_service.py index 989d8c799955..ef56c4142321 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_dw_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_dw_linked_service.py @@ -33,8 +33,10 @@ class AzureSqlDWLinkedService(LinkedService): :type annotations: list[object] :param type: Required. Constant filled by server. :type type: str - :param connection_string: Required. The connection string. - :type connection_string: ~azure.mgmt.datafactory.models.SecretBase + :param connection_string: Required. The connection string. Type: string, + SecureString or AzureKeyVaultSecretReference. Type: string, SecureString + or AzureKeyVaultSecretReference. + :type connection_string: object :param service_principal_id: The ID of the service principal used to authenticate against Azure SQL Data Warehouse. Type: string (or Expression with resultType string). @@ -63,7 +65,7 @@ class AzureSqlDWLinkedService(LinkedService): 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, 'type': {'key': 'type', 'type': 'str'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'SecretBase'}, + 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, 'service_principal_id': {'key': 'typeProperties.servicePrincipalId', 'type': 'object'}, 'service_principal_key': {'key': 'typeProperties.servicePrincipalKey', 'type': 'SecretBase'}, 'tenant': {'key': 'typeProperties.tenant', 'type': 'object'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_dw_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_dw_linked_service_py3.py index 6d0644a4d37d..4da9cfe20a49 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_dw_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_dw_linked_service_py3.py @@ -33,8 +33,10 @@ class AzureSqlDWLinkedService(LinkedService): :type annotations: list[object] :param type: Required. Constant filled by server. :type type: str - :param connection_string: Required. The connection string. - :type connection_string: ~azure.mgmt.datafactory.models.SecretBase + :param connection_string: Required. The connection string. Type: string, + SecureString or AzureKeyVaultSecretReference. Type: string, SecureString + or AzureKeyVaultSecretReference. + :type connection_string: object :param service_principal_id: The ID of the service principal used to authenticate against Azure SQL Data Warehouse. Type: string (or Expression with resultType string). @@ -63,7 +65,7 @@ class AzureSqlDWLinkedService(LinkedService): 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, 'type': {'key': 'type', 'type': 'str'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'SecretBase'}, + 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, 'service_principal_id': {'key': 'typeProperties.servicePrincipalId', 'type': 'object'}, 'service_principal_key': {'key': 'typeProperties.servicePrincipalKey', 'type': 'SecretBase'}, 'tenant': {'key': 'typeProperties.tenant', 'type': 'object'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_storage_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_storage_linked_service.py index bd78469cfcf5..d21da53dc341 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_storage_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_storage_linked_service.py @@ -34,8 +34,9 @@ class AzureStorageLinkedService(LinkedService): :param type: Required. Constant filled by server. :type type: str :param connection_string: The connection string. It is mutually exclusive - with sasUri property. - :type connection_string: ~azure.mgmt.datafactory.models.SecretBase + with sasUri property. Type: string, SecureString or + AzureKeyVaultSecretReference. + :type connection_string: object :param sas_uri: SAS URI of the Azure Storage resource. It is mutually exclusive with connectionString property. :type sas_uri: ~azure.mgmt.datafactory.models.SecretBase @@ -56,7 +57,7 @@ class AzureStorageLinkedService(LinkedService): 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, 'type': {'key': 'type', 'type': 'str'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'SecretBase'}, + 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, 'sas_uri': {'key': 'typeProperties.sasUri', 'type': 'SecretBase'}, 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_storage_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_storage_linked_service_py3.py index 2cbb53b69dcb..a09c83fabf9a 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_storage_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_storage_linked_service_py3.py @@ -34,8 +34,9 @@ class AzureStorageLinkedService(LinkedService): :param type: Required. Constant filled by server. :type type: str :param connection_string: The connection string. It is mutually exclusive - with sasUri property. - :type connection_string: ~azure.mgmt.datafactory.models.SecretBase + with sasUri property. Type: string, SecureString or + AzureKeyVaultSecretReference. + :type connection_string: object :param sas_uri: SAS URI of the Azure Storage resource. It is mutually exclusive with connectionString property. :type sas_uri: ~azure.mgmt.datafactory.models.SecretBase @@ -56,7 +57,7 @@ class AzureStorageLinkedService(LinkedService): 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, 'type': {'key': 'type', 'type': 'str'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'SecretBase'}, + 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, 'sas_uri': {'key': 'typeProperties.sasUri', 'type': 'SecretBase'}, 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_source.py index e7c55fbe9e9c..9a11107fc8e8 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_source.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_source.py @@ -16,18 +16,19 @@ class CopySource(Model): """A copy activity source. You probably want to use the sub-classes and not this class directly. Known - sub-classes are: AmazonRedshiftSource, SalesforceMarketingCloudSource, - VerticaSource, NetezzaSource, ZohoSource, XeroSource, SquareSource, - SparkSource, ShopifySource, ServiceNowSource, QuickBooksSource, - PrestoSource, PhoenixSource, PaypalSource, MarketoSource, MariaDBSource, - MagentoSource, JiraSource, ImpalaSource, HubspotSource, HiveSource, - HBaseSource, GreenplumSource, GoogleBigQuerySource, EloquaSource, - DrillSource, CouchbaseSource, ConcurSource, AzurePostgreSqlSource, - AmazonMWSSource, HttpSource, AzureDataLakeStoreSource, MongoDbSource, - CassandraSource, WebSource, OracleSource, AzureMySqlSource, HdfsSource, - FileSystemSource, SqlDWSource, SqlSource, SapEccSource, - SapCloudForCustomerSource, SalesforceSource, RelationalSource, - DynamicsSource, DocumentDbCollectionSource, BlobSource, AzureTableSource + sub-classes are: AmazonRedshiftSource, ResponsysSource, + SalesforceMarketingCloudSource, VerticaSource, NetezzaSource, ZohoSource, + XeroSource, SquareSource, SparkSource, ShopifySource, ServiceNowSource, + QuickBooksSource, PrestoSource, PhoenixSource, PaypalSource, MarketoSource, + MariaDBSource, MagentoSource, JiraSource, ImpalaSource, HubspotSource, + HiveSource, HBaseSource, GreenplumSource, GoogleBigQuerySource, + EloquaSource, DrillSource, CouchbaseSource, ConcurSource, + AzurePostgreSqlSource, AmazonMWSSource, HttpSource, + AzureDataLakeStoreSource, MongoDbSource, CassandraSource, WebSource, + OracleSource, AzureMySqlSource, HdfsSource, FileSystemSource, SqlDWSource, + SqlSource, SapEccSource, SapCloudForCustomerSource, SalesforceSource, + RelationalSource, DynamicsSource, DocumentDbCollectionSource, BlobSource, + AzureTableSource All required parameters must be populated in order to send to Azure. @@ -57,7 +58,7 @@ class CopySource(Model): } _subtype_map = { - 'type': {'AmazonRedshiftSource': 'AmazonRedshiftSource', 'SalesforceMarketingCloudSource': 'SalesforceMarketingCloudSource', 'VerticaSource': 'VerticaSource', 'NetezzaSource': 'NetezzaSource', 'ZohoSource': 'ZohoSource', 'XeroSource': 'XeroSource', 'SquareSource': 'SquareSource', 'SparkSource': 'SparkSource', 'ShopifySource': 'ShopifySource', 'ServiceNowSource': 'ServiceNowSource', 'QuickBooksSource': 'QuickBooksSource', 'PrestoSource': 'PrestoSource', 'PhoenixSource': 'PhoenixSource', 'PaypalSource': 'PaypalSource', 'MarketoSource': 'MarketoSource', 'MariaDBSource': 'MariaDBSource', 'MagentoSource': 'MagentoSource', 'JiraSource': 'JiraSource', 'ImpalaSource': 'ImpalaSource', 'HubspotSource': 'HubspotSource', 'HiveSource': 'HiveSource', 'HBaseSource': 'HBaseSource', 'GreenplumSource': 'GreenplumSource', 'GoogleBigQuerySource': 'GoogleBigQuerySource', 'EloquaSource': 'EloquaSource', 'DrillSource': 'DrillSource', 'CouchbaseSource': 'CouchbaseSource', 'ConcurSource': 'ConcurSource', 'AzurePostgreSqlSource': 'AzurePostgreSqlSource', 'AmazonMWSSource': 'AmazonMWSSource', 'HttpSource': 'HttpSource', 'AzureDataLakeStoreSource': 'AzureDataLakeStoreSource', 'MongoDbSource': 'MongoDbSource', 'CassandraSource': 'CassandraSource', 'WebSource': 'WebSource', 'OracleSource': 'OracleSource', 'AzureMySqlSource': 'AzureMySqlSource', 'HdfsSource': 'HdfsSource', 'FileSystemSource': 'FileSystemSource', 'SqlDWSource': 'SqlDWSource', 'SqlSource': 'SqlSource', 'SapEccSource': 'SapEccSource', 'SapCloudForCustomerSource': 'SapCloudForCustomerSource', 'SalesforceSource': 'SalesforceSource', 'RelationalSource': 'RelationalSource', 'DynamicsSource': 'DynamicsSource', 'DocumentDbCollectionSource': 'DocumentDbCollectionSource', 'BlobSource': 'BlobSource', 'AzureTableSource': 'AzureTableSource'} + 'type': {'AmazonRedshiftSource': 'AmazonRedshiftSource', 'ResponsysSource': 'ResponsysSource', 'SalesforceMarketingCloudSource': 'SalesforceMarketingCloudSource', 'VerticaSource': 'VerticaSource', 'NetezzaSource': 'NetezzaSource', 'ZohoSource': 'ZohoSource', 'XeroSource': 'XeroSource', 'SquareSource': 'SquareSource', 'SparkSource': 'SparkSource', 'ShopifySource': 'ShopifySource', 'ServiceNowSource': 'ServiceNowSource', 'QuickBooksSource': 'QuickBooksSource', 'PrestoSource': 'PrestoSource', 'PhoenixSource': 'PhoenixSource', 'PaypalSource': 'PaypalSource', 'MarketoSource': 'MarketoSource', 'MariaDBSource': 'MariaDBSource', 'MagentoSource': 'MagentoSource', 'JiraSource': 'JiraSource', 'ImpalaSource': 'ImpalaSource', 'HubspotSource': 'HubspotSource', 'HiveSource': 'HiveSource', 'HBaseSource': 'HBaseSource', 'GreenplumSource': 'GreenplumSource', 'GoogleBigQuerySource': 'GoogleBigQuerySource', 'EloquaSource': 'EloquaSource', 'DrillSource': 'DrillSource', 'CouchbaseSource': 'CouchbaseSource', 'ConcurSource': 'ConcurSource', 'AzurePostgreSqlSource': 'AzurePostgreSqlSource', 'AmazonMWSSource': 'AmazonMWSSource', 'HttpSource': 'HttpSource', 'AzureDataLakeStoreSource': 'AzureDataLakeStoreSource', 'MongoDbSource': 'MongoDbSource', 'CassandraSource': 'CassandraSource', 'WebSource': 'WebSource', 'OracleSource': 'OracleSource', 'AzureMySqlSource': 'AzureMySqlSource', 'HdfsSource': 'HdfsSource', 'FileSystemSource': 'FileSystemSource', 'SqlDWSource': 'SqlDWSource', 'SqlSource': 'SqlSource', 'SapEccSource': 'SapEccSource', 'SapCloudForCustomerSource': 'SapCloudForCustomerSource', 'SalesforceSource': 'SalesforceSource', 'RelationalSource': 'RelationalSource', 'DynamicsSource': 'DynamicsSource', 'DocumentDbCollectionSource': 'DocumentDbCollectionSource', 'BlobSource': 'BlobSource', 'AzureTableSource': 'AzureTableSource'} } def __init__(self, **kwargs): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_source_py3.py index f117bfbe7c3b..7c1a96b2897a 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_source_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_source_py3.py @@ -16,18 +16,19 @@ class CopySource(Model): """A copy activity source. You probably want to use the sub-classes and not this class directly. Known - sub-classes are: AmazonRedshiftSource, SalesforceMarketingCloudSource, - VerticaSource, NetezzaSource, ZohoSource, XeroSource, SquareSource, - SparkSource, ShopifySource, ServiceNowSource, QuickBooksSource, - PrestoSource, PhoenixSource, PaypalSource, MarketoSource, MariaDBSource, - MagentoSource, JiraSource, ImpalaSource, HubspotSource, HiveSource, - HBaseSource, GreenplumSource, GoogleBigQuerySource, EloquaSource, - DrillSource, CouchbaseSource, ConcurSource, AzurePostgreSqlSource, - AmazonMWSSource, HttpSource, AzureDataLakeStoreSource, MongoDbSource, - CassandraSource, WebSource, OracleSource, AzureMySqlSource, HdfsSource, - FileSystemSource, SqlDWSource, SqlSource, SapEccSource, - SapCloudForCustomerSource, SalesforceSource, RelationalSource, - DynamicsSource, DocumentDbCollectionSource, BlobSource, AzureTableSource + sub-classes are: AmazonRedshiftSource, ResponsysSource, + SalesforceMarketingCloudSource, VerticaSource, NetezzaSource, ZohoSource, + XeroSource, SquareSource, SparkSource, ShopifySource, ServiceNowSource, + QuickBooksSource, PrestoSource, PhoenixSource, PaypalSource, MarketoSource, + MariaDBSource, MagentoSource, JiraSource, ImpalaSource, HubspotSource, + HiveSource, HBaseSource, GreenplumSource, GoogleBigQuerySource, + EloquaSource, DrillSource, CouchbaseSource, ConcurSource, + AzurePostgreSqlSource, AmazonMWSSource, HttpSource, + AzureDataLakeStoreSource, MongoDbSource, CassandraSource, WebSource, + OracleSource, AzureMySqlSource, HdfsSource, FileSystemSource, SqlDWSource, + SqlSource, SapEccSource, SapCloudForCustomerSource, SalesforceSource, + RelationalSource, DynamicsSource, DocumentDbCollectionSource, BlobSource, + AzureTableSource All required parameters must be populated in order to send to Azure. @@ -57,7 +58,7 @@ class CopySource(Model): } _subtype_map = { - 'type': {'AmazonRedshiftSource': 'AmazonRedshiftSource', 'SalesforceMarketingCloudSource': 'SalesforceMarketingCloudSource', 'VerticaSource': 'VerticaSource', 'NetezzaSource': 'NetezzaSource', 'ZohoSource': 'ZohoSource', 'XeroSource': 'XeroSource', 'SquareSource': 'SquareSource', 'SparkSource': 'SparkSource', 'ShopifySource': 'ShopifySource', 'ServiceNowSource': 'ServiceNowSource', 'QuickBooksSource': 'QuickBooksSource', 'PrestoSource': 'PrestoSource', 'PhoenixSource': 'PhoenixSource', 'PaypalSource': 'PaypalSource', 'MarketoSource': 'MarketoSource', 'MariaDBSource': 'MariaDBSource', 'MagentoSource': 'MagentoSource', 'JiraSource': 'JiraSource', 'ImpalaSource': 'ImpalaSource', 'HubspotSource': 'HubspotSource', 'HiveSource': 'HiveSource', 'HBaseSource': 'HBaseSource', 'GreenplumSource': 'GreenplumSource', 'GoogleBigQuerySource': 'GoogleBigQuerySource', 'EloquaSource': 'EloquaSource', 'DrillSource': 'DrillSource', 'CouchbaseSource': 'CouchbaseSource', 'ConcurSource': 'ConcurSource', 'AzurePostgreSqlSource': 'AzurePostgreSqlSource', 'AmazonMWSSource': 'AmazonMWSSource', 'HttpSource': 'HttpSource', 'AzureDataLakeStoreSource': 'AzureDataLakeStoreSource', 'MongoDbSource': 'MongoDbSource', 'CassandraSource': 'CassandraSource', 'WebSource': 'WebSource', 'OracleSource': 'OracleSource', 'AzureMySqlSource': 'AzureMySqlSource', 'HdfsSource': 'HdfsSource', 'FileSystemSource': 'FileSystemSource', 'SqlDWSource': 'SqlDWSource', 'SqlSource': 'SqlSource', 'SapEccSource': 'SapEccSource', 'SapCloudForCustomerSource': 'SapCloudForCustomerSource', 'SalesforceSource': 'SalesforceSource', 'RelationalSource': 'RelationalSource', 'DynamicsSource': 'DynamicsSource', 'DocumentDbCollectionSource': 'DocumentDbCollectionSource', 'BlobSource': 'BlobSource', 'AzureTableSource': 'AzureTableSource'} + 'type': {'AmazonRedshiftSource': 'AmazonRedshiftSource', 'ResponsysSource': 'ResponsysSource', 'SalesforceMarketingCloudSource': 'SalesforceMarketingCloudSource', 'VerticaSource': 'VerticaSource', 'NetezzaSource': 'NetezzaSource', 'ZohoSource': 'ZohoSource', 'XeroSource': 'XeroSource', 'SquareSource': 'SquareSource', 'SparkSource': 'SparkSource', 'ShopifySource': 'ShopifySource', 'ServiceNowSource': 'ServiceNowSource', 'QuickBooksSource': 'QuickBooksSource', 'PrestoSource': 'PrestoSource', 'PhoenixSource': 'PhoenixSource', 'PaypalSource': 'PaypalSource', 'MarketoSource': 'MarketoSource', 'MariaDBSource': 'MariaDBSource', 'MagentoSource': 'MagentoSource', 'JiraSource': 'JiraSource', 'ImpalaSource': 'ImpalaSource', 'HubspotSource': 'HubspotSource', 'HiveSource': 'HiveSource', 'HBaseSource': 'HBaseSource', 'GreenplumSource': 'GreenplumSource', 'GoogleBigQuerySource': 'GoogleBigQuerySource', 'EloquaSource': 'EloquaSource', 'DrillSource': 'DrillSource', 'CouchbaseSource': 'CouchbaseSource', 'ConcurSource': 'ConcurSource', 'AzurePostgreSqlSource': 'AzurePostgreSqlSource', 'AmazonMWSSource': 'AmazonMWSSource', 'HttpSource': 'HttpSource', 'AzureDataLakeStoreSource': 'AzureDataLakeStoreSource', 'MongoDbSource': 'MongoDbSource', 'CassandraSource': 'CassandraSource', 'WebSource': 'WebSource', 'OracleSource': 'OracleSource', 'AzureMySqlSource': 'AzureMySqlSource', 'HdfsSource': 'HdfsSource', 'FileSystemSource': 'FileSystemSource', 'SqlDWSource': 'SqlDWSource', 'SqlSource': 'SqlSource', 'SapEccSource': 'SapEccSource', 'SapCloudForCustomerSource': 'SapCloudForCustomerSource', 'SalesforceSource': 'SalesforceSource', 'RelationalSource': 'RelationalSource', 'DynamicsSource': 'DynamicsSource', 'DocumentDbCollectionSource': 'DocumentDbCollectionSource', 'BlobSource': 'BlobSource', 'AzureTableSource': 'AzureTableSource'} } def __init__(self, *, additional_properties=None, source_retry_count=None, source_retry_wait=None, **kwargs) -> None: diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cosmos_db_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cosmos_db_linked_service.py index 1ca42b56e86d..98d6e9a81eea 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cosmos_db_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cosmos_db_linked_service.py @@ -33,8 +33,9 @@ class CosmosDbLinkedService(LinkedService): :type annotations: list[object] :param type: Required. Constant filled by server. :type type: str - :param connection_string: Required. The connection string. - :type connection_string: ~azure.mgmt.datafactory.models.SecretBase + :param connection_string: Required. The connection string. Type: string, + SecureString or AzureKeyVaultSecretReference. + :type connection_string: object :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -53,7 +54,7 @@ class CosmosDbLinkedService(LinkedService): 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, 'type': {'key': 'type', 'type': 'str'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'SecretBase'}, + 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cosmos_db_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cosmos_db_linked_service_py3.py index 8bf8ce4bf74a..b875cc98374c 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cosmos_db_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cosmos_db_linked_service_py3.py @@ -33,8 +33,9 @@ class CosmosDbLinkedService(LinkedService): :type annotations: list[object] :param type: Required. Constant filled by server. :type type: str - :param connection_string: Required. The connection string. - :type connection_string: ~azure.mgmt.datafactory.models.SecretBase + :param connection_string: Required. The connection string. Type: string, + SecureString or AzureKeyVaultSecretReference. + :type connection_string: object :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -53,7 +54,7 @@ class CosmosDbLinkedService(LinkedService): 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, 'type': {'key': 'type', 'type': 'str'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'SecretBase'}, + 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_linked_service.py index 63787690a7ab..7c8aceb401dd 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_linked_service.py @@ -33,8 +33,9 @@ class CouchbaseLinkedService(LinkedService): :type annotations: list[object] :param type: Required. Constant filled by server. :type type: str - :param connection_string: An ODBC connection string. - :type connection_string: ~azure.mgmt.datafactory.models.SecretBase + :param connection_string: An ODBC connection string. Type: string, + SecureString or AzureKeyVaultSecretReference. + :type connection_string: object :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -52,7 +53,7 @@ class CouchbaseLinkedService(LinkedService): 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, 'type': {'key': 'type', 'type': 'str'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'SecretBase'}, + 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_linked_service_py3.py index 3c53fb08cd4e..05af16e9f3d9 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_linked_service_py3.py @@ -33,8 +33,9 @@ class CouchbaseLinkedService(LinkedService): :type annotations: list[object] :param type: Required. Constant filled by server. :type type: str - :param connection_string: An ODBC connection string. - :type connection_string: ~azure.mgmt.datafactory.models.SecretBase + :param connection_string: An ODBC connection string. Type: string, + SecureString or AzureKeyVaultSecretReference. + :type connection_string: object :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -52,7 +53,7 @@ class CouchbaseLinkedService(LinkedService): 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, 'type': {'key': 'type', 'type': 'str'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'SecretBase'}, + 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset.py index 9fc6e4a433f6..4ebebafdda73 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset.py @@ -17,18 +17,19 @@ class Dataset(Model): data stores, such as tables, files, folders, and documents. You probably want to use the sub-classes and not this class directly. Known - sub-classes are: SalesforceMarketingCloudObjectDataset, - VerticaTableDataset, NetezzaTableDataset, ZohoObjectDataset, - XeroObjectDataset, SquareObjectDataset, SparkObjectDataset, - ShopifyObjectDataset, ServiceNowObjectDataset, QuickBooksObjectDataset, - PrestoObjectDataset, PhoenixObjectDataset, PaypalObjectDataset, - MarketoObjectDataset, MariaDBTableDataset, MagentoObjectDataset, - JiraObjectDataset, ImpalaObjectDataset, HubspotObjectDataset, - HiveObjectDataset, HBaseObjectDataset, GreenplumTableDataset, - GoogleBigQueryObjectDataset, EloquaObjectDataset, DrillTableDataset, - CouchbaseTableDataset, ConcurObjectDataset, AzurePostgreSqlTableDataset, - AmazonMWSObjectDataset, HttpDataset, AzureSearchIndexDataset, - WebTableDataset, SqlServerTableDataset, SapEccResourceDataset, + sub-classes are: ResponsysObjectDataset, + SalesforceMarketingCloudObjectDataset, VerticaTableDataset, + NetezzaTableDataset, ZohoObjectDataset, XeroObjectDataset, + SquareObjectDataset, SparkObjectDataset, ShopifyObjectDataset, + ServiceNowObjectDataset, QuickBooksObjectDataset, PrestoObjectDataset, + PhoenixObjectDataset, PaypalObjectDataset, MarketoObjectDataset, + MariaDBTableDataset, MagentoObjectDataset, JiraObjectDataset, + ImpalaObjectDataset, HubspotObjectDataset, HiveObjectDataset, + HBaseObjectDataset, GreenplumTableDataset, GoogleBigQueryObjectDataset, + EloquaObjectDataset, DrillTableDataset, CouchbaseTableDataset, + ConcurObjectDataset, AzurePostgreSqlTableDataset, AmazonMWSObjectDataset, + HttpDataset, AzureSearchIndexDataset, WebTableDataset, + SqlServerTableDataset, SapEccResourceDataset, SapCloudForCustomerResourceDataset, SalesforceObjectDataset, RelationalTableDataset, AzureMySqlTableDataset, OracleTableDataset, ODataResourceDataset, MongoDbCollectionDataset, FileShareDataset, @@ -76,7 +77,7 @@ class Dataset(Model): } _subtype_map = { - 'type': {'SalesforceMarketingCloudObject': 'SalesforceMarketingCloudObjectDataset', 'VerticaTable': 'VerticaTableDataset', 'NetezzaTable': 'NetezzaTableDataset', 'ZohoObject': 'ZohoObjectDataset', 'XeroObject': 'XeroObjectDataset', 'SquareObject': 'SquareObjectDataset', 'SparkObject': 'SparkObjectDataset', 'ShopifyObject': 'ShopifyObjectDataset', 'ServiceNowObject': 'ServiceNowObjectDataset', 'QuickBooksObject': 'QuickBooksObjectDataset', 'PrestoObject': 'PrestoObjectDataset', 'PhoenixObject': 'PhoenixObjectDataset', 'PaypalObject': 'PaypalObjectDataset', 'MarketoObject': 'MarketoObjectDataset', 'MariaDBTable': 'MariaDBTableDataset', 'MagentoObject': 'MagentoObjectDataset', 'JiraObject': 'JiraObjectDataset', 'ImpalaObject': 'ImpalaObjectDataset', 'HubspotObject': 'HubspotObjectDataset', 'HiveObject': 'HiveObjectDataset', 'HBaseObject': 'HBaseObjectDataset', 'GreenplumTable': 'GreenplumTableDataset', 'GoogleBigQueryObject': 'GoogleBigQueryObjectDataset', 'EloquaObject': 'EloquaObjectDataset', 'DrillTable': 'DrillTableDataset', 'CouchbaseTable': 'CouchbaseTableDataset', 'ConcurObject': 'ConcurObjectDataset', 'AzurePostgreSqlTable': 'AzurePostgreSqlTableDataset', 'AmazonMWSObject': 'AmazonMWSObjectDataset', 'HttpFile': 'HttpDataset', 'AzureSearchIndex': 'AzureSearchIndexDataset', 'WebTable': 'WebTableDataset', 'SqlServerTable': 'SqlServerTableDataset', 'SapEccResource': 'SapEccResourceDataset', 'SapCloudForCustomerResource': 'SapCloudForCustomerResourceDataset', 'SalesforceObject': 'SalesforceObjectDataset', 'RelationalTable': 'RelationalTableDataset', 'AzureMySqlTable': 'AzureMySqlTableDataset', 'OracleTable': 'OracleTableDataset', 'ODataResource': 'ODataResourceDataset', 'MongoDbCollection': 'MongoDbCollectionDataset', 'FileShare': 'FileShareDataset', 'AzureDataLakeStoreFile': 'AzureDataLakeStoreDataset', 'DynamicsEntity': 'DynamicsEntityDataset', 'DocumentDbCollection': 'DocumentDbCollectionDataset', 'CustomDataset': 'CustomDataset', 'CassandraTable': 'CassandraTableDataset', 'AzureSqlDWTable': 'AzureSqlDWTableDataset', 'AzureSqlTable': 'AzureSqlTableDataset', 'AzureTable': 'AzureTableDataset', 'AzureBlob': 'AzureBlobDataset', 'AmazonS3Object': 'AmazonS3Dataset'} + 'type': {'ResponsysObject': 'ResponsysObjectDataset', 'SalesforceMarketingCloudObject': 'SalesforceMarketingCloudObjectDataset', 'VerticaTable': 'VerticaTableDataset', 'NetezzaTable': 'NetezzaTableDataset', 'ZohoObject': 'ZohoObjectDataset', 'XeroObject': 'XeroObjectDataset', 'SquareObject': 'SquareObjectDataset', 'SparkObject': 'SparkObjectDataset', 'ShopifyObject': 'ShopifyObjectDataset', 'ServiceNowObject': 'ServiceNowObjectDataset', 'QuickBooksObject': 'QuickBooksObjectDataset', 'PrestoObject': 'PrestoObjectDataset', 'PhoenixObject': 'PhoenixObjectDataset', 'PaypalObject': 'PaypalObjectDataset', 'MarketoObject': 'MarketoObjectDataset', 'MariaDBTable': 'MariaDBTableDataset', 'MagentoObject': 'MagentoObjectDataset', 'JiraObject': 'JiraObjectDataset', 'ImpalaObject': 'ImpalaObjectDataset', 'HubspotObject': 'HubspotObjectDataset', 'HiveObject': 'HiveObjectDataset', 'HBaseObject': 'HBaseObjectDataset', 'GreenplumTable': 'GreenplumTableDataset', 'GoogleBigQueryObject': 'GoogleBigQueryObjectDataset', 'EloquaObject': 'EloquaObjectDataset', 'DrillTable': 'DrillTableDataset', 'CouchbaseTable': 'CouchbaseTableDataset', 'ConcurObject': 'ConcurObjectDataset', 'AzurePostgreSqlTable': 'AzurePostgreSqlTableDataset', 'AmazonMWSObject': 'AmazonMWSObjectDataset', 'HttpFile': 'HttpDataset', 'AzureSearchIndex': 'AzureSearchIndexDataset', 'WebTable': 'WebTableDataset', 'SqlServerTable': 'SqlServerTableDataset', 'SapEccResource': 'SapEccResourceDataset', 'SapCloudForCustomerResource': 'SapCloudForCustomerResourceDataset', 'SalesforceObject': 'SalesforceObjectDataset', 'RelationalTable': 'RelationalTableDataset', 'AzureMySqlTable': 'AzureMySqlTableDataset', 'OracleTable': 'OracleTableDataset', 'ODataResource': 'ODataResourceDataset', 'MongoDbCollection': 'MongoDbCollectionDataset', 'FileShare': 'FileShareDataset', 'AzureDataLakeStoreFile': 'AzureDataLakeStoreDataset', 'DynamicsEntity': 'DynamicsEntityDataset', 'DocumentDbCollection': 'DocumentDbCollectionDataset', 'CustomDataset': 'CustomDataset', 'CassandraTable': 'CassandraTableDataset', 'AzureSqlDWTable': 'AzureSqlDWTableDataset', 'AzureSqlTable': 'AzureSqlTableDataset', 'AzureTable': 'AzureTableDataset', 'AzureBlob': 'AzureBlobDataset', 'AmazonS3Object': 'AmazonS3Dataset'} } def __init__(self, **kwargs): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_py3.py index 18d3a0d74068..65f43811d7a7 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_py3.py @@ -17,18 +17,19 @@ class Dataset(Model): data stores, such as tables, files, folders, and documents. You probably want to use the sub-classes and not this class directly. Known - sub-classes are: SalesforceMarketingCloudObjectDataset, - VerticaTableDataset, NetezzaTableDataset, ZohoObjectDataset, - XeroObjectDataset, SquareObjectDataset, SparkObjectDataset, - ShopifyObjectDataset, ServiceNowObjectDataset, QuickBooksObjectDataset, - PrestoObjectDataset, PhoenixObjectDataset, PaypalObjectDataset, - MarketoObjectDataset, MariaDBTableDataset, MagentoObjectDataset, - JiraObjectDataset, ImpalaObjectDataset, HubspotObjectDataset, - HiveObjectDataset, HBaseObjectDataset, GreenplumTableDataset, - GoogleBigQueryObjectDataset, EloquaObjectDataset, DrillTableDataset, - CouchbaseTableDataset, ConcurObjectDataset, AzurePostgreSqlTableDataset, - AmazonMWSObjectDataset, HttpDataset, AzureSearchIndexDataset, - WebTableDataset, SqlServerTableDataset, SapEccResourceDataset, + sub-classes are: ResponsysObjectDataset, + SalesforceMarketingCloudObjectDataset, VerticaTableDataset, + NetezzaTableDataset, ZohoObjectDataset, XeroObjectDataset, + SquareObjectDataset, SparkObjectDataset, ShopifyObjectDataset, + ServiceNowObjectDataset, QuickBooksObjectDataset, PrestoObjectDataset, + PhoenixObjectDataset, PaypalObjectDataset, MarketoObjectDataset, + MariaDBTableDataset, MagentoObjectDataset, JiraObjectDataset, + ImpalaObjectDataset, HubspotObjectDataset, HiveObjectDataset, + HBaseObjectDataset, GreenplumTableDataset, GoogleBigQueryObjectDataset, + EloquaObjectDataset, DrillTableDataset, CouchbaseTableDataset, + ConcurObjectDataset, AzurePostgreSqlTableDataset, AmazonMWSObjectDataset, + HttpDataset, AzureSearchIndexDataset, WebTableDataset, + SqlServerTableDataset, SapEccResourceDataset, SapCloudForCustomerResourceDataset, SalesforceObjectDataset, RelationalTableDataset, AzureMySqlTableDataset, OracleTableDataset, ODataResourceDataset, MongoDbCollectionDataset, FileShareDataset, @@ -76,7 +77,7 @@ class Dataset(Model): } _subtype_map = { - 'type': {'SalesforceMarketingCloudObject': 'SalesforceMarketingCloudObjectDataset', 'VerticaTable': 'VerticaTableDataset', 'NetezzaTable': 'NetezzaTableDataset', 'ZohoObject': 'ZohoObjectDataset', 'XeroObject': 'XeroObjectDataset', 'SquareObject': 'SquareObjectDataset', 'SparkObject': 'SparkObjectDataset', 'ShopifyObject': 'ShopifyObjectDataset', 'ServiceNowObject': 'ServiceNowObjectDataset', 'QuickBooksObject': 'QuickBooksObjectDataset', 'PrestoObject': 'PrestoObjectDataset', 'PhoenixObject': 'PhoenixObjectDataset', 'PaypalObject': 'PaypalObjectDataset', 'MarketoObject': 'MarketoObjectDataset', 'MariaDBTable': 'MariaDBTableDataset', 'MagentoObject': 'MagentoObjectDataset', 'JiraObject': 'JiraObjectDataset', 'ImpalaObject': 'ImpalaObjectDataset', 'HubspotObject': 'HubspotObjectDataset', 'HiveObject': 'HiveObjectDataset', 'HBaseObject': 'HBaseObjectDataset', 'GreenplumTable': 'GreenplumTableDataset', 'GoogleBigQueryObject': 'GoogleBigQueryObjectDataset', 'EloquaObject': 'EloquaObjectDataset', 'DrillTable': 'DrillTableDataset', 'CouchbaseTable': 'CouchbaseTableDataset', 'ConcurObject': 'ConcurObjectDataset', 'AzurePostgreSqlTable': 'AzurePostgreSqlTableDataset', 'AmazonMWSObject': 'AmazonMWSObjectDataset', 'HttpFile': 'HttpDataset', 'AzureSearchIndex': 'AzureSearchIndexDataset', 'WebTable': 'WebTableDataset', 'SqlServerTable': 'SqlServerTableDataset', 'SapEccResource': 'SapEccResourceDataset', 'SapCloudForCustomerResource': 'SapCloudForCustomerResourceDataset', 'SalesforceObject': 'SalesforceObjectDataset', 'RelationalTable': 'RelationalTableDataset', 'AzureMySqlTable': 'AzureMySqlTableDataset', 'OracleTable': 'OracleTableDataset', 'ODataResource': 'ODataResourceDataset', 'MongoDbCollection': 'MongoDbCollectionDataset', 'FileShare': 'FileShareDataset', 'AzureDataLakeStoreFile': 'AzureDataLakeStoreDataset', 'DynamicsEntity': 'DynamicsEntityDataset', 'DocumentDbCollection': 'DocumentDbCollectionDataset', 'CustomDataset': 'CustomDataset', 'CassandraTable': 'CassandraTableDataset', 'AzureSqlDWTable': 'AzureSqlDWTableDataset', 'AzureSqlTable': 'AzureSqlTableDataset', 'AzureTable': 'AzureTableDataset', 'AzureBlob': 'AzureBlobDataset', 'AmazonS3Object': 'AmazonS3Dataset'} + 'type': {'ResponsysObject': 'ResponsysObjectDataset', 'SalesforceMarketingCloudObject': 'SalesforceMarketingCloudObjectDataset', 'VerticaTable': 'VerticaTableDataset', 'NetezzaTable': 'NetezzaTableDataset', 'ZohoObject': 'ZohoObjectDataset', 'XeroObject': 'XeroObjectDataset', 'SquareObject': 'SquareObjectDataset', 'SparkObject': 'SparkObjectDataset', 'ShopifyObject': 'ShopifyObjectDataset', 'ServiceNowObject': 'ServiceNowObjectDataset', 'QuickBooksObject': 'QuickBooksObjectDataset', 'PrestoObject': 'PrestoObjectDataset', 'PhoenixObject': 'PhoenixObjectDataset', 'PaypalObject': 'PaypalObjectDataset', 'MarketoObject': 'MarketoObjectDataset', 'MariaDBTable': 'MariaDBTableDataset', 'MagentoObject': 'MagentoObjectDataset', 'JiraObject': 'JiraObjectDataset', 'ImpalaObject': 'ImpalaObjectDataset', 'HubspotObject': 'HubspotObjectDataset', 'HiveObject': 'HiveObjectDataset', 'HBaseObject': 'HBaseObjectDataset', 'GreenplumTable': 'GreenplumTableDataset', 'GoogleBigQueryObject': 'GoogleBigQueryObjectDataset', 'EloquaObject': 'EloquaObjectDataset', 'DrillTable': 'DrillTableDataset', 'CouchbaseTable': 'CouchbaseTableDataset', 'ConcurObject': 'ConcurObjectDataset', 'AzurePostgreSqlTable': 'AzurePostgreSqlTableDataset', 'AmazonMWSObject': 'AmazonMWSObjectDataset', 'HttpFile': 'HttpDataset', 'AzureSearchIndex': 'AzureSearchIndexDataset', 'WebTable': 'WebTableDataset', 'SqlServerTable': 'SqlServerTableDataset', 'SapEccResource': 'SapEccResourceDataset', 'SapCloudForCustomerResource': 'SapCloudForCustomerResourceDataset', 'SalesforceObject': 'SalesforceObjectDataset', 'RelationalTable': 'RelationalTableDataset', 'AzureMySqlTable': 'AzureMySqlTableDataset', 'OracleTable': 'OracleTableDataset', 'ODataResource': 'ODataResourceDataset', 'MongoDbCollection': 'MongoDbCollectionDataset', 'FileShare': 'FileShareDataset', 'AzureDataLakeStoreFile': 'AzureDataLakeStoreDataset', 'DynamicsEntity': 'DynamicsEntityDataset', 'DocumentDbCollection': 'DocumentDbCollectionDataset', 'CustomDataset': 'CustomDataset', 'CassandraTable': 'CassandraTableDataset', 'AzureSqlDWTable': 'AzureSqlDWTableDataset', 'AzureSqlTable': 'AzureSqlTableDataset', 'AzureTable': 'AzureTableDataset', 'AzureBlob': 'AzureBlobDataset', 'AmazonS3Object': 'AmazonS3Dataset'} } def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_resource_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_resource_py3.py index 1e82ef161d47..a7b3d20e2ec6 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_resource_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_resource_py3.py @@ -49,5 +49,5 @@ class DatasetResource(SubResource): } def __init__(self, *, properties, **kwargs) -> None: - super(DatasetResource, self).__init__(, **kwargs) + super(DatasetResource, self).__init__(**kwargs) self.properties = properties diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_linked_service.py index 2cd19dd6fdd8..0ece6ef6ad05 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_linked_service.py @@ -33,8 +33,9 @@ class DrillLinkedService(LinkedService): :type annotations: list[object] :param type: Required. Constant filled by server. :type type: str - :param connection_string: An ODBC connection string. - :type connection_string: ~azure.mgmt.datafactory.models.SecretBase + :param connection_string: An ODBC connection string. Type: string, + SecureString or AzureKeyVaultSecretReference. + :type connection_string: object :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -52,7 +53,7 @@ class DrillLinkedService(LinkedService): 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, 'type': {'key': 'type', 'type': 'str'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'SecretBase'}, + 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_linked_service_py3.py index 7deba2b1e73d..5761efa8687a 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_linked_service_py3.py @@ -33,8 +33,9 @@ class DrillLinkedService(LinkedService): :type annotations: list[object] :param type: Required. Constant filled by server. :type type: str - :param connection_string: An ODBC connection string. - :type connection_string: ~azure.mgmt.datafactory.models.SecretBase + :param connection_string: An ODBC connection string. Type: string, + SecureString or AzureKeyVaultSecretReference. + :type connection_string: object :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -52,7 +53,7 @@ class DrillLinkedService(LinkedService): 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, 'type': {'key': 'type', 'type': 'str'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'SecretBase'}, + 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_linked_service.py index cedb1a371923..d7507545f9e5 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_linked_service.py @@ -33,8 +33,9 @@ class GreenplumLinkedService(LinkedService): :type annotations: list[object] :param type: Required. Constant filled by server. :type type: str - :param connection_string: An ODBC connection string. - :type connection_string: ~azure.mgmt.datafactory.models.SecretBase + :param connection_string: An ODBC connection string. Type: string, + SecureString or AzureKeyVaultSecretReference. + :type connection_string: object :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -52,7 +53,7 @@ class GreenplumLinkedService(LinkedService): 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, 'type': {'key': 'type', 'type': 'str'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'SecretBase'}, + 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_linked_service_py3.py index bf0d5a8f9150..f85c146c4fc3 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_linked_service_py3.py @@ -33,8 +33,9 @@ class GreenplumLinkedService(LinkedService): :type annotations: list[object] :param type: Required. Constant filled by server. :type type: str - :param connection_string: An ODBC connection string. - :type connection_string: ~azure.mgmt.datafactory.models.SecretBase + :param connection_string: An ODBC connection string. Type: string, + SecureString or AzureKeyVaultSecretReference. + :type connection_string: object :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -52,7 +53,7 @@ class GreenplumLinkedService(LinkedService): 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, 'type': {'key': 'type', 'type': 'str'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'SecretBase'}, + 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_resource_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_resource_py3.py index acf162de2562..c506360d6357 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_resource_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_resource_py3.py @@ -49,5 +49,5 @@ class IntegrationRuntimeResource(SubResource): } def __init__(self, *, properties, **kwargs) -> None: - super(IntegrationRuntimeResource, self).__init__(, **kwargs) + super(IntegrationRuntimeResource, self).__init__(**kwargs) self.properties = properties diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_key_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_key_py3.py index 3e8f4dca2051..9cc45cf95b54 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_key_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_key_py3.py @@ -34,6 +34,6 @@ class LinkedIntegrationRuntimeKey(LinkedIntegrationRuntimeProperties): } def __init__(self, *, key, **kwargs) -> None: - super(LinkedIntegrationRuntimeKey, self).__init__(, **kwargs) + super(LinkedIntegrationRuntimeKey, self).__init__(**kwargs) self.key = key self.authorization_type = 'Key' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_rbac_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_rbac_py3.py index 16c070d71c47..d06dd029b0f1 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_rbac_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_rbac_py3.py @@ -35,6 +35,6 @@ class LinkedIntegrationRuntimeRbac(LinkedIntegrationRuntimeProperties): } def __init__(self, *, resource_id: str, **kwargs) -> None: - super(LinkedIntegrationRuntimeRbac, self).__init__(, **kwargs) + super(LinkedIntegrationRuntimeRbac, self).__init__(**kwargs) self.resource_id = resource_id self.authorization_type = 'RBAC' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service.py index efb89ee0eacd..ecb6b42e145a 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service.py @@ -18,7 +18,7 @@ class LinkedService(Model): resource. You probably want to use the sub-classes and not this class directly. Known - sub-classes are: AzureDatabricksLinkedService, + sub-classes are: ResponsysLinkedService, AzureDatabricksLinkedService, AzureDataLakeAnalyticsLinkedService, HDInsightOnDemandLinkedService, SalesforceMarketingCloudLinkedService, NetezzaLinkedService, VerticaLinkedService, ZohoLinkedService, XeroLinkedService, @@ -79,7 +79,7 @@ class LinkedService(Model): } _subtype_map = { - 'type': {'AzureDatabricks': 'AzureDatabricksLinkedService', 'AzureDataLakeAnalytics': 'AzureDataLakeAnalyticsLinkedService', 'HDInsightOnDemand': 'HDInsightOnDemandLinkedService', 'SalesforceMarketingCloud': 'SalesforceMarketingCloudLinkedService', 'Netezza': 'NetezzaLinkedService', 'Vertica': 'VerticaLinkedService', 'Zoho': 'ZohoLinkedService', 'Xero': 'XeroLinkedService', 'Square': 'SquareLinkedService', 'Spark': 'SparkLinkedService', 'Shopify': 'ShopifyLinkedService', 'ServiceNow': 'ServiceNowLinkedService', 'QuickBooks': 'QuickBooksLinkedService', 'Presto': 'PrestoLinkedService', 'Phoenix': 'PhoenixLinkedService', 'Paypal': 'PaypalLinkedService', 'Marketo': 'MarketoLinkedService', 'MariaDB': 'MariaDBLinkedService', 'Magento': 'MagentoLinkedService', 'Jira': 'JiraLinkedService', 'Impala': 'ImpalaLinkedService', 'Hubspot': 'HubspotLinkedService', 'Hive': 'HiveLinkedService', 'HBase': 'HBaseLinkedService', 'Greenplum': 'GreenplumLinkedService', 'GoogleBigQuery': 'GoogleBigQueryLinkedService', 'Eloqua': 'EloquaLinkedService', 'Drill': 'DrillLinkedService', 'Couchbase': 'CouchbaseLinkedService', 'Concur': 'ConcurLinkedService', 'AzurePostgreSql': 'AzurePostgreSqlLinkedService', 'AmazonMWS': 'AmazonMWSLinkedService', 'SapHana': 'SapHanaLinkedService', 'SapBW': 'SapBWLinkedService', 'Sftp': 'SftpServerLinkedService', 'FtpServer': 'FtpServerLinkedService', 'HttpServer': 'HttpLinkedService', 'AzureSearch': 'AzureSearchLinkedService', 'CustomDataSource': 'CustomDataSourceLinkedService', 'AmazonRedshift': 'AmazonRedshiftLinkedService', 'AmazonS3': 'AmazonS3LinkedService', 'SapEcc': 'SapEccLinkedService', 'SapCloudForCustomer': 'SapCloudForCustomerLinkedService', 'Salesforce': 'SalesforceLinkedService', 'AzureDataLakeStore': 'AzureDataLakeStoreLinkedService', 'MongoDb': 'MongoDbLinkedService', 'Cassandra': 'CassandraLinkedService', 'Web': 'WebLinkedService', 'OData': 'ODataLinkedService', 'Hdfs': 'HdfsLinkedService', 'Odbc': 'OdbcLinkedService', 'AzureML': 'AzureMLLinkedService', 'Teradata': 'TeradataLinkedService', 'Db2': 'Db2LinkedService', 'Sybase': 'SybaseLinkedService', 'PostgreSql': 'PostgreSqlLinkedService', 'MySql': 'MySqlLinkedService', 'AzureMySql': 'AzureMySqlLinkedService', 'Oracle': 'OracleLinkedService', 'FileServer': 'FileServerLinkedService', 'HDInsight': 'HDInsightLinkedService', 'Dynamics': 'DynamicsLinkedService', 'CosmosDb': 'CosmosDbLinkedService', 'AzureKeyVault': 'AzureKeyVaultLinkedService', 'AzureBatch': 'AzureBatchLinkedService', 'AzureSqlDatabase': 'AzureSqlDatabaseLinkedService', 'SqlServer': 'SqlServerLinkedService', 'AzureSqlDW': 'AzureSqlDWLinkedService', 'AzureStorage': 'AzureStorageLinkedService'} + 'type': {'Responsys': 'ResponsysLinkedService', 'AzureDatabricks': 'AzureDatabricksLinkedService', 'AzureDataLakeAnalytics': 'AzureDataLakeAnalyticsLinkedService', 'HDInsightOnDemand': 'HDInsightOnDemandLinkedService', 'SalesforceMarketingCloud': 'SalesforceMarketingCloudLinkedService', 'Netezza': 'NetezzaLinkedService', 'Vertica': 'VerticaLinkedService', 'Zoho': 'ZohoLinkedService', 'Xero': 'XeroLinkedService', 'Square': 'SquareLinkedService', 'Spark': 'SparkLinkedService', 'Shopify': 'ShopifyLinkedService', 'ServiceNow': 'ServiceNowLinkedService', 'QuickBooks': 'QuickBooksLinkedService', 'Presto': 'PrestoLinkedService', 'Phoenix': 'PhoenixLinkedService', 'Paypal': 'PaypalLinkedService', 'Marketo': 'MarketoLinkedService', 'MariaDB': 'MariaDBLinkedService', 'Magento': 'MagentoLinkedService', 'Jira': 'JiraLinkedService', 'Impala': 'ImpalaLinkedService', 'Hubspot': 'HubspotLinkedService', 'Hive': 'HiveLinkedService', 'HBase': 'HBaseLinkedService', 'Greenplum': 'GreenplumLinkedService', 'GoogleBigQuery': 'GoogleBigQueryLinkedService', 'Eloqua': 'EloquaLinkedService', 'Drill': 'DrillLinkedService', 'Couchbase': 'CouchbaseLinkedService', 'Concur': 'ConcurLinkedService', 'AzurePostgreSql': 'AzurePostgreSqlLinkedService', 'AmazonMWS': 'AmazonMWSLinkedService', 'SapHana': 'SapHanaLinkedService', 'SapBW': 'SapBWLinkedService', 'Sftp': 'SftpServerLinkedService', 'FtpServer': 'FtpServerLinkedService', 'HttpServer': 'HttpLinkedService', 'AzureSearch': 'AzureSearchLinkedService', 'CustomDataSource': 'CustomDataSourceLinkedService', 'AmazonRedshift': 'AmazonRedshiftLinkedService', 'AmazonS3': 'AmazonS3LinkedService', 'SapEcc': 'SapEccLinkedService', 'SapCloudForCustomer': 'SapCloudForCustomerLinkedService', 'Salesforce': 'SalesforceLinkedService', 'AzureDataLakeStore': 'AzureDataLakeStoreLinkedService', 'MongoDb': 'MongoDbLinkedService', 'Cassandra': 'CassandraLinkedService', 'Web': 'WebLinkedService', 'OData': 'ODataLinkedService', 'Hdfs': 'HdfsLinkedService', 'Odbc': 'OdbcLinkedService', 'AzureML': 'AzureMLLinkedService', 'Teradata': 'TeradataLinkedService', 'Db2': 'Db2LinkedService', 'Sybase': 'SybaseLinkedService', 'PostgreSql': 'PostgreSqlLinkedService', 'MySql': 'MySqlLinkedService', 'AzureMySql': 'AzureMySqlLinkedService', 'Oracle': 'OracleLinkedService', 'FileServer': 'FileServerLinkedService', 'HDInsight': 'HDInsightLinkedService', 'Dynamics': 'DynamicsLinkedService', 'CosmosDb': 'CosmosDbLinkedService', 'AzureKeyVault': 'AzureKeyVaultLinkedService', 'AzureBatch': 'AzureBatchLinkedService', 'AzureSqlDatabase': 'AzureSqlDatabaseLinkedService', 'SqlServer': 'SqlServerLinkedService', 'AzureSqlDW': 'AzureSqlDWLinkedService', 'AzureStorage': 'AzureStorageLinkedService'} } def __init__(self, **kwargs): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service_py3.py index 202165ec4982..39119635cba5 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service_py3.py @@ -18,7 +18,7 @@ class LinkedService(Model): resource. You probably want to use the sub-classes and not this class directly. Known - sub-classes are: AzureDatabricksLinkedService, + sub-classes are: ResponsysLinkedService, AzureDatabricksLinkedService, AzureDataLakeAnalyticsLinkedService, HDInsightOnDemandLinkedService, SalesforceMarketingCloudLinkedService, NetezzaLinkedService, VerticaLinkedService, ZohoLinkedService, XeroLinkedService, @@ -79,7 +79,7 @@ class LinkedService(Model): } _subtype_map = { - 'type': {'AzureDatabricks': 'AzureDatabricksLinkedService', 'AzureDataLakeAnalytics': 'AzureDataLakeAnalyticsLinkedService', 'HDInsightOnDemand': 'HDInsightOnDemandLinkedService', 'SalesforceMarketingCloud': 'SalesforceMarketingCloudLinkedService', 'Netezza': 'NetezzaLinkedService', 'Vertica': 'VerticaLinkedService', 'Zoho': 'ZohoLinkedService', 'Xero': 'XeroLinkedService', 'Square': 'SquareLinkedService', 'Spark': 'SparkLinkedService', 'Shopify': 'ShopifyLinkedService', 'ServiceNow': 'ServiceNowLinkedService', 'QuickBooks': 'QuickBooksLinkedService', 'Presto': 'PrestoLinkedService', 'Phoenix': 'PhoenixLinkedService', 'Paypal': 'PaypalLinkedService', 'Marketo': 'MarketoLinkedService', 'MariaDB': 'MariaDBLinkedService', 'Magento': 'MagentoLinkedService', 'Jira': 'JiraLinkedService', 'Impala': 'ImpalaLinkedService', 'Hubspot': 'HubspotLinkedService', 'Hive': 'HiveLinkedService', 'HBase': 'HBaseLinkedService', 'Greenplum': 'GreenplumLinkedService', 'GoogleBigQuery': 'GoogleBigQueryLinkedService', 'Eloqua': 'EloquaLinkedService', 'Drill': 'DrillLinkedService', 'Couchbase': 'CouchbaseLinkedService', 'Concur': 'ConcurLinkedService', 'AzurePostgreSql': 'AzurePostgreSqlLinkedService', 'AmazonMWS': 'AmazonMWSLinkedService', 'SapHana': 'SapHanaLinkedService', 'SapBW': 'SapBWLinkedService', 'Sftp': 'SftpServerLinkedService', 'FtpServer': 'FtpServerLinkedService', 'HttpServer': 'HttpLinkedService', 'AzureSearch': 'AzureSearchLinkedService', 'CustomDataSource': 'CustomDataSourceLinkedService', 'AmazonRedshift': 'AmazonRedshiftLinkedService', 'AmazonS3': 'AmazonS3LinkedService', 'SapEcc': 'SapEccLinkedService', 'SapCloudForCustomer': 'SapCloudForCustomerLinkedService', 'Salesforce': 'SalesforceLinkedService', 'AzureDataLakeStore': 'AzureDataLakeStoreLinkedService', 'MongoDb': 'MongoDbLinkedService', 'Cassandra': 'CassandraLinkedService', 'Web': 'WebLinkedService', 'OData': 'ODataLinkedService', 'Hdfs': 'HdfsLinkedService', 'Odbc': 'OdbcLinkedService', 'AzureML': 'AzureMLLinkedService', 'Teradata': 'TeradataLinkedService', 'Db2': 'Db2LinkedService', 'Sybase': 'SybaseLinkedService', 'PostgreSql': 'PostgreSqlLinkedService', 'MySql': 'MySqlLinkedService', 'AzureMySql': 'AzureMySqlLinkedService', 'Oracle': 'OracleLinkedService', 'FileServer': 'FileServerLinkedService', 'HDInsight': 'HDInsightLinkedService', 'Dynamics': 'DynamicsLinkedService', 'CosmosDb': 'CosmosDbLinkedService', 'AzureKeyVault': 'AzureKeyVaultLinkedService', 'AzureBatch': 'AzureBatchLinkedService', 'AzureSqlDatabase': 'AzureSqlDatabaseLinkedService', 'SqlServer': 'SqlServerLinkedService', 'AzureSqlDW': 'AzureSqlDWLinkedService', 'AzureStorage': 'AzureStorageLinkedService'} + 'type': {'Responsys': 'ResponsysLinkedService', 'AzureDatabricks': 'AzureDatabricksLinkedService', 'AzureDataLakeAnalytics': 'AzureDataLakeAnalyticsLinkedService', 'HDInsightOnDemand': 'HDInsightOnDemandLinkedService', 'SalesforceMarketingCloud': 'SalesforceMarketingCloudLinkedService', 'Netezza': 'NetezzaLinkedService', 'Vertica': 'VerticaLinkedService', 'Zoho': 'ZohoLinkedService', 'Xero': 'XeroLinkedService', 'Square': 'SquareLinkedService', 'Spark': 'SparkLinkedService', 'Shopify': 'ShopifyLinkedService', 'ServiceNow': 'ServiceNowLinkedService', 'QuickBooks': 'QuickBooksLinkedService', 'Presto': 'PrestoLinkedService', 'Phoenix': 'PhoenixLinkedService', 'Paypal': 'PaypalLinkedService', 'Marketo': 'MarketoLinkedService', 'MariaDB': 'MariaDBLinkedService', 'Magento': 'MagentoLinkedService', 'Jira': 'JiraLinkedService', 'Impala': 'ImpalaLinkedService', 'Hubspot': 'HubspotLinkedService', 'Hive': 'HiveLinkedService', 'HBase': 'HBaseLinkedService', 'Greenplum': 'GreenplumLinkedService', 'GoogleBigQuery': 'GoogleBigQueryLinkedService', 'Eloqua': 'EloquaLinkedService', 'Drill': 'DrillLinkedService', 'Couchbase': 'CouchbaseLinkedService', 'Concur': 'ConcurLinkedService', 'AzurePostgreSql': 'AzurePostgreSqlLinkedService', 'AmazonMWS': 'AmazonMWSLinkedService', 'SapHana': 'SapHanaLinkedService', 'SapBW': 'SapBWLinkedService', 'Sftp': 'SftpServerLinkedService', 'FtpServer': 'FtpServerLinkedService', 'HttpServer': 'HttpLinkedService', 'AzureSearch': 'AzureSearchLinkedService', 'CustomDataSource': 'CustomDataSourceLinkedService', 'AmazonRedshift': 'AmazonRedshiftLinkedService', 'AmazonS3': 'AmazonS3LinkedService', 'SapEcc': 'SapEccLinkedService', 'SapCloudForCustomer': 'SapCloudForCustomerLinkedService', 'Salesforce': 'SalesforceLinkedService', 'AzureDataLakeStore': 'AzureDataLakeStoreLinkedService', 'MongoDb': 'MongoDbLinkedService', 'Cassandra': 'CassandraLinkedService', 'Web': 'WebLinkedService', 'OData': 'ODataLinkedService', 'Hdfs': 'HdfsLinkedService', 'Odbc': 'OdbcLinkedService', 'AzureML': 'AzureMLLinkedService', 'Teradata': 'TeradataLinkedService', 'Db2': 'Db2LinkedService', 'Sybase': 'SybaseLinkedService', 'PostgreSql': 'PostgreSqlLinkedService', 'MySql': 'MySqlLinkedService', 'AzureMySql': 'AzureMySqlLinkedService', 'Oracle': 'OracleLinkedService', 'FileServer': 'FileServerLinkedService', 'HDInsight': 'HDInsightLinkedService', 'Dynamics': 'DynamicsLinkedService', 'CosmosDb': 'CosmosDbLinkedService', 'AzureKeyVault': 'AzureKeyVaultLinkedService', 'AzureBatch': 'AzureBatchLinkedService', 'AzureSqlDatabase': 'AzureSqlDatabaseLinkedService', 'SqlServer': 'SqlServerLinkedService', 'AzureSqlDW': 'AzureSqlDWLinkedService', 'AzureStorage': 'AzureStorageLinkedService'} } def __init__(self, *, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, **kwargs) -> None: diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service_resource_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service_resource_py3.py index c3a20944a87d..b9ceff820b34 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service_resource_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service_resource_py3.py @@ -49,5 +49,5 @@ class LinkedServiceResource(SubResource): } def __init__(self, *, properties, **kwargs) -> None: - super(LinkedServiceResource, self).__init__(, **kwargs) + super(LinkedServiceResource, self).__init__(**kwargs) self.properties = properties diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_linked_service.py index 24c0b0eef87d..853df4dbd376 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_linked_service.py @@ -33,8 +33,9 @@ class MariaDBLinkedService(LinkedService): :type annotations: list[object] :param type: Required. Constant filled by server. :type type: str - :param connection_string: An ODBC connection string. - :type connection_string: ~azure.mgmt.datafactory.models.SecretBase + :param connection_string: An ODBC connection string. Type: string, + SecureString or AzureKeyVaultSecretReference. + :type connection_string: object :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -52,7 +53,7 @@ class MariaDBLinkedService(LinkedService): 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, 'type': {'key': 'type', 'type': 'str'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'SecretBase'}, + 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_linked_service_py3.py index 7a428d54120b..8cfa0f73c07a 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_linked_service_py3.py @@ -33,8 +33,9 @@ class MariaDBLinkedService(LinkedService): :type annotations: list[object] :param type: Required. Constant filled by server. :type type: str - :param connection_string: An ODBC connection string. - :type connection_string: ~azure.mgmt.datafactory.models.SecretBase + :param connection_string: An ODBC connection string. Type: string, + SecureString or AzureKeyVaultSecretReference. + :type connection_string: object :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -52,7 +53,7 @@ class MariaDBLinkedService(LinkedService): 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, 'type': {'key': 'type', 'type': 'str'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'SecretBase'}, + 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/netezza_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/netezza_linked_service.py index a848845b448e..1d7b72403b54 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/netezza_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/netezza_linked_service.py @@ -33,8 +33,9 @@ class NetezzaLinkedService(LinkedService): :type annotations: list[object] :param type: Required. Constant filled by server. :type type: str - :param connection_string: An ODBC connection string. - :type connection_string: ~azure.mgmt.datafactory.models.SecretBase + :param connection_string: An ODBC connection string. Type: string, + SecureString or AzureKeyVaultSecretReference. + :type connection_string: object :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -52,7 +53,7 @@ class NetezzaLinkedService(LinkedService): 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, 'type': {'key': 'type', 'type': 'str'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'SecretBase'}, + 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/netezza_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/netezza_linked_service_py3.py index 52b3cc6932a0..d4c47cdd3c35 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/netezza_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/netezza_linked_service_py3.py @@ -33,8 +33,9 @@ class NetezzaLinkedService(LinkedService): :type annotations: list[object] :param type: Required. Constant filled by server. :type type: str - :param connection_string: An ODBC connection string. - :type connection_string: ~azure.mgmt.datafactory.models.SecretBase + :param connection_string: An ODBC connection string. Type: string, + SecureString or AzureKeyVaultSecretReference. + :type connection_string: object :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -52,7 +53,7 @@ class NetezzaLinkedService(LinkedService): 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, 'type': {'key': 'type', 'type': 'str'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'SecretBase'}, + 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odbc_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odbc_linked_service.py index 4f04d86f8f8f..43559b76e0e0 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odbc_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odbc_linked_service.py @@ -34,8 +34,9 @@ class OdbcLinkedService(LinkedService): :param type: Required. Constant filled by server. :type type: str :param connection_string: Required. The non-access credential portion of - the connection string as well as an optional encrypted credential. - :type connection_string: ~azure.mgmt.datafactory.models.SecretBase + the connection string as well as an optional encrypted credential. Type: + string, SecureString or AzureKeyVaultSecretReference. + :type connection_string: object :param authentication_type: Type of authentication used to connect to the ODBC data store. Possible values are: Anonymous and Basic. Type: string (or Expression with resultType string). @@ -66,7 +67,7 @@ class OdbcLinkedService(LinkedService): 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, 'type': {'key': 'type', 'type': 'str'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'SecretBase'}, + 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'object'}, 'credential': {'key': 'typeProperties.credential', 'type': 'SecretBase'}, 'user_name': {'key': 'typeProperties.userName', 'type': 'object'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odbc_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odbc_linked_service_py3.py index 3d30c01d7397..2e7ab9a8244a 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odbc_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odbc_linked_service_py3.py @@ -34,8 +34,9 @@ class OdbcLinkedService(LinkedService): :param type: Required. Constant filled by server. :type type: str :param connection_string: Required. The non-access credential portion of - the connection string as well as an optional encrypted credential. - :type connection_string: ~azure.mgmt.datafactory.models.SecretBase + the connection string as well as an optional encrypted credential. Type: + string, SecureString or AzureKeyVaultSecretReference. + :type connection_string: object :param authentication_type: Type of authentication used to connect to the ODBC data store. Possible values are: Anonymous and Basic. Type: string (or Expression with resultType string). @@ -66,7 +67,7 @@ class OdbcLinkedService(LinkedService): 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, 'type': {'key': 'type', 'type': 'str'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'SecretBase'}, + 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'object'}, 'credential': {'key': 'typeProperties.credential', 'type': 'SecretBase'}, 'user_name': {'key': 'typeProperties.userName', 'type': 'object'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_linked_service.py index 9e78c4ef42c0..2aca7d0a7215 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_linked_service.py @@ -33,8 +33,9 @@ class OracleLinkedService(LinkedService): :type annotations: list[object] :param type: Required. Constant filled by server. :type type: str - :param connection_string: Required. The connection string. - :type connection_string: ~azure.mgmt.datafactory.models.SecretBase + :param connection_string: Required. The connection string. Type: string, + SecureString or AzureKeyVaultSecretReference. + :type connection_string: object :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -53,7 +54,7 @@ class OracleLinkedService(LinkedService): 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, 'type': {'key': 'type', 'type': 'str'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'SecretBase'}, + 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_linked_service_py3.py index 4072d25d329c..350e5d29412c 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_linked_service_py3.py @@ -33,8 +33,9 @@ class OracleLinkedService(LinkedService): :type annotations: list[object] :param type: Required. Constant filled by server. :type type: str - :param connection_string: Required. The connection string. - :type connection_string: ~azure.mgmt.datafactory.models.SecretBase + :param connection_string: Required. The connection string. Type: string, + SecureString or AzureKeyVaultSecretReference. + :type connection_string: object :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -53,7 +54,7 @@ class OracleLinkedService(LinkedService): 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, 'type': {'key': 'type', 'type': 'str'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'SecretBase'}, + 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_resource_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_resource_py3.py index 4da049f1f3b5..ae1be9a29b07 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_resource_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_resource_py3.py @@ -65,7 +65,7 @@ class PipelineResource(SubResource): } def __init__(self, *, additional_properties=None, description: str=None, activities=None, parameters=None, concurrency: int=None, annotations=None, **kwargs) -> None: - super(PipelineResource, self).__init__(, **kwargs) + super(PipelineResource, self).__init__(**kwargs) self.additional_properties = additional_properties self.description = description self.activities = activities diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_linked_service.py new file mode 100644 index 000000000000..9c1b8e4c3cbd --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_linked_service.py @@ -0,0 +1,94 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class ResponsysLinkedService(LinkedService): + """Responsys linked service. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param endpoint: Required. The endpoint of the Responsys server. + :type endpoint: object + :param client_id: Required. The client ID associated with the Responsys + application. Type: string (or Expression with resultType string). + :type client_id: object + :param client_secret: The client secret associated with the Responsys + application. Type: string (or Expression with resultType string). + :type client_secret: ~azure.mgmt.datafactory.models.SecretBase + :param use_encrypted_endpoints: Specifies whether the data source + endpoints are encrypted using HTTPS. The default value is true. Type: + boolean (or Expression with resultType boolean). + :type use_encrypted_endpoints: object + :param use_host_verification: Specifies whether to require the host name + in the server's certificate to match the host name of the server when + connecting over SSL. The default value is true. Type: boolean (or + Expression with resultType boolean). + :type use_host_verification: object + :param use_peer_verification: Specifies whether to verify the identity of + the server when connecting over SSL. The default value is true. Type: + boolean (or Expression with resultType boolean). + :type use_peer_verification: object + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'endpoint': {'required': True}, + 'client_id': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'endpoint': {'key': 'typeProperties.endpoint', 'type': 'object'}, + 'client_id': {'key': 'typeProperties.clientId', 'type': 'object'}, + 'client_secret': {'key': 'typeProperties.clientSecret', 'type': 'SecretBase'}, + 'use_encrypted_endpoints': {'key': 'typeProperties.useEncryptedEndpoints', 'type': 'object'}, + 'use_host_verification': {'key': 'typeProperties.useHostVerification', 'type': 'object'}, + 'use_peer_verification': {'key': 'typeProperties.usePeerVerification', 'type': 'object'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, **kwargs): + super(ResponsysLinkedService, self).__init__(**kwargs) + self.endpoint = kwargs.get('endpoint', None) + self.client_id = kwargs.get('client_id', None) + self.client_secret = kwargs.get('client_secret', None) + self.use_encrypted_endpoints = kwargs.get('use_encrypted_endpoints', None) + self.use_host_verification = kwargs.get('use_host_verification', None) + self.use_peer_verification = kwargs.get('use_peer_verification', None) + self.encrypted_credential = kwargs.get('encrypted_credential', None) + self.type = 'Responsys' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_linked_service_py3.py new file mode 100644 index 000000000000..efad0984a99f --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_linked_service_py3.py @@ -0,0 +1,94 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class ResponsysLinkedService(LinkedService): + """Responsys linked service. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param endpoint: Required. The endpoint of the Responsys server. + :type endpoint: object + :param client_id: Required. The client ID associated with the Responsys + application. Type: string (or Expression with resultType string). + :type client_id: object + :param client_secret: The client secret associated with the Responsys + application. Type: string (or Expression with resultType string). + :type client_secret: ~azure.mgmt.datafactory.models.SecretBase + :param use_encrypted_endpoints: Specifies whether the data source + endpoints are encrypted using HTTPS. The default value is true. Type: + boolean (or Expression with resultType boolean). + :type use_encrypted_endpoints: object + :param use_host_verification: Specifies whether to require the host name + in the server's certificate to match the host name of the server when + connecting over SSL. The default value is true. Type: boolean (or + Expression with resultType boolean). + :type use_host_verification: object + :param use_peer_verification: Specifies whether to verify the identity of + the server when connecting over SSL. The default value is true. Type: + boolean (or Expression with resultType boolean). + :type use_peer_verification: object + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'endpoint': {'required': True}, + 'client_id': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'endpoint': {'key': 'typeProperties.endpoint', 'type': 'object'}, + 'client_id': {'key': 'typeProperties.clientId', 'type': 'object'}, + 'client_secret': {'key': 'typeProperties.clientSecret', 'type': 'SecretBase'}, + 'use_encrypted_endpoints': {'key': 'typeProperties.useEncryptedEndpoints', 'type': 'object'}, + 'use_host_verification': {'key': 'typeProperties.useHostVerification', 'type': 'object'}, + 'use_peer_verification': {'key': 'typeProperties.usePeerVerification', 'type': 'object'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, *, endpoint, client_id, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, client_secret=None, use_encrypted_endpoints=None, use_host_verification=None, use_peer_verification=None, encrypted_credential=None, **kwargs) -> None: + super(ResponsysLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.endpoint = endpoint + self.client_id = client_id + self.client_secret = client_secret + self.use_encrypted_endpoints = use_encrypted_endpoints + self.use_host_verification = use_host_verification + self.use_peer_verification = use_peer_verification + self.encrypted_credential = encrypted_credential + self.type = 'Responsys' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_object_dataset.py new file mode 100644 index 000000000000..eef47fbf10a8 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_object_dataset.py @@ -0,0 +1,58 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class ResponsysObjectDataset(Dataset): + """Responsys dataset. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Required. Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, **kwargs): + super(ResponsysObjectDataset, self).__init__(**kwargs) + self.type = 'ResponsysObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_object_dataset_py3.py new file mode 100644 index 000000000000..e7b88bd52cec --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_object_dataset_py3.py @@ -0,0 +1,58 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class ResponsysObjectDataset(Dataset): + """Responsys dataset. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Required. Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: + super(ResponsysObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + self.type = 'ResponsysObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_source.py new file mode 100644 index 000000000000..1e1a9397a6ba --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_source.py @@ -0,0 +1,52 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class ResponsysSource(CopySource): + """A copy activity Responsys source. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Required. Constant filled by server. + :type type: str + :param query: A query to retrieve data from source. Type: string (or + Expression with resultType string). + :type query: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'query': {'key': 'query', 'type': 'object'}, + } + + def __init__(self, **kwargs): + super(ResponsysSource, self).__init__(**kwargs) + self.query = kwargs.get('query', None) + self.type = 'ResponsysSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_source_py3.py new file mode 100644 index 000000000000..1934f830d515 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_source_py3.py @@ -0,0 +1,52 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class ResponsysSource(CopySource): + """A copy activity Responsys source. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Required. Constant filled by server. + :type type: str + :param query: A query to retrieve data from source. Type: string (or + Expression with resultType string). + :type query: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'query': {'key': 'query', 'type': 'object'}, + } + + def __init__(self, *, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None, **kwargs) -> None: + super(ResponsysSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, **kwargs) + self.query = query + self.type = 'ResponsysSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/secure_string_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/secure_string_py3.py index 1ca20053acd1..87513c02573c 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/secure_string_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/secure_string_py3.py @@ -35,6 +35,6 @@ class SecureString(SecretBase): } def __init__(self, *, value: str, **kwargs) -> None: - super(SecureString, self).__init__(, **kwargs) + super(SecureString, self).__init__(**kwargs) self.value = value self.type = 'SecureString' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_linked_service.py index 5e6a25007f3f..36230c046278 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_linked_service.py @@ -33,8 +33,9 @@ class SqlServerLinkedService(LinkedService): :type annotations: list[object] :param type: Required. Constant filled by server. :type type: str - :param connection_string: Required. The connection string. - :type connection_string: ~azure.mgmt.datafactory.models.SecretBase + :param connection_string: Required. The connection string. Type: string, + SecureString or AzureKeyVaultSecretReference. + :type connection_string: object :param user_name: The on-premises Windows authentication user name. Type: string (or Expression with resultType string). :type user_name: object @@ -58,7 +59,7 @@ class SqlServerLinkedService(LinkedService): 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, 'type': {'key': 'type', 'type': 'str'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'SecretBase'}, + 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, 'user_name': {'key': 'typeProperties.userName', 'type': 'object'}, 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_linked_service_py3.py index de0bb00b1cbd..0ec79b1eefe7 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_linked_service_py3.py @@ -33,8 +33,9 @@ class SqlServerLinkedService(LinkedService): :type annotations: list[object] :param type: Required. Constant filled by server. :type type: str - :param connection_string: Required. The connection string. - :type connection_string: ~azure.mgmt.datafactory.models.SecretBase + :param connection_string: Required. The connection string. Type: string, + SecureString or AzureKeyVaultSecretReference. + :type connection_string: object :param user_name: The on-premises Windows authentication user name. Type: string (or Expression with resultType string). :type user_name: object @@ -58,7 +59,7 @@ class SqlServerLinkedService(LinkedService): 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, 'type': {'key': 'type', 'type': 'str'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'SecretBase'}, + 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, 'user_name': {'key': 'typeProperties.userName', 'type': 'object'}, 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger_resource_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger_resource_py3.py index e90d1a8010fb..232374e962c5 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger_resource_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger_resource_py3.py @@ -49,5 +49,5 @@ class TriggerResource(SubResource): } def __init__(self, *, properties, **kwargs) -> None: - super(TriggerResource, self).__init__(, **kwargs) + super(TriggerResource, self).__init__(**kwargs) self.properties = properties diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/vertica_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/vertica_linked_service.py index b6bb2c2d389d..d872a0d547cc 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/vertica_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/vertica_linked_service.py @@ -33,8 +33,9 @@ class VerticaLinkedService(LinkedService): :type annotations: list[object] :param type: Required. Constant filled by server. :type type: str - :param connection_string: An ODBC connection string. - :type connection_string: ~azure.mgmt.datafactory.models.SecretBase + :param connection_string: An ODBC connection string. Type: string, + SecureString or AzureKeyVaultSecretReference. + :type connection_string: object :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -52,7 +53,7 @@ class VerticaLinkedService(LinkedService): 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, 'type': {'key': 'type', 'type': 'str'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'SecretBase'}, + 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/vertica_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/vertica_linked_service_py3.py index 8162656a5e20..b3380cf3ccda 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/vertica_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/vertica_linked_service_py3.py @@ -33,8 +33,9 @@ class VerticaLinkedService(LinkedService): :type annotations: list[object] :param type: Required. Constant filled by server. :type type: str - :param connection_string: An ODBC connection string. - :type connection_string: ~azure.mgmt.datafactory.models.SecretBase + :param connection_string: An ODBC connection string. Type: string, + SecureString or AzureKeyVaultSecretReference. + :type connection_string: object :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -52,7 +53,7 @@ class VerticaLinkedService(LinkedService): 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, 'type': {'key': 'type', 'type': 'str'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'SecretBase'}, + 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/version.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/version.py index a410d4adb270..5a7feab42d26 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/version.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/version.py @@ -9,5 +9,5 @@ # regenerated. # -------------------------------------------------------------------------- -VERSION = "2017-09-01-preview" +VERSION = "0.6.0" From 716dbd3814401649ebdb94c579d60f60e7c4c15d Mon Sep 17 00:00:00 2001 From: Azure SDK for Python bot Date: Wed, 16 May 2018 12:57:00 -0700 Subject: [PATCH 04/54] [AutoPR datafactory/resource-manager] [DataFactory] blob events trigger (#2527) * Generated from ce3cce80a8ae6434caf934e7baa601ba36c41874 blob events trigger * Generated from aaf689f1bc6b98c0795c300871da64956f4a3479 fixes to events * Generated from 8f703fbbae01ca72cfc3d1025f774125c0cc3540 address review feedback * Generated from 764ab738e12c8a7d388ec9f216230f664a102681 address review feedback * Generated from af2af5124c1789a46dce1086f97b9e7a6e2b2928 address review feedback * Generated from 31a95abf24bd5a79c06d22c611e949e83d0d41c2 resolve blob path property issue --- .../azure/mgmt/datafactory/models/__init__.py | 5 ++ .../models/amazon_mws_linked_service_py3.py | 2 +- .../models/amazon_mws_object_dataset_py3.py | 2 +- .../models/amazon_mws_source_py3.py | 2 +- .../amazon_redshift_linked_service_py3.py | 2 +- .../models/amazon_redshift_source_py3.py | 2 +- .../models/amazon_s3_dataset_py3.py | 2 +- .../models/amazon_s3_linked_service_py3.py | 2 +- .../datafactory/models/avro_format_py3.py | 2 +- .../models/azure_batch_linked_service_py3.py | 2 +- .../models/azure_blob_dataset_py3.py | 2 +- ..._data_lake_analytics_linked_service_py3.py | 2 +- .../azure_data_lake_store_dataset_py3.py | 2 +- ...zure_data_lake_store_linked_service_py3.py | 2 +- .../models/azure_data_lake_store_sink_py3.py | 2 +- .../azure_data_lake_store_source_py3.py | 2 +- .../azure_databricks_linked_service_py3.py | 2 +- .../azure_key_vault_linked_service_py3.py | 2 +- .../azure_key_vault_secret_reference_py3.py | 2 +- .../azure_ml_batch_execution_activity_py3.py | 2 +- .../models/azure_ml_linked_service_py3.py | 2 +- .../azure_ml_update_resource_activity_py3.py | 2 +- .../models/azure_my_sql_linked_service_py3.py | 2 +- .../models/azure_my_sql_source_py3.py | 2 +- .../models/azure_my_sql_table_dataset_py3.py | 2 +- .../azure_postgre_sql_linked_service_py3.py | 2 +- .../models/azure_postgre_sql_source_py3.py | 2 +- .../azure_postgre_sql_table_dataset_py3.py | 2 +- .../models/azure_queue_sink_py3.py | 2 +- .../models/azure_search_index_dataset_py3.py | 2 +- .../models/azure_search_index_sink_py3.py | 2 +- .../models/azure_search_linked_service_py3.py | 2 +- .../azure_sql_database_linked_service_py3.py | 2 +- .../models/azure_sql_dw_linked_service_py3.py | 2 +- .../models/azure_sql_dw_table_dataset_py3.py | 2 +- .../models/azure_sql_table_dataset_py3.py | 2 +- .../azure_storage_linked_service_py3.py | 2 +- .../models/azure_table_dataset_py3.py | 2 +- .../models/azure_table_sink_py3.py | 2 +- .../models/azure_table_source_py3.py | 2 +- .../datafactory/models/blob_events_trigger.py | 81 +++++++++++++++++++ .../models/blob_events_trigger_py3.py | 81 +++++++++++++++++++ .../mgmt/datafactory/models/blob_sink_py3.py | 2 +- .../datafactory/models/blob_source_py3.py | 2 +- .../datafactory/models/blob_trigger_py3.py | 2 +- .../models/cassandra_linked_service_py3.py | 2 +- .../models/cassandra_source_py3.py | 2 +- .../models/cassandra_table_dataset_py3.py | 2 +- .../models/concur_linked_service_py3.py | 2 +- .../models/concur_object_dataset_py3.py | 2 +- .../datafactory/models/concur_source_py3.py | 2 +- .../models/control_activity_py3.py | 2 +- .../datafactory/models/copy_activity_py3.py | 2 +- .../models/cosmos_db_linked_service_py3.py | 2 +- .../models/couchbase_linked_service_py3.py | 2 +- .../models/couchbase_source_py3.py | 2 +- .../models/couchbase_table_dataset_py3.py | 2 +- .../datafactory/models/custom_activity_py3.py | 2 +- .../custom_data_source_linked_service_py3.py | 2 +- .../datafactory/models/custom_dataset_py3.py | 2 +- .../data_factory_management_client_enums.py | 6 ++ .../data_lake_analytics_usql_activity_py3.py | 2 +- .../databricks_notebook_activity_py3.py | 2 +- .../models/dataset_bzip2_compression_py3.py | 2 +- .../models/dataset_deflate_compression_py3.py | 2 +- .../models/dataset_gzip_compression_py3.py | 2 +- .../models/dataset_resource_py3.py | 2 +- .../dataset_zip_deflate_compression_py3.py | 2 +- .../datafactory/models/db2_linked_service.py | 5 -- .../models/db2_linked_service_py3.py | 9 +-- .../document_db_collection_dataset_py3.py | 2 +- .../models/document_db_collection_sink_py3.py | 2 +- .../document_db_collection_source_py3.py | 2 +- .../models/drill_linked_service_py3.py | 2 +- .../datafactory/models/drill_source_py3.py | 2 +- .../models/drill_table_dataset_py3.py | 2 +- .../models/dynamics_entity_dataset_py3.py | 2 +- .../models/dynamics_linked_service_py3.py | 2 +- .../datafactory/models/dynamics_sink_py3.py | 2 +- .../datafactory/models/dynamics_source_py3.py | 2 +- .../models/eloqua_linked_service_py3.py | 2 +- .../models/eloqua_object_dataset_py3.py | 2 +- .../datafactory/models/eloqua_source_py3.py | 2 +- .../models/execute_pipeline_activity_py3.py | 2 +- .../execute_ssis_package_activity_py3.py | 2 +- .../models/execution_activity_py3.py | 2 +- .../mgmt/datafactory/models/factory_py3.py | 2 +- .../models/file_server_linked_service_py3.py | 2 +- .../models/file_share_dataset_py3.py | 2 +- .../models/file_system_sink_py3.py | 2 +- .../models/file_system_source_py3.py | 2 +- .../datafactory/models/filter_activity_py3.py | 2 +- .../models/for_each_activity_py3.py | 2 +- .../models/ftp_server_linked_service_py3.py | 2 +- .../models/get_metadata_activity_py3.py | 2 +- .../google_big_query_linked_service_py3.py | 2 +- .../google_big_query_object_dataset_py3.py | 2 +- .../models/google_big_query_source_py3.py | 2 +- .../models/greenplum_linked_service_py3.py | 2 +- .../models/greenplum_source_py3.py | 2 +- .../models/greenplum_table_dataset_py3.py | 2 +- .../models/hbase_linked_service_py3.py | 2 +- .../models/hbase_object_dataset_py3.py | 2 +- .../datafactory/models/hbase_source_py3.py | 2 +- .../models/hd_insight_hive_activity_py3.py | 2 +- .../models/hd_insight_linked_service_py3.py | 2 +- .../hd_insight_map_reduce_activity_py3.py | 2 +- ...hd_insight_on_demand_linked_service_py3.py | 2 +- .../models/hd_insight_pig_activity_py3.py | 2 +- .../models/hd_insight_spark_activity_py3.py | 2 +- .../hd_insight_streaming_activity_py3.py | 2 +- .../models/hdfs_linked_service_py3.py | 2 +- .../datafactory/models/hdfs_source_py3.py | 2 +- .../models/hive_linked_service_py3.py | 2 +- .../models/hive_object_dataset_py3.py | 2 +- .../datafactory/models/hive_source_py3.py | 2 +- .../datafactory/models/http_dataset_py3.py | 2 +- .../models/http_linked_service_py3.py | 2 +- .../datafactory/models/http_source_py3.py | 2 +- .../models/hubspot_linked_service_py3.py | 2 +- .../models/hubspot_object_dataset_py3.py | 2 +- .../datafactory/models/hubspot_source_py3.py | 2 +- .../models/if_condition_activity_py3.py | 2 +- .../models/impala_linked_service_py3.py | 2 +- .../models/impala_object_dataset_py3.py | 2 +- .../datafactory/models/impala_source_py3.py | 2 +- .../integration_runtime_resource_py3.py | 2 +- .../models/jira_linked_service_py3.py | 2 +- .../models/jira_object_dataset_py3.py | 2 +- .../datafactory/models/jira_source_py3.py | 2 +- .../datafactory/models/json_format_py3.py | 2 +- .../linked_integration_runtime_key_py3.py | 2 +- .../linked_integration_runtime_rbac_py3.py | 2 +- .../models/linked_service_resource_py3.py | 2 +- .../datafactory/models/lookup_activity_py3.py | 2 +- .../models/magento_linked_service_py3.py | 2 +- .../models/magento_object_dataset_py3.py | 2 +- .../datafactory/models/magento_source_py3.py | 2 +- .../models/managed_integration_runtime_py3.py | 2 +- .../managed_integration_runtime_status_py3.py | 2 +- .../models/maria_db_linked_service_py3.py | 2 +- .../datafactory/models/maria_db_source_py3.py | 2 +- .../models/maria_db_table_dataset_py3.py | 2 +- .../models/marketo_linked_service_py3.py | 2 +- .../models/marketo_object_dataset_py3.py | 2 +- .../datafactory/models/marketo_source_py3.py | 2 +- .../models/mongo_db_collection_dataset_py3.py | 2 +- .../models/mongo_db_linked_service_py3.py | 2 +- .../datafactory/models/mongo_db_source_py3.py | 2 +- .../models/multiple_pipeline_trigger.py | 4 +- .../models/multiple_pipeline_trigger_py3.py | 6 +- .../models/my_sql_linked_service.py | 31 ++----- .../models/my_sql_linked_service_py3.py | 35 ++------ .../models/netezza_linked_service_py3.py | 2 +- .../datafactory/models/netezza_source_py3.py | 2 +- .../models/netezza_table_dataset_py3.py | 2 +- .../models/odata_linked_service_py3.py | 2 +- .../models/odata_resource_dataset_py3.py | 2 +- .../models/odbc_linked_service_py3.py | 2 +- .../mgmt/datafactory/models/odbc_sink_py3.py | 2 +- .../models/oracle_linked_service_py3.py | 2 +- .../datafactory/models/oracle_sink_py3.py | 2 +- .../datafactory/models/oracle_source_py3.py | 2 +- .../models/oracle_table_dataset_py3.py | 2 +- .../mgmt/datafactory/models/orc_format_py3.py | 2 +- .../datafactory/models/parquet_format_py3.py | 2 +- .../models/paypal_linked_service_py3.py | 2 +- .../models/paypal_object_dataset_py3.py | 2 +- .../datafactory/models/paypal_source_py3.py | 2 +- .../models/phoenix_linked_service_py3.py | 2 +- .../models/phoenix_object_dataset_py3.py | 2 +- .../datafactory/models/phoenix_source_py3.py | 2 +- .../models/pipeline_resource_py3.py | 2 +- .../models/postgre_sql_linked_service.py | 31 ++----- .../models/postgre_sql_linked_service_py3.py | 35 ++------ .../models/presto_linked_service_py3.py | 2 +- .../models/presto_object_dataset_py3.py | 2 +- .../datafactory/models/presto_source_py3.py | 2 +- .../models/quick_books_linked_service_py3.py | 2 +- .../models/quick_books_object_dataset_py3.py | 2 +- .../models/quick_books_source_py3.py | 2 +- .../models/relational_source_py3.py | 2 +- .../models/relational_table_dataset_py3.py | 2 +- .../models/responsys_linked_service_py3.py | 2 +- .../models/responsys_object_dataset_py3.py | 2 +- .../models/responsys_source_py3.py | 2 +- .../models/salesforce_linked_service_py3.py | 2 +- ...orce_marketing_cloud_linked_service_py3.py | 2 +- ...orce_marketing_cloud_object_dataset_py3.py | 2 +- .../salesforce_marketing_cloud_source_py3.py | 2 +- .../models/salesforce_object_dataset_py3.py | 2 +- .../datafactory/models/salesforce_sink_py3.py | 2 +- .../models/salesforce_source_py3.py | 2 +- .../models/sap_bw_linked_service_py3.py | 2 +- ...p_cloud_for_customer_linked_service_py3.py | 2 +- ...cloud_for_customer_resource_dataset_py3.py | 2 +- .../models/sap_cloud_for_customer_sink_py3.py | 2 +- .../sap_cloud_for_customer_source_py3.py | 2 +- .../models/sap_ecc_linked_service_py3.py | 2 +- .../models/sap_ecc_resource_dataset_py3.py | 2 +- .../datafactory/models/sap_ecc_source_py3.py | 2 +- .../models/sap_hana_linked_service_py3.py | 2 +- .../models/schedule_trigger_py3.py | 2 +- .../datafactory/models/secure_string_py3.py | 2 +- .../self_hosted_integration_runtime_py3.py | 2 +- ...f_hosted_integration_runtime_status_py3.py | 2 +- .../models/service_now_linked_service_py3.py | 2 +- .../models/service_now_object_dataset_py3.py | 2 +- .../models/service_now_source_py3.py | 2 +- .../models/sftp_server_linked_service_py3.py | 2 +- .../models/shopify_linked_service_py3.py | 2 +- .../models/shopify_object_dataset_py3.py | 2 +- .../datafactory/models/shopify_source_py3.py | 2 +- .../models/spark_linked_service_py3.py | 2 +- .../models/spark_object_dataset_py3.py | 2 +- .../datafactory/models/spark_source_py3.py | 2 +- .../datafactory/models/sql_dw_sink_py3.py | 2 +- .../datafactory/models/sql_dw_source_py3.py | 2 +- .../models/sql_server_linked_service_py3.py | 2 +- ...ql_server_stored_procedure_activity_py3.py | 2 +- .../models/sql_server_table_dataset_py3.py | 2 +- .../mgmt/datafactory/models/sql_sink_py3.py | 2 +- .../mgmt/datafactory/models/sql_source_py3.py | 2 +- .../models/square_linked_service_py3.py | 2 +- .../models/square_object_dataset_py3.py | 2 +- .../datafactory/models/square_source_py3.py | 2 +- .../models/sybase_linked_service_py3.py | 2 +- .../models/tabular_translator_py3.py | 2 +- .../models/teradata_linked_service.py | 5 -- .../models/teradata_linked_service_py3.py | 9 +-- .../datafactory/models/text_format_py3.py | 2 +- .../models/trigger_resource_py3.py | 2 +- .../models/tumbling_window_trigger_py3.py | 2 +- .../datafactory/models/until_activity_py3.py | 2 +- .../models/vertica_linked_service_py3.py | 2 +- .../datafactory/models/vertica_source_py3.py | 2 +- .../models/vertica_table_dataset_py3.py | 2 +- .../datafactory/models/wait_activity_py3.py | 2 +- .../datafactory/models/web_activity_py3.py | 2 +- .../web_anonymous_authentication_py3.py | 2 +- .../models/web_basic_authentication_py3.py | 2 +- ...b_client_certificate_authentication_py3.py | 2 +- .../models/web_linked_service_py3.py | 2 +- .../mgmt/datafactory/models/web_source_py3.py | 2 +- .../models/web_table_dataset_py3.py | 2 +- .../models/xero_linked_service_py3.py | 2 +- .../models/xero_object_dataset_py3.py | 2 +- .../datafactory/models/xero_source_py3.py | 2 +- .../models/zoho_linked_service_py3.py | 2 +- .../models/zoho_object_dataset_py3.py | 2 +- .../datafactory/models/zoho_source_py3.py | 2 +- 251 files changed, 443 insertions(+), 374 deletions(-) create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/blob_events_trigger.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/blob_events_trigger_py3.py diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py index 21a24faf8a96..fdd54957b36f 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py @@ -198,6 +198,7 @@ from .amazon_s3_dataset_py3 import AmazonS3Dataset from .retry_policy_py3 import RetryPolicy from .tumbling_window_trigger_py3 import TumblingWindowTrigger + from .blob_events_trigger_py3 import BlobEventsTrigger from .blob_trigger_py3 import BlobTrigger from .recurrence_schedule_occurrence_py3 import RecurrenceScheduleOccurrence from .recurrence_schedule_py3 import RecurrenceSchedule @@ -522,6 +523,7 @@ from .amazon_s3_dataset import AmazonS3Dataset from .retry_policy import RetryPolicy from .tumbling_window_trigger import TumblingWindowTrigger + from .blob_events_trigger import BlobEventsTrigger from .blob_trigger import BlobTrigger from .recurrence_schedule_occurrence import RecurrenceScheduleOccurrence from .recurrence_schedule import RecurrenceSchedule @@ -700,6 +702,7 @@ DatasetCompressionLevel, JsonFormatFilePattern, TumblingWindowFrequency, + BlobEventTypes, DayOfWeek, DaysOfWeek, RecurrenceFrequency, @@ -914,6 +917,7 @@ 'AmazonS3Dataset', 'RetryPolicy', 'TumblingWindowTrigger', + 'BlobEventsTrigger', 'BlobTrigger', 'RecurrenceScheduleOccurrence', 'RecurrenceSchedule', @@ -1091,6 +1095,7 @@ 'DatasetCompressionLevel', 'JsonFormatFilePattern', 'TumblingWindowFrequency', + 'BlobEventTypes', 'DayOfWeek', 'DaysOfWeek', 'RecurrenceFrequency', diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_linked_service_py3.py index e55cd710ca7c..e1a5ba11f560 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_linked_service_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .linked_service import LinkedService +from .linked_service_py3 import LinkedService class AmazonMWSLinkedService(LinkedService): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_object_dataset_py3.py index 980bc9445650..7250b12f0fb9 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_object_dataset_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .dataset import Dataset +from .dataset_py3 import Dataset class AmazonMWSObjectDataset(Dataset): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_source_py3.py index 524a047cc6fc..895281f9af51 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_source_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_source_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .copy_source import CopySource +from .copy_source_py3 import CopySource class AmazonMWSSource(CopySource): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_redshift_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_redshift_linked_service_py3.py index 1d6e7a098786..7912ad040946 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_redshift_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_redshift_linked_service_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .linked_service import LinkedService +from .linked_service_py3 import LinkedService class AmazonRedshiftLinkedService(LinkedService): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_redshift_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_redshift_source_py3.py index de43ec3bdff0..9542e56e4850 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_redshift_source_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_redshift_source_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .copy_source import CopySource +from .copy_source_py3 import CopySource class AmazonRedshiftSource(CopySource): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_s3_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_s3_dataset_py3.py index 3ca9f3ca5d86..b9cc4fc4782b 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_s3_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_s3_dataset_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .dataset import Dataset +from .dataset_py3 import Dataset class AmazonS3Dataset(Dataset): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_s3_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_s3_linked_service_py3.py index fca34c82fb88..044e8bc299cf 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_s3_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_s3_linked_service_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .linked_service import LinkedService +from .linked_service_py3 import LinkedService class AmazonS3LinkedService(LinkedService): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/avro_format_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/avro_format_py3.py index eb1f141852a9..35d459c4b2a6 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/avro_format_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/avro_format_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .dataset_storage_format import DatasetStorageFormat +from .dataset_storage_format_py3 import DatasetStorageFormat class AvroFormat(DatasetStorageFormat): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_batch_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_batch_linked_service_py3.py index 928dfecf2664..63724f76f13f 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_batch_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_batch_linked_service_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .linked_service import LinkedService +from .linked_service_py3 import LinkedService class AzureBatchLinkedService(LinkedService): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_dataset_py3.py index fb3265a92c6f..fe28cffcac03 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_dataset_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .dataset import Dataset +from .dataset_py3 import Dataset class AzureBlobDataset(Dataset): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_analytics_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_analytics_linked_service_py3.py index f5b32873f157..b6c4b993cae7 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_analytics_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_analytics_linked_service_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .linked_service import LinkedService +from .linked_service_py3 import LinkedService class AzureDataLakeAnalyticsLinkedService(LinkedService): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_dataset_py3.py index 0fcf6571644f..5053cbd9a7a6 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_dataset_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .dataset import Dataset +from .dataset_py3 import Dataset class AzureDataLakeStoreDataset(Dataset): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_linked_service_py3.py index 3244f5d35bb6..10e3b72e654e 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_linked_service_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .linked_service import LinkedService +from .linked_service_py3 import LinkedService class AzureDataLakeStoreLinkedService(LinkedService): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_sink_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_sink_py3.py index 1ea62762385f..449c7b0a2a3e 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_sink_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_sink_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .copy_sink import CopySink +from .copy_sink_py3 import CopySink class AzureDataLakeStoreSink(CopySink): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_source_py3.py index 36896f7f69e3..d228d787bff4 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_source_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_source_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .copy_source import CopySource +from .copy_source_py3 import CopySource class AzureDataLakeStoreSource(CopySource): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_databricks_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_databricks_linked_service_py3.py index f41395a54f74..f6113d0209b9 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_databricks_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_databricks_linked_service_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .linked_service import LinkedService +from .linked_service_py3 import LinkedService class AzureDatabricksLinkedService(LinkedService): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_key_vault_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_key_vault_linked_service_py3.py index 0f66d72b1d75..e13cf7fb527a 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_key_vault_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_key_vault_linked_service_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .linked_service import LinkedService +from .linked_service_py3 import LinkedService class AzureKeyVaultLinkedService(LinkedService): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_key_vault_secret_reference_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_key_vault_secret_reference_py3.py index af03d495f3fd..c5fe4c7afbd4 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_key_vault_secret_reference_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_key_vault_secret_reference_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .secret_base import SecretBase +from .secret_base_py3 import SecretBase class AzureKeyVaultSecretReference(SecretBase): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_batch_execution_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_batch_execution_activity_py3.py index 7dc7f9d3a154..67f0f605ca5d 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_batch_execution_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_batch_execution_activity_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .execution_activity import ExecutionActivity +from .execution_activity_py3 import ExecutionActivity class AzureMLBatchExecutionActivity(ExecutionActivity): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_linked_service_py3.py index 9bea24015ba7..0fff3cea9b8a 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_linked_service_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .linked_service import LinkedService +from .linked_service_py3 import LinkedService class AzureMLLinkedService(LinkedService): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_update_resource_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_update_resource_activity_py3.py index d035b2ac4049..ff0dce23464f 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_update_resource_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_update_resource_activity_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .execution_activity import ExecutionActivity +from .execution_activity_py3 import ExecutionActivity class AzureMLUpdateResourceActivity(ExecutionActivity): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_linked_service_py3.py index f63266d2211e..928a621ff74b 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_linked_service_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .linked_service import LinkedService +from .linked_service_py3 import LinkedService class AzureMySqlLinkedService(LinkedService): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_source_py3.py index aaa6377548ef..4e1d35981f78 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_source_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_source_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .copy_source import CopySource +from .copy_source_py3 import CopySource class AzureMySqlSource(CopySource): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_table_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_table_dataset_py3.py index cb8e662e5941..3807f7484878 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_table_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_table_dataset_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .dataset import Dataset +from .dataset_py3 import Dataset class AzureMySqlTableDataset(Dataset): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_linked_service_py3.py index 85ccf186006a..e64ab78696d4 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_linked_service_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .linked_service import LinkedService +from .linked_service_py3 import LinkedService class AzurePostgreSqlLinkedService(LinkedService): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_source_py3.py index 6dbc32ed38a7..2af53cf91da2 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_source_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_source_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .copy_source import CopySource +from .copy_source_py3 import CopySource class AzurePostgreSqlSource(CopySource): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_table_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_table_dataset_py3.py index 8aff95ae35f7..9f3a62d269f2 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_table_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_table_dataset_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .dataset import Dataset +from .dataset_py3 import Dataset class AzurePostgreSqlTableDataset(Dataset): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_queue_sink_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_queue_sink_py3.py index 4357c290185c..debc14c0c7e1 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_queue_sink_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_queue_sink_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .copy_sink import CopySink +from .copy_sink_py3 import CopySink class AzureQueueSink(CopySink): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_search_index_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_search_index_dataset_py3.py index f97335cba8ee..b949c1878a80 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_search_index_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_search_index_dataset_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .dataset import Dataset +from .dataset_py3 import Dataset class AzureSearchIndexDataset(Dataset): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_search_index_sink_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_search_index_sink_py3.py index 90ed08a6f910..9ed48b36a588 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_search_index_sink_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_search_index_sink_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .copy_sink import CopySink +from .copy_sink_py3 import CopySink class AzureSearchIndexSink(CopySink): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_search_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_search_linked_service_py3.py index 2cd8ef33be9f..6cc3cdc98b89 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_search_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_search_linked_service_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .linked_service import LinkedService +from .linked_service_py3 import LinkedService class AzureSearchLinkedService(LinkedService): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_database_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_database_linked_service_py3.py index f816df691d6d..80286301c6a8 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_database_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_database_linked_service_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .linked_service import LinkedService +from .linked_service_py3 import LinkedService class AzureSqlDatabaseLinkedService(LinkedService): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_dw_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_dw_linked_service_py3.py index 4da9cfe20a49..2f8d316510aa 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_dw_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_dw_linked_service_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .linked_service import LinkedService +from .linked_service_py3 import LinkedService class AzureSqlDWLinkedService(LinkedService): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_dw_table_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_dw_table_dataset_py3.py index 2a5040f3cdcc..145d409e2b76 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_dw_table_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_dw_table_dataset_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .dataset import Dataset +from .dataset_py3 import Dataset class AzureSqlDWTableDataset(Dataset): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_table_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_table_dataset_py3.py index 67d236b072fd..8a226a2e673b 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_table_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_table_dataset_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .dataset import Dataset +from .dataset_py3 import Dataset class AzureSqlTableDataset(Dataset): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_storage_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_storage_linked_service_py3.py index a09c83fabf9a..6852af1d13ac 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_storage_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_storage_linked_service_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .linked_service import LinkedService +from .linked_service_py3 import LinkedService class AzureStorageLinkedService(LinkedService): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_table_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_table_dataset_py3.py index 98956e68bc72..b0c1cd13e693 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_table_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_table_dataset_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .dataset import Dataset +from .dataset_py3 import Dataset class AzureTableDataset(Dataset): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_table_sink_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_table_sink_py3.py index 2ba396eead2f..630df4f1f606 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_table_sink_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_table_sink_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .copy_sink import CopySink +from .copy_sink_py3 import CopySink class AzureTableSink(CopySink): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_table_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_table_source_py3.py index 00c0fb93f3dc..30ca05775f27 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_table_source_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_table_source_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .copy_source import CopySource +from .copy_source_py3 import CopySource class AzureTableSource(CopySource): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/blob_events_trigger.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/blob_events_trigger.py new file mode 100644 index 000000000000..cb8647fc1577 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/blob_events_trigger.py @@ -0,0 +1,81 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .multiple_pipeline_trigger import MultiplePipelineTrigger + + +class BlobEventsTrigger(MultiplePipelineTrigger): + """Trigger that runs everytime a Blob event occurs. + + Variables are only populated by the server, and will be ignored when + sending a request. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param description: Trigger description. + :type description: str + :ivar runtime_state: Indicates if trigger is running or not. Updated when + Start/Stop APIs are called on the Trigger. Possible values include: + 'Started', 'Stopped', 'Disabled' + :vartype runtime_state: str or + ~azure.mgmt.datafactory.models.TriggerRuntimeState + :param type: Required. Constant filled by server. + :type type: str + :param pipelines: Pipelines that need to be started. + :type pipelines: + list[~azure.mgmt.datafactory.models.TriggerPipelineReference] + :param blob_path_begins_with: The blob path must begin with the pattern + provided for trigger to fire. For example, '/records/blobs/december/' will + only fire the trigger for blobs in the december folder under the records + container. At least one of these must be provided: blobPathBeginsWith, + blobPathEndsWith. + :type blob_path_begins_with: str + :param blob_path_ends_with: The blob path must end with the pattern + provided for trigger to fire. For example, 'december/boxes.csv' will only + fire the trigger for blobs named boxes in a december folder. At least one + of these must be provided: blobPathBeginsWith, blobPathEndsWith. + :type blob_path_ends_with: str + :param events: Required. The type of events that cause this trigger to + fire. + :type events: list[str or ~azure.mgmt.datafactory.models.BlobEventTypes] + :param scope: Required. The ARM resource ID of the Storage Account. + :type scope: str + """ + + _validation = { + 'runtime_state': {'readonly': True}, + 'type': {'required': True}, + 'events': {'required': True}, + 'scope': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'runtime_state': {'key': 'runtimeState', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'str'}, + 'pipelines': {'key': 'pipelines', 'type': '[TriggerPipelineReference]'}, + 'blob_path_begins_with': {'key': 'typeProperties.blobPathBeginsWith', 'type': 'str'}, + 'blob_path_ends_with': {'key': 'typeProperties.blobPathEndsWith', 'type': 'str'}, + 'events': {'key': 'typeProperties.events', 'type': '[str]'}, + 'scope': {'key': 'typeProperties.scope', 'type': 'str'}, + } + + def __init__(self, **kwargs): + super(BlobEventsTrigger, self).__init__(**kwargs) + self.blob_path_begins_with = kwargs.get('blob_path_begins_with', None) + self.blob_path_ends_with = kwargs.get('blob_path_ends_with', None) + self.events = kwargs.get('events', None) + self.scope = kwargs.get('scope', None) + self.type = 'BlobEventsTrigger' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/blob_events_trigger_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/blob_events_trigger_py3.py new file mode 100644 index 000000000000..150f9eed2012 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/blob_events_trigger_py3.py @@ -0,0 +1,81 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .multiple_pipeline_trigger_py3 import MultiplePipelineTrigger + + +class BlobEventsTrigger(MultiplePipelineTrigger): + """Trigger that runs everytime a Blob event occurs. + + Variables are only populated by the server, and will be ignored when + sending a request. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param description: Trigger description. + :type description: str + :ivar runtime_state: Indicates if trigger is running or not. Updated when + Start/Stop APIs are called on the Trigger. Possible values include: + 'Started', 'Stopped', 'Disabled' + :vartype runtime_state: str or + ~azure.mgmt.datafactory.models.TriggerRuntimeState + :param type: Required. Constant filled by server. + :type type: str + :param pipelines: Pipelines that need to be started. + :type pipelines: + list[~azure.mgmt.datafactory.models.TriggerPipelineReference] + :param blob_path_begins_with: The blob path must begin with the pattern + provided for trigger to fire. For example, '/records/blobs/december/' will + only fire the trigger for blobs in the december folder under the records + container. At least one of these must be provided: blobPathBeginsWith, + blobPathEndsWith. + :type blob_path_begins_with: str + :param blob_path_ends_with: The blob path must end with the pattern + provided for trigger to fire. For example, 'december/boxes.csv' will only + fire the trigger for blobs named boxes in a december folder. At least one + of these must be provided: blobPathBeginsWith, blobPathEndsWith. + :type blob_path_ends_with: str + :param events: Required. The type of events that cause this trigger to + fire. + :type events: list[str or ~azure.mgmt.datafactory.models.BlobEventTypes] + :param scope: Required. The ARM resource ID of the Storage Account. + :type scope: str + """ + + _validation = { + 'runtime_state': {'readonly': True}, + 'type': {'required': True}, + 'events': {'required': True}, + 'scope': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'runtime_state': {'key': 'runtimeState', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'str'}, + 'pipelines': {'key': 'pipelines', 'type': '[TriggerPipelineReference]'}, + 'blob_path_begins_with': {'key': 'typeProperties.blobPathBeginsWith', 'type': 'str'}, + 'blob_path_ends_with': {'key': 'typeProperties.blobPathEndsWith', 'type': 'str'}, + 'events': {'key': 'typeProperties.events', 'type': '[str]'}, + 'scope': {'key': 'typeProperties.scope', 'type': 'str'}, + } + + def __init__(self, *, events, scope: str, additional_properties=None, description: str=None, pipelines=None, blob_path_begins_with: str=None, blob_path_ends_with: str=None, **kwargs) -> None: + super(BlobEventsTrigger, self).__init__(additional_properties=additional_properties, description=description, pipelines=pipelines, **kwargs) + self.blob_path_begins_with = blob_path_begins_with + self.blob_path_ends_with = blob_path_ends_with + self.events = events + self.scope = scope + self.type = 'BlobEventsTrigger' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/blob_sink_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/blob_sink_py3.py index 7d8d3dc642fd..1d6ac96aff6e 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/blob_sink_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/blob_sink_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .copy_sink import CopySink +from .copy_sink_py3 import CopySink class BlobSink(CopySink): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/blob_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/blob_source_py3.py index 8c442e8f408f..5b9dc775f069 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/blob_source_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/blob_source_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .copy_source import CopySource +from .copy_source_py3 import CopySource class BlobSource(CopySource): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/blob_trigger_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/blob_trigger_py3.py index f7123fbcf9bd..cf53b39ba72f 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/blob_trigger_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/blob_trigger_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .multiple_pipeline_trigger import MultiplePipelineTrigger +from .multiple_pipeline_trigger_py3 import MultiplePipelineTrigger class BlobTrigger(MultiplePipelineTrigger): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cassandra_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cassandra_linked_service_py3.py index 43a1e607b58f..dbc74f10002f 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cassandra_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cassandra_linked_service_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .linked_service import LinkedService +from .linked_service_py3 import LinkedService class CassandraLinkedService(LinkedService): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cassandra_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cassandra_source_py3.py index 6c62419c1e29..323d85d1e742 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cassandra_source_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cassandra_source_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .copy_source import CopySource +from .copy_source_py3 import CopySource class CassandraSource(CopySource): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cassandra_table_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cassandra_table_dataset_py3.py index e8932fe02571..6927e69a6e0d 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cassandra_table_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cassandra_table_dataset_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .dataset import Dataset +from .dataset_py3 import Dataset class CassandraTableDataset(Dataset): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_linked_service_py3.py index ba57ab2874c4..790e6bc76b76 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_linked_service_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .linked_service import LinkedService +from .linked_service_py3 import LinkedService class ConcurLinkedService(LinkedService): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_object_dataset_py3.py index 50b5325f3656..b77695bdc037 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_object_dataset_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .dataset import Dataset +from .dataset_py3 import Dataset class ConcurObjectDataset(Dataset): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_source_py3.py index d019deb98e5a..5205786a4585 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_source_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_source_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .copy_source import CopySource +from .copy_source_py3 import CopySource class ConcurSource(CopySource): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/control_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/control_activity_py3.py index 72eb7caec9fb..8623e8d151e2 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/control_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/control_activity_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .activity import Activity +from .activity_py3 import Activity class ControlActivity(Activity): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_activity_py3.py index 6da5626be374..30dd29e42139 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_activity_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .execution_activity import ExecutionActivity +from .execution_activity_py3 import ExecutionActivity class CopyActivity(ExecutionActivity): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cosmos_db_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cosmos_db_linked_service_py3.py index b875cc98374c..a338f3dc093b 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cosmos_db_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cosmos_db_linked_service_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .linked_service import LinkedService +from .linked_service_py3 import LinkedService class CosmosDbLinkedService(LinkedService): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_linked_service_py3.py index 05af16e9f3d9..7f94dca390eb 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_linked_service_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .linked_service import LinkedService +from .linked_service_py3 import LinkedService class CouchbaseLinkedService(LinkedService): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_source_py3.py index 21535b3f1987..cc661253a13d 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_source_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_source_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .copy_source import CopySource +from .copy_source_py3 import CopySource class CouchbaseSource(CopySource): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_table_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_table_dataset_py3.py index 99f9d2e8d218..aab909106680 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_table_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_table_dataset_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .dataset import Dataset +from .dataset_py3 import Dataset class CouchbaseTableDataset(Dataset): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_activity_py3.py index 65520468a96c..10542a55610f 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_activity_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .execution_activity import ExecutionActivity +from .execution_activity_py3 import ExecutionActivity class CustomActivity(ExecutionActivity): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_data_source_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_data_source_linked_service_py3.py index d11e5f8f8076..2ec05f7a32d9 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_data_source_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_data_source_linked_service_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .linked_service import LinkedService +from .linked_service_py3 import LinkedService class CustomDataSourceLinkedService(LinkedService): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_dataset_py3.py index e8c97da85a2d..8070049fa8cd 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_dataset_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .dataset import Dataset +from .dataset_py3 import Dataset class CustomDataset(Dataset): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_factory_management_client_enums.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_factory_management_client_enums.py index cb828e14276b..fb948ae66a19 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_factory_management_client_enums.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_factory_management_client_enums.py @@ -248,6 +248,12 @@ class TumblingWindowFrequency(str, Enum): hour = "Hour" +class BlobEventTypes(str, Enum): + + microsoft_storage_blob_created = "Microsoft.Storage.BlobCreated" + microsoft_storage_blob_deleted = "Microsoft.Storage.BlobDeleted" + + class DayOfWeek(str, Enum): sunday = "Sunday" diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_lake_analytics_usql_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_lake_analytics_usql_activity_py3.py index 022ec770081d..3b0ff6dbe94f 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_lake_analytics_usql_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_lake_analytics_usql_activity_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .execution_activity import ExecutionActivity +from .execution_activity_py3 import ExecutionActivity class DataLakeAnalyticsUSQLActivity(ExecutionActivity): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_notebook_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_notebook_activity_py3.py index a2beeb2fe9dd..cefdc1681830 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_notebook_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_notebook_activity_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .execution_activity import ExecutionActivity +from .execution_activity_py3 import ExecutionActivity class DatabricksNotebookActivity(ExecutionActivity): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_bzip2_compression_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_bzip2_compression_py3.py index fbf34be77561..f97af4588e0a 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_bzip2_compression_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_bzip2_compression_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .dataset_compression import DatasetCompression +from .dataset_compression_py3 import DatasetCompression class DatasetBZip2Compression(DatasetCompression): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_deflate_compression_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_deflate_compression_py3.py index 14a7e8be45c9..715fe91a12a3 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_deflate_compression_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_deflate_compression_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .dataset_compression import DatasetCompression +from .dataset_compression_py3 import DatasetCompression class DatasetDeflateCompression(DatasetCompression): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_gzip_compression_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_gzip_compression_py3.py index 7e8f11cf8e87..99b1081469f8 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_gzip_compression_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_gzip_compression_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .dataset_compression import DatasetCompression +from .dataset_compression_py3 import DatasetCompression class DatasetGZipCompression(DatasetCompression): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_resource_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_resource_py3.py index a7b3d20e2ec6..6eb099dcb884 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_resource_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_resource_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .sub_resource import SubResource +from .sub_resource_py3 import SubResource class DatasetResource(SubResource): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_zip_deflate_compression_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_zip_deflate_compression_py3.py index c6b0bae79f0c..74fbb92ce1ab 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_zip_deflate_compression_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_zip_deflate_compression_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .dataset_compression import DatasetCompression +from .dataset_compression_py3 import DatasetCompression class DatasetZipDeflateCompression(DatasetCompression): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/db2_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/db2_linked_service.py index a3349c5b7ab2..9349bbcba5e0 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/db2_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/db2_linked_service.py @@ -39,9 +39,6 @@ class Db2LinkedService(LinkedService): :param database: Required. Database name for connection. Type: string (or Expression with resultType string). :type database: object - :param schema: Schema name for connection. Type: string (or Expression - with resultType string). - :type schema: object :param authentication_type: AuthenticationType to be used for connection. Possible values include: 'Basic' :type authentication_type: str or @@ -72,7 +69,6 @@ class Db2LinkedService(LinkedService): 'type': {'key': 'type', 'type': 'str'}, 'server': {'key': 'typeProperties.server', 'type': 'object'}, 'database': {'key': 'typeProperties.database', 'type': 'object'}, - 'schema': {'key': 'typeProperties.schema', 'type': 'object'}, 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, 'username': {'key': 'typeProperties.username', 'type': 'object'}, 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, @@ -83,7 +79,6 @@ def __init__(self, **kwargs): super(Db2LinkedService, self).__init__(**kwargs) self.server = kwargs.get('server', None) self.database = kwargs.get('database', None) - self.schema = kwargs.get('schema', None) self.authentication_type = kwargs.get('authentication_type', None) self.username = kwargs.get('username', None) self.password = kwargs.get('password', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/db2_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/db2_linked_service_py3.py index ff86655a9386..d339860c3229 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/db2_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/db2_linked_service_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .linked_service import LinkedService +from .linked_service_py3 import LinkedService class Db2LinkedService(LinkedService): @@ -39,9 +39,6 @@ class Db2LinkedService(LinkedService): :param database: Required. Database name for connection. Type: string (or Expression with resultType string). :type database: object - :param schema: Schema name for connection. Type: string (or Expression - with resultType string). - :type schema: object :param authentication_type: AuthenticationType to be used for connection. Possible values include: 'Basic' :type authentication_type: str or @@ -72,18 +69,16 @@ class Db2LinkedService(LinkedService): 'type': {'key': 'type', 'type': 'str'}, 'server': {'key': 'typeProperties.server', 'type': 'object'}, 'database': {'key': 'typeProperties.database', 'type': 'object'}, - 'schema': {'key': 'typeProperties.schema', 'type': 'object'}, 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, 'username': {'key': 'typeProperties.username', 'type': 'object'}, 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, *, server, database, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, schema=None, authentication_type=None, username=None, password=None, encrypted_credential=None, **kwargs) -> None: + def __init__(self, *, server, database, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, authentication_type=None, username=None, password=None, encrypted_credential=None, **kwargs) -> None: super(Db2LinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) self.server = server self.database = database - self.schema = schema self.authentication_type = authentication_type self.username = username self.password = password diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/document_db_collection_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/document_db_collection_dataset_py3.py index e05756941c1c..e30288172785 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/document_db_collection_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/document_db_collection_dataset_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .dataset import Dataset +from .dataset_py3 import Dataset class DocumentDbCollectionDataset(Dataset): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/document_db_collection_sink_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/document_db_collection_sink_py3.py index be10d5704bb4..5377d4ed5aa5 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/document_db_collection_sink_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/document_db_collection_sink_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .copy_sink import CopySink +from .copy_sink_py3 import CopySink class DocumentDbCollectionSink(CopySink): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/document_db_collection_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/document_db_collection_source_py3.py index 7e72aa6f9800..9c20bfbfa132 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/document_db_collection_source_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/document_db_collection_source_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .copy_source import CopySource +from .copy_source_py3 import CopySource class DocumentDbCollectionSource(CopySource): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_linked_service_py3.py index 5761efa8687a..7e6a7a64bde2 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_linked_service_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .linked_service import LinkedService +from .linked_service_py3 import LinkedService class DrillLinkedService(LinkedService): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_source_py3.py index a68ea79a5238..ea67bbef64fb 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_source_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_source_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .copy_source import CopySource +from .copy_source_py3 import CopySource class DrillSource(CopySource): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_table_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_table_dataset_py3.py index 325085161353..cdff7f8360ca 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_table_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_table_dataset_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .dataset import Dataset +from .dataset_py3 import Dataset class DrillTableDataset(Dataset): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_entity_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_entity_dataset_py3.py index 50c02ae54167..4a5e6f3f542b 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_entity_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_entity_dataset_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .dataset import Dataset +from .dataset_py3 import Dataset class DynamicsEntityDataset(Dataset): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_linked_service_py3.py index f5349731c817..4971dabfba16 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_linked_service_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .linked_service import LinkedService +from .linked_service_py3 import LinkedService class DynamicsLinkedService(LinkedService): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_sink_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_sink_py3.py index 6b301a3f75d7..2e2a64169797 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_sink_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_sink_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .copy_sink import CopySink +from .copy_sink_py3 import CopySink class DynamicsSink(CopySink): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_source_py3.py index a01986da60db..9c921cf40f3a 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_source_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_source_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .copy_source import CopySource +from .copy_source_py3 import CopySource class DynamicsSource(CopySource): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/eloqua_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/eloqua_linked_service_py3.py index ec0af56a9dd1..1c6bd97ecf9d 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/eloqua_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/eloqua_linked_service_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .linked_service import LinkedService +from .linked_service_py3 import LinkedService class EloquaLinkedService(LinkedService): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/eloqua_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/eloqua_object_dataset_py3.py index 4150adf6e7ad..fe8be1261fcd 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/eloqua_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/eloqua_object_dataset_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .dataset import Dataset +from .dataset_py3 import Dataset class EloquaObjectDataset(Dataset): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/eloqua_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/eloqua_source_py3.py index f613b9a0e905..c9d96711743f 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/eloqua_source_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/eloqua_source_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .copy_source import CopySource +from .copy_source_py3 import CopySource class EloquaSource(CopySource): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_pipeline_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_pipeline_activity_py3.py index 72fd5fa3015e..2550875243b6 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_pipeline_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_pipeline_activity_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .control_activity import ControlActivity +from .control_activity_py3 import ControlActivity class ExecutePipelineActivity(ControlActivity): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_ssis_package_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_ssis_package_activity_py3.py index 9c1de33e7433..6b58bf34b57b 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_ssis_package_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_ssis_package_activity_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .execution_activity import ExecutionActivity +from .execution_activity_py3 import ExecutionActivity class ExecuteSSISPackageActivity(ExecutionActivity): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execution_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execution_activity_py3.py index 2d9e3324ffd0..0617ad31da47 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execution_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execution_activity_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .activity import Activity +from .activity_py3 import Activity class ExecutionActivity(Activity): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_py3.py index 57bc1be61aae..5dd8f2e502ea 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .resource import Resource +from .resource_py3 import Resource class Factory(Resource): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_server_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_server_linked_service_py3.py index 705927000617..a9793d5b44fc 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_server_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_server_linked_service_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .linked_service import LinkedService +from .linked_service_py3 import LinkedService class FileServerLinkedService(LinkedService): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_share_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_share_dataset_py3.py index ff76331de81a..08524cbc2a4b 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_share_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_share_dataset_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .dataset import Dataset +from .dataset_py3 import Dataset class FileShareDataset(Dataset): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_system_sink_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_system_sink_py3.py index 59bd6ec12f41..a940e39878f8 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_system_sink_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_system_sink_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .copy_sink import CopySink +from .copy_sink_py3 import CopySink class FileSystemSink(CopySink): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_system_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_system_source_py3.py index bf6eeb2e7fd5..6db0072329d4 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_system_source_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_system_source_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .copy_source import CopySource +from .copy_source_py3 import CopySource class FileSystemSource(CopySource): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/filter_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/filter_activity_py3.py index 87b0d1ccdeb7..acc4617a3f72 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/filter_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/filter_activity_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .control_activity import ControlActivity +from .control_activity_py3 import ControlActivity class FilterActivity(ControlActivity): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/for_each_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/for_each_activity_py3.py index 88d75bc00f3b..e1ebd7ff426b 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/for_each_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/for_each_activity_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .control_activity import ControlActivity +from .control_activity_py3 import ControlActivity class ForEachActivity(ControlActivity): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/ftp_server_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/ftp_server_linked_service_py3.py index b80223faf92f..21fd1168165f 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/ftp_server_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/ftp_server_linked_service_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .linked_service import LinkedService +from .linked_service_py3 import LinkedService class FtpServerLinkedService(LinkedService): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/get_metadata_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/get_metadata_activity_py3.py index 2db128a4dc7f..79e16f1a8c14 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/get_metadata_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/get_metadata_activity_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .execution_activity import ExecutionActivity +from .execution_activity_py3 import ExecutionActivity class GetMetadataActivity(ExecutionActivity): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_big_query_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_big_query_linked_service_py3.py index 549dcb7abf1b..a8582aca98b5 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_big_query_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_big_query_linked_service_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .linked_service import LinkedService +from .linked_service_py3 import LinkedService class GoogleBigQueryLinkedService(LinkedService): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_big_query_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_big_query_object_dataset_py3.py index 1bd794bc719c..d51dfe5de0f9 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_big_query_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_big_query_object_dataset_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .dataset import Dataset +from .dataset_py3 import Dataset class GoogleBigQueryObjectDataset(Dataset): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_big_query_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_big_query_source_py3.py index 7bd1a5256c19..eb5727bd43a5 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_big_query_source_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_big_query_source_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .copy_source import CopySource +from .copy_source_py3 import CopySource class GoogleBigQuerySource(CopySource): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_linked_service_py3.py index f85c146c4fc3..ea15aa2690dd 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_linked_service_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .linked_service import LinkedService +from .linked_service_py3 import LinkedService class GreenplumLinkedService(LinkedService): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_source_py3.py index 9adb06fb16ce..6a373bf9d6ae 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_source_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_source_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .copy_source import CopySource +from .copy_source_py3 import CopySource class GreenplumSource(CopySource): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_table_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_table_dataset_py3.py index e68a07b2d05b..a266e22dae9c 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_table_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_table_dataset_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .dataset import Dataset +from .dataset_py3 import Dataset class GreenplumTableDataset(Dataset): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hbase_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hbase_linked_service_py3.py index e8bc89f401b7..7963b3fc643c 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hbase_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hbase_linked_service_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .linked_service import LinkedService +from .linked_service_py3 import LinkedService class HBaseLinkedService(LinkedService): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hbase_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hbase_object_dataset_py3.py index 30c2ed7de587..a4386af6240f 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hbase_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hbase_object_dataset_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .dataset import Dataset +from .dataset_py3 import Dataset class HBaseObjectDataset(Dataset): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hbase_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hbase_source_py3.py index a6834f23afec..c17d8cf07003 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hbase_source_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hbase_source_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .copy_source import CopySource +from .copy_source_py3 import CopySource class HBaseSource(CopySource): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_hive_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_hive_activity_py3.py index 2bcb4b1c5df4..0d3e900a4398 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_hive_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_hive_activity_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .execution_activity import ExecutionActivity +from .execution_activity_py3 import ExecutionActivity class HDInsightHiveActivity(ExecutionActivity): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_linked_service_py3.py index 0b5374553827..e7e60b5c8333 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_linked_service_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .linked_service import LinkedService +from .linked_service_py3 import LinkedService class HDInsightLinkedService(LinkedService): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_map_reduce_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_map_reduce_activity_py3.py index 1b781fd08bd9..2ac97c2e8453 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_map_reduce_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_map_reduce_activity_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .execution_activity import ExecutionActivity +from .execution_activity_py3 import ExecutionActivity class HDInsightMapReduceActivity(ExecutionActivity): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_on_demand_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_on_demand_linked_service_py3.py index 8bb2cfdb1e91..75f601f6a97b 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_on_demand_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_on_demand_linked_service_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .linked_service import LinkedService +from .linked_service_py3 import LinkedService class HDInsightOnDemandLinkedService(LinkedService): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_pig_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_pig_activity_py3.py index a2cbf90ac518..b57d8da7b682 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_pig_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_pig_activity_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .execution_activity import ExecutionActivity +from .execution_activity_py3 import ExecutionActivity class HDInsightPigActivity(ExecutionActivity): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_spark_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_spark_activity_py3.py index 118772db8539..322d99f7c063 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_spark_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_spark_activity_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .execution_activity import ExecutionActivity +from .execution_activity_py3 import ExecutionActivity class HDInsightSparkActivity(ExecutionActivity): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_streaming_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_streaming_activity_py3.py index e42e543bebac..10756ecd093d 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_streaming_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_streaming_activity_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .execution_activity import ExecutionActivity +from .execution_activity_py3 import ExecutionActivity class HDInsightStreamingActivity(ExecutionActivity): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hdfs_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hdfs_linked_service_py3.py index 550b4824c181..3b854d945e27 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hdfs_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hdfs_linked_service_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .linked_service import LinkedService +from .linked_service_py3 import LinkedService class HdfsLinkedService(LinkedService): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hdfs_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hdfs_source_py3.py index e7125240a2ac..34b194f92d64 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hdfs_source_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hdfs_source_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .copy_source import CopySource +from .copy_source_py3 import CopySource class HdfsSource(CopySource): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hive_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hive_linked_service_py3.py index d6e1a46cbe7e..2f742d72594c 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hive_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hive_linked_service_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .linked_service import LinkedService +from .linked_service_py3 import LinkedService class HiveLinkedService(LinkedService): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hive_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hive_object_dataset_py3.py index 5b245c3543f8..d4bebbbcfe7b 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hive_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hive_object_dataset_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .dataset import Dataset +from .dataset_py3 import Dataset class HiveObjectDataset(Dataset): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hive_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hive_source_py3.py index aa52f335d530..7dc54994b25a 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hive_source_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hive_source_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .copy_source import CopySource +from .copy_source_py3 import CopySource class HiveSource(CopySource): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/http_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/http_dataset_py3.py index a3e7a17b95d3..05c340ab447a 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/http_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/http_dataset_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .dataset import Dataset +from .dataset_py3 import Dataset class HttpDataset(Dataset): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/http_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/http_linked_service_py3.py index 044ab5352683..bd4f03006513 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/http_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/http_linked_service_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .linked_service import LinkedService +from .linked_service_py3 import LinkedService class HttpLinkedService(LinkedService): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/http_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/http_source_py3.py index 53fd82025b11..78bfe7216da6 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/http_source_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/http_source_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .copy_source import CopySource +from .copy_source_py3 import CopySource class HttpSource(CopySource): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_linked_service_py3.py index 216adb7fe228..5c92b45e0252 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_linked_service_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .linked_service import LinkedService +from .linked_service_py3 import LinkedService class HubspotLinkedService(LinkedService): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_object_dataset_py3.py index b8591962805b..f0e8d1a74809 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_object_dataset_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .dataset import Dataset +from .dataset_py3 import Dataset class HubspotObjectDataset(Dataset): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_source_py3.py index 3efc8e70fdd1..4162be4216a6 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_source_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_source_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .copy_source import CopySource +from .copy_source_py3 import CopySource class HubspotSource(CopySource): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/if_condition_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/if_condition_activity_py3.py index 8f63f627ed98..cbef9a900284 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/if_condition_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/if_condition_activity_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .control_activity import ControlActivity +from .control_activity_py3 import ControlActivity class IfConditionActivity(ControlActivity): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/impala_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/impala_linked_service_py3.py index 3605a3443e04..9d79f13b9708 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/impala_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/impala_linked_service_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .linked_service import LinkedService +from .linked_service_py3 import LinkedService class ImpalaLinkedService(LinkedService): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/impala_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/impala_object_dataset_py3.py index 840e4082b726..964bfbd71748 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/impala_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/impala_object_dataset_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .dataset import Dataset +from .dataset_py3 import Dataset class ImpalaObjectDataset(Dataset): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/impala_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/impala_source_py3.py index 63c3e8a3492d..5bdb3391c2fc 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/impala_source_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/impala_source_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .copy_source import CopySource +from .copy_source_py3 import CopySource class ImpalaSource(CopySource): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_resource_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_resource_py3.py index c506360d6357..9239f54166f9 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_resource_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_resource_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .sub_resource import SubResource +from .sub_resource_py3 import SubResource class IntegrationRuntimeResource(SubResource): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_linked_service_py3.py index 824a30ea1b2b..327ce4d0793e 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_linked_service_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .linked_service import LinkedService +from .linked_service_py3 import LinkedService class JiraLinkedService(LinkedService): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_object_dataset_py3.py index d1f5a34f516d..e0c3007bca89 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_object_dataset_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .dataset import Dataset +from .dataset_py3 import Dataset class JiraObjectDataset(Dataset): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_source_py3.py index 80e3e50a062d..5835c7eceef4 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_source_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_source_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .copy_source import CopySource +from .copy_source_py3 import CopySource class JiraSource(CopySource): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/json_format_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/json_format_py3.py index 5213a17d8e4c..a9a7f20ea103 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/json_format_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/json_format_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .dataset_storage_format import DatasetStorageFormat +from .dataset_storage_format_py3 import DatasetStorageFormat class JsonFormat(DatasetStorageFormat): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_key_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_key_py3.py index 9cc45cf95b54..c223b18281cd 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_key_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_key_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .linked_integration_runtime_properties import LinkedIntegrationRuntimeProperties +from .linked_integration_runtime_properties_py3 import LinkedIntegrationRuntimeProperties class LinkedIntegrationRuntimeKey(LinkedIntegrationRuntimeProperties): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_rbac_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_rbac_py3.py index d06dd029b0f1..24d2b946d97e 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_rbac_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_rbac_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .linked_integration_runtime_properties import LinkedIntegrationRuntimeProperties +from .linked_integration_runtime_properties_py3 import LinkedIntegrationRuntimeProperties class LinkedIntegrationRuntimeRbac(LinkedIntegrationRuntimeProperties): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service_resource_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service_resource_py3.py index b9ceff820b34..1fa964b51f57 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service_resource_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service_resource_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .sub_resource import SubResource +from .sub_resource_py3 import SubResource class LinkedServiceResource(SubResource): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/lookup_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/lookup_activity_py3.py index d0ab996d6f3b..fdbcebcd0164 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/lookup_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/lookup_activity_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .execution_activity import ExecutionActivity +from .execution_activity_py3 import ExecutionActivity class LookupActivity(ExecutionActivity): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/magento_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/magento_linked_service_py3.py index ef85812fa1e1..420656103983 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/magento_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/magento_linked_service_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .linked_service import LinkedService +from .linked_service_py3 import LinkedService class MagentoLinkedService(LinkedService): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/magento_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/magento_object_dataset_py3.py index 4386f3704d13..ab3c0bf4f39c 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/magento_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/magento_object_dataset_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .dataset import Dataset +from .dataset_py3 import Dataset class MagentoObjectDataset(Dataset): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/magento_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/magento_source_py3.py index 3ce50a3ed452..a01cf80a969a 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/magento_source_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/magento_source_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .copy_source import CopySource +from .copy_source_py3 import CopySource class MagentoSource(CopySource): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/managed_integration_runtime_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/managed_integration_runtime_py3.py index d2e2468aa741..741564e6e531 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/managed_integration_runtime_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/managed_integration_runtime_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .integration_runtime import IntegrationRuntime +from .integration_runtime_py3 import IntegrationRuntime class ManagedIntegrationRuntime(IntegrationRuntime): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/managed_integration_runtime_status_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/managed_integration_runtime_status_py3.py index 9944628a9997..873c885e7c5d 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/managed_integration_runtime_status_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/managed_integration_runtime_status_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .integration_runtime_status import IntegrationRuntimeStatus +from .integration_runtime_status_py3 import IntegrationRuntimeStatus class ManagedIntegrationRuntimeStatus(IntegrationRuntimeStatus): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_linked_service_py3.py index 8cfa0f73c07a..b0bbd9c4f733 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_linked_service_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .linked_service import LinkedService +from .linked_service_py3 import LinkedService class MariaDBLinkedService(LinkedService): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_source_py3.py index 99fd864e7e62..1dbb6f327d04 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_source_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_source_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .copy_source import CopySource +from .copy_source_py3 import CopySource class MariaDBSource(CopySource): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_table_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_table_dataset_py3.py index d9a30f2deccc..e557822e5fca 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_table_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_table_dataset_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .dataset import Dataset +from .dataset_py3 import Dataset class MariaDBTableDataset(Dataset): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/marketo_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/marketo_linked_service_py3.py index 6ee722e1577c..b4e360931809 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/marketo_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/marketo_linked_service_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .linked_service import LinkedService +from .linked_service_py3 import LinkedService class MarketoLinkedService(LinkedService): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/marketo_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/marketo_object_dataset_py3.py index 3add966719c6..dfb08a716949 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/marketo_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/marketo_object_dataset_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .dataset import Dataset +from .dataset_py3 import Dataset class MarketoObjectDataset(Dataset): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/marketo_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/marketo_source_py3.py index 34a97998f96e..52c16eae0437 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/marketo_source_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/marketo_source_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .copy_source import CopySource +from .copy_source_py3 import CopySource class MarketoSource(CopySource): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/mongo_db_collection_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/mongo_db_collection_dataset_py3.py index db839b82af25..dac4b49da87a 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/mongo_db_collection_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/mongo_db_collection_dataset_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .dataset import Dataset +from .dataset_py3 import Dataset class MongoDbCollectionDataset(Dataset): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/mongo_db_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/mongo_db_linked_service_py3.py index 4f260bf37e90..c1d96a5465b9 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/mongo_db_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/mongo_db_linked_service_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .linked_service import LinkedService +from .linked_service_py3 import LinkedService class MongoDbLinkedService(LinkedService): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/mongo_db_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/mongo_db_source_py3.py index 4313cac8c921..b4f01d8d7ffb 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/mongo_db_source_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/mongo_db_source_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .copy_source import CopySource +from .copy_source_py3 import CopySource class MongoDbSource(CopySource): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/multiple_pipeline_trigger.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/multiple_pipeline_trigger.py index ba053baee118..dd279ab6baa3 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/multiple_pipeline_trigger.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/multiple_pipeline_trigger.py @@ -17,7 +17,7 @@ class MultiplePipelineTrigger(Trigger): pipeline. You probably want to use the sub-classes and not this class directly. Known - sub-classes are: BlobTrigger, ScheduleTrigger + sub-classes are: BlobEventsTrigger, BlobTrigger, ScheduleTrigger Variables are only populated by the server, and will be ignored when sending a request. @@ -55,7 +55,7 @@ class MultiplePipelineTrigger(Trigger): } _subtype_map = { - 'type': {'BlobTrigger': 'BlobTrigger', 'ScheduleTrigger': 'ScheduleTrigger'} + 'type': {'BlobEventsTrigger': 'BlobEventsTrigger', 'BlobTrigger': 'BlobTrigger', 'ScheduleTrigger': 'ScheduleTrigger'} } def __init__(self, **kwargs): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/multiple_pipeline_trigger_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/multiple_pipeline_trigger_py3.py index 8724414f4e0b..3400431e49e2 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/multiple_pipeline_trigger_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/multiple_pipeline_trigger_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .trigger import Trigger +from .trigger_py3 import Trigger class MultiplePipelineTrigger(Trigger): @@ -17,7 +17,7 @@ class MultiplePipelineTrigger(Trigger): pipeline. You probably want to use the sub-classes and not this class directly. Known - sub-classes are: BlobTrigger, ScheduleTrigger + sub-classes are: BlobEventsTrigger, BlobTrigger, ScheduleTrigger Variables are only populated by the server, and will be ignored when sending a request. @@ -55,7 +55,7 @@ class MultiplePipelineTrigger(Trigger): } _subtype_map = { - 'type': {'BlobTrigger': 'BlobTrigger', 'ScheduleTrigger': 'ScheduleTrigger'} + 'type': {'BlobEventsTrigger': 'BlobEventsTrigger', 'BlobTrigger': 'BlobTrigger', 'ScheduleTrigger': 'ScheduleTrigger'} } def __init__(self, *, additional_properties=None, description: str=None, pipelines=None, **kwargs) -> None: diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/my_sql_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/my_sql_linked_service.py index bc5d7e820333..b643e1c7266b 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/my_sql_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/my_sql_linked_service.py @@ -33,20 +33,8 @@ class MySqlLinkedService(LinkedService): :type annotations: list[object] :param type: Required. Constant filled by server. :type type: str - :param server: Required. Server name for connection. Type: string (or - Expression with resultType string). - :type server: object - :param database: Required. Database name for connection. Type: string (or - Expression with resultType string). - :type database: object - :param schema: Schema name for connection. Type: string (or Expression - with resultType string). - :type schema: object - :param username: Username for authentication. Type: string (or Expression - with resultType string). - :type username: object - :param password: Password for authentication. - :type password: ~azure.mgmt.datafactory.models.SecretBase + :param connection_string: Required. The connection string. + :type connection_string: ~azure.mgmt.datafactory.models.SecretBase :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -55,8 +43,7 @@ class MySqlLinkedService(LinkedService): _validation = { 'type': {'required': True}, - 'server': {'required': True}, - 'database': {'required': True}, + 'connection_string': {'required': True}, } _attribute_map = { @@ -66,20 +53,12 @@ class MySqlLinkedService(LinkedService): 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, 'type': {'key': 'type', 'type': 'str'}, - 'server': {'key': 'typeProperties.server', 'type': 'object'}, - 'database': {'key': 'typeProperties.database', 'type': 'object'}, - 'schema': {'key': 'typeProperties.schema', 'type': 'object'}, - 'username': {'key': 'typeProperties.username', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, + 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'SecretBase'}, 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } def __init__(self, **kwargs): super(MySqlLinkedService, self).__init__(**kwargs) - self.server = kwargs.get('server', None) - self.database = kwargs.get('database', None) - self.schema = kwargs.get('schema', None) - self.username = kwargs.get('username', None) - self.password = kwargs.get('password', None) + self.connection_string = kwargs.get('connection_string', None) self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'MySql' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/my_sql_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/my_sql_linked_service_py3.py index 44170220fdfe..38cf9b34198a 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/my_sql_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/my_sql_linked_service_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .linked_service import LinkedService +from .linked_service_py3 import LinkedService class MySqlLinkedService(LinkedService): @@ -33,20 +33,8 @@ class MySqlLinkedService(LinkedService): :type annotations: list[object] :param type: Required. Constant filled by server. :type type: str - :param server: Required. Server name for connection. Type: string (or - Expression with resultType string). - :type server: object - :param database: Required. Database name for connection. Type: string (or - Expression with resultType string). - :type database: object - :param schema: Schema name for connection. Type: string (or Expression - with resultType string). - :type schema: object - :param username: Username for authentication. Type: string (or Expression - with resultType string). - :type username: object - :param password: Password for authentication. - :type password: ~azure.mgmt.datafactory.models.SecretBase + :param connection_string: Required. The connection string. + :type connection_string: ~azure.mgmt.datafactory.models.SecretBase :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -55,8 +43,7 @@ class MySqlLinkedService(LinkedService): _validation = { 'type': {'required': True}, - 'server': {'required': True}, - 'database': {'required': True}, + 'connection_string': {'required': True}, } _attribute_map = { @@ -66,20 +53,12 @@ class MySqlLinkedService(LinkedService): 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, 'type': {'key': 'type', 'type': 'str'}, - 'server': {'key': 'typeProperties.server', 'type': 'object'}, - 'database': {'key': 'typeProperties.database', 'type': 'object'}, - 'schema': {'key': 'typeProperties.schema', 'type': 'object'}, - 'username': {'key': 'typeProperties.username', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, + 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'SecretBase'}, 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, *, server, database, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, schema=None, username=None, password=None, encrypted_credential=None, **kwargs) -> None: + def __init__(self, *, connection_string, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, encrypted_credential=None, **kwargs) -> None: super(MySqlLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.server = server - self.database = database - self.schema = schema - self.username = username - self.password = password + self.connection_string = connection_string self.encrypted_credential = encrypted_credential self.type = 'MySql' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/netezza_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/netezza_linked_service_py3.py index d4c47cdd3c35..b6576954e26b 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/netezza_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/netezza_linked_service_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .linked_service import LinkedService +from .linked_service_py3 import LinkedService class NetezzaLinkedService(LinkedService): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/netezza_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/netezza_source_py3.py index 393dedf4dc55..2b4c38f708ee 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/netezza_source_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/netezza_source_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .copy_source import CopySource +from .copy_source_py3 import CopySource class NetezzaSource(CopySource): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/netezza_table_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/netezza_table_dataset_py3.py index 4acbace2398e..28de2143add2 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/netezza_table_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/netezza_table_dataset_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .dataset import Dataset +from .dataset_py3 import Dataset class NetezzaTableDataset(Dataset): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odata_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odata_linked_service_py3.py index abb4fedf5adf..688bb4e4ffda 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odata_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odata_linked_service_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .linked_service import LinkedService +from .linked_service_py3 import LinkedService class ODataLinkedService(LinkedService): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odata_resource_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odata_resource_dataset_py3.py index 03227b214202..9cd01c414218 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odata_resource_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odata_resource_dataset_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .dataset import Dataset +from .dataset_py3 import Dataset class ODataResourceDataset(Dataset): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odbc_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odbc_linked_service_py3.py index 2e7ab9a8244a..e0147881f3d0 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odbc_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odbc_linked_service_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .linked_service import LinkedService +from .linked_service_py3 import LinkedService class OdbcLinkedService(LinkedService): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odbc_sink_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odbc_sink_py3.py index 8d098f6a96a8..430329bdf2b9 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odbc_sink_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odbc_sink_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .copy_sink import CopySink +from .copy_sink_py3 import CopySink class OdbcSink(CopySink): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_linked_service_py3.py index 350e5d29412c..bc4765e0b2e2 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_linked_service_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .linked_service import LinkedService +from .linked_service_py3 import LinkedService class OracleLinkedService(LinkedService): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_sink_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_sink_py3.py index 07bd11f38775..a6b666d31ed7 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_sink_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_sink_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .copy_sink import CopySink +from .copy_sink_py3 import CopySink class OracleSink(CopySink): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_source_py3.py index 24af230164c5..89252615e6e5 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_source_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_source_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .copy_source import CopySource +from .copy_source_py3 import CopySource class OracleSource(CopySource): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_table_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_table_dataset_py3.py index 2737f7df09a3..0c12187802c9 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_table_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_table_dataset_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .dataset import Dataset +from .dataset_py3 import Dataset class OracleTableDataset(Dataset): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/orc_format_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/orc_format_py3.py index 17db530edc64..40a0e389ccc3 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/orc_format_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/orc_format_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .dataset_storage_format import DatasetStorageFormat +from .dataset_storage_format_py3 import DatasetStorageFormat class OrcFormat(DatasetStorageFormat): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/parquet_format_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/parquet_format_py3.py index ba2b2428f140..36a6f5c88c4d 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/parquet_format_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/parquet_format_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .dataset_storage_format import DatasetStorageFormat +from .dataset_storage_format_py3 import DatasetStorageFormat class ParquetFormat(DatasetStorageFormat): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_linked_service_py3.py index 39c83d422068..41fdb52595ab 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_linked_service_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .linked_service import LinkedService +from .linked_service_py3 import LinkedService class PaypalLinkedService(LinkedService): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_object_dataset_py3.py index 475c579602a7..cb0e92035dfd 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_object_dataset_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .dataset import Dataset +from .dataset_py3 import Dataset class PaypalObjectDataset(Dataset): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_source_py3.py index 9d6d7e6f30fa..347b0e3e550e 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_source_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_source_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .copy_source import CopySource +from .copy_source_py3 import CopySource class PaypalSource(CopySource): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/phoenix_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/phoenix_linked_service_py3.py index fe18934da90a..aeb89e4fdd4a 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/phoenix_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/phoenix_linked_service_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .linked_service import LinkedService +from .linked_service_py3 import LinkedService class PhoenixLinkedService(LinkedService): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/phoenix_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/phoenix_object_dataset_py3.py index a279a7763bd6..73cb4ee2184d 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/phoenix_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/phoenix_object_dataset_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .dataset import Dataset +from .dataset_py3 import Dataset class PhoenixObjectDataset(Dataset): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/phoenix_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/phoenix_source_py3.py index 6497d3dfc80e..619e7220dd09 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/phoenix_source_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/phoenix_source_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .copy_source import CopySource +from .copy_source_py3 import CopySource class PhoenixSource(CopySource): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_resource_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_resource_py3.py index ae1be9a29b07..2b3bf380782a 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_resource_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_resource_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .sub_resource import SubResource +from .sub_resource_py3 import SubResource class PipelineResource(SubResource): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/postgre_sql_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/postgre_sql_linked_service.py index 7a96d59e87e1..6a572c24c1cd 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/postgre_sql_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/postgre_sql_linked_service.py @@ -33,20 +33,8 @@ class PostgreSqlLinkedService(LinkedService): :type annotations: list[object] :param type: Required. Constant filled by server. :type type: str - :param server: Required. Server name for connection. Type: string (or - Expression with resultType string). - :type server: object - :param database: Required. Database name for connection. Type: string (or - Expression with resultType string). - :type database: object - :param schema: Schema name for connection. Type: string (or Expression - with resultType string). - :type schema: object - :param username: Username for authentication. Type: string (or Expression - with resultType string). - :type username: object - :param password: Password for authentication. - :type password: ~azure.mgmt.datafactory.models.SecretBase + :param connection_string: Required. The connection string. + :type connection_string: ~azure.mgmt.datafactory.models.SecretBase :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -55,8 +43,7 @@ class PostgreSqlLinkedService(LinkedService): _validation = { 'type': {'required': True}, - 'server': {'required': True}, - 'database': {'required': True}, + 'connection_string': {'required': True}, } _attribute_map = { @@ -66,20 +53,12 @@ class PostgreSqlLinkedService(LinkedService): 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, 'type': {'key': 'type', 'type': 'str'}, - 'server': {'key': 'typeProperties.server', 'type': 'object'}, - 'database': {'key': 'typeProperties.database', 'type': 'object'}, - 'schema': {'key': 'typeProperties.schema', 'type': 'object'}, - 'username': {'key': 'typeProperties.username', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, + 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'SecretBase'}, 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } def __init__(self, **kwargs): super(PostgreSqlLinkedService, self).__init__(**kwargs) - self.server = kwargs.get('server', None) - self.database = kwargs.get('database', None) - self.schema = kwargs.get('schema', None) - self.username = kwargs.get('username', None) - self.password = kwargs.get('password', None) + self.connection_string = kwargs.get('connection_string', None) self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'PostgreSql' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/postgre_sql_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/postgre_sql_linked_service_py3.py index 976e29aebbde..ae16efbbfdad 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/postgre_sql_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/postgre_sql_linked_service_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .linked_service import LinkedService +from .linked_service_py3 import LinkedService class PostgreSqlLinkedService(LinkedService): @@ -33,20 +33,8 @@ class PostgreSqlLinkedService(LinkedService): :type annotations: list[object] :param type: Required. Constant filled by server. :type type: str - :param server: Required. Server name for connection. Type: string (or - Expression with resultType string). - :type server: object - :param database: Required. Database name for connection. Type: string (or - Expression with resultType string). - :type database: object - :param schema: Schema name for connection. Type: string (or Expression - with resultType string). - :type schema: object - :param username: Username for authentication. Type: string (or Expression - with resultType string). - :type username: object - :param password: Password for authentication. - :type password: ~azure.mgmt.datafactory.models.SecretBase + :param connection_string: Required. The connection string. + :type connection_string: ~azure.mgmt.datafactory.models.SecretBase :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -55,8 +43,7 @@ class PostgreSqlLinkedService(LinkedService): _validation = { 'type': {'required': True}, - 'server': {'required': True}, - 'database': {'required': True}, + 'connection_string': {'required': True}, } _attribute_map = { @@ -66,20 +53,12 @@ class PostgreSqlLinkedService(LinkedService): 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, 'type': {'key': 'type', 'type': 'str'}, - 'server': {'key': 'typeProperties.server', 'type': 'object'}, - 'database': {'key': 'typeProperties.database', 'type': 'object'}, - 'schema': {'key': 'typeProperties.schema', 'type': 'object'}, - 'username': {'key': 'typeProperties.username', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, + 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'SecretBase'}, 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, *, server, database, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, schema=None, username=None, password=None, encrypted_credential=None, **kwargs) -> None: + def __init__(self, *, connection_string, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, encrypted_credential=None, **kwargs) -> None: super(PostgreSqlLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.server = server - self.database = database - self.schema = schema - self.username = username - self.password = password + self.connection_string = connection_string self.encrypted_credential = encrypted_credential self.type = 'PostgreSql' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/presto_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/presto_linked_service_py3.py index e9721fcba562..fe178f62df4f 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/presto_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/presto_linked_service_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .linked_service import LinkedService +from .linked_service_py3 import LinkedService class PrestoLinkedService(LinkedService): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/presto_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/presto_object_dataset_py3.py index a0dce3cd789b..dfb69ded9c4c 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/presto_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/presto_object_dataset_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .dataset import Dataset +from .dataset_py3 import Dataset class PrestoObjectDataset(Dataset): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/presto_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/presto_source_py3.py index be957bd2bfcd..ad16115ef8f3 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/presto_source_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/presto_source_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .copy_source import CopySource +from .copy_source_py3 import CopySource class PrestoSource(CopySource): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_linked_service_py3.py index b56e34367553..8c23934112bd 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_linked_service_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .linked_service import LinkedService +from .linked_service_py3 import LinkedService class QuickBooksLinkedService(LinkedService): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_object_dataset_py3.py index e852f5bcd624..7897881b5cbb 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_object_dataset_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .dataset import Dataset +from .dataset_py3 import Dataset class QuickBooksObjectDataset(Dataset): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_source_py3.py index e2ecafefc3bf..b6bb7a260d1d 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_source_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_source_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .copy_source import CopySource +from .copy_source_py3 import CopySource class QuickBooksSource(CopySource): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/relational_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/relational_source_py3.py index efac5c7b9050..9e7a75043b8c 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/relational_source_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/relational_source_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .copy_source import CopySource +from .copy_source_py3 import CopySource class RelationalSource(CopySource): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/relational_table_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/relational_table_dataset_py3.py index 908ca1d8657d..0b492686356d 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/relational_table_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/relational_table_dataset_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .dataset import Dataset +from .dataset_py3 import Dataset class RelationalTableDataset(Dataset): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_linked_service_py3.py index efad0984a99f..4c1997e6ab26 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_linked_service_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .linked_service import LinkedService +from .linked_service_py3 import LinkedService class ResponsysLinkedService(LinkedService): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_object_dataset_py3.py index e7b88bd52cec..ffdd38dea188 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_object_dataset_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .dataset import Dataset +from .dataset_py3 import Dataset class ResponsysObjectDataset(Dataset): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_source_py3.py index 1934f830d515..3bfb9c19a2a7 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_source_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_source_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .copy_source import CopySource +from .copy_source_py3 import CopySource class ResponsysSource(CopySource): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_linked_service_py3.py index 2668ed113708..9fa5287aa3b4 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_linked_service_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .linked_service import LinkedService +from .linked_service_py3 import LinkedService class SalesforceLinkedService(LinkedService): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_marketing_cloud_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_marketing_cloud_linked_service_py3.py index 2cf5a947413e..863b679398e1 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_marketing_cloud_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_marketing_cloud_linked_service_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .linked_service import LinkedService +from .linked_service_py3 import LinkedService class SalesforceMarketingCloudLinkedService(LinkedService): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_marketing_cloud_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_marketing_cloud_object_dataset_py3.py index 08b243e3c826..3180cbd37ea1 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_marketing_cloud_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_marketing_cloud_object_dataset_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .dataset import Dataset +from .dataset_py3 import Dataset class SalesforceMarketingCloudObjectDataset(Dataset): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_marketing_cloud_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_marketing_cloud_source_py3.py index 4a7e17da4f06..0a3d26cfb43b 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_marketing_cloud_source_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_marketing_cloud_source_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .copy_source import CopySource +from .copy_source_py3 import CopySource class SalesforceMarketingCloudSource(CopySource): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_object_dataset_py3.py index ccebb62fd5b3..d31a9754f738 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_object_dataset_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .dataset import Dataset +from .dataset_py3 import Dataset class SalesforceObjectDataset(Dataset): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_sink_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_sink_py3.py index 10f9aea9c69d..6db44ebb4228 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_sink_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_sink_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .copy_sink import CopySink +from .copy_sink_py3 import CopySink class SalesforceSink(CopySink): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_source_py3.py index 729c342b37f7..9ebc65ddeec8 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_source_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_source_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .copy_source import CopySource +from .copy_source_py3 import CopySource class SalesforceSource(CopySource): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_bw_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_bw_linked_service_py3.py index a3e686b16405..a1f6133e558d 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_bw_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_bw_linked_service_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .linked_service import LinkedService +from .linked_service_py3 import LinkedService class SapBWLinkedService(LinkedService): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_linked_service_py3.py index 319a5f31e99a..85c1100d01eb 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_linked_service_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .linked_service import LinkedService +from .linked_service_py3 import LinkedService class SapCloudForCustomerLinkedService(LinkedService): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_resource_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_resource_dataset_py3.py index 4a7c19252dd1..fc52d1b3bdaf 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_resource_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_resource_dataset_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .dataset import Dataset +from .dataset_py3 import Dataset class SapCloudForCustomerResourceDataset(Dataset): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_sink_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_sink_py3.py index c366c6c00a4c..f3cd45263f3e 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_sink_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_sink_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .copy_sink import CopySink +from .copy_sink_py3 import CopySink class SapCloudForCustomerSink(CopySink): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_source_py3.py index 8edc564d0b8d..ab5bddf21be3 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_source_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_source_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .copy_source import CopySource +from .copy_source_py3 import CopySource class SapCloudForCustomerSource(CopySource): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_ecc_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_ecc_linked_service_py3.py index ef2fe2e60573..24490fb39a9a 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_ecc_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_ecc_linked_service_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .linked_service import LinkedService +from .linked_service_py3 import LinkedService class SapEccLinkedService(LinkedService): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_ecc_resource_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_ecc_resource_dataset_py3.py index 321d4d85bc9a..6836e6206041 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_ecc_resource_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_ecc_resource_dataset_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .dataset import Dataset +from .dataset_py3 import Dataset class SapEccResourceDataset(Dataset): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_ecc_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_ecc_source_py3.py index e7ec1d6b495b..f8993720428c 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_ecc_source_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_ecc_source_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .copy_source import CopySource +from .copy_source_py3 import CopySource class SapEccSource(CopySource): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_hana_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_hana_linked_service_py3.py index 60c8d0283d61..c906d74d0c2b 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_hana_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_hana_linked_service_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .linked_service import LinkedService +from .linked_service_py3 import LinkedService class SapHanaLinkedService(LinkedService): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/schedule_trigger_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/schedule_trigger_py3.py index 217195a5930f..1fc148a81b29 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/schedule_trigger_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/schedule_trigger_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .multiple_pipeline_trigger import MultiplePipelineTrigger +from .multiple_pipeline_trigger_py3 import MultiplePipelineTrigger class ScheduleTrigger(MultiplePipelineTrigger): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/secure_string_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/secure_string_py3.py index 87513c02573c..d7ebd5e13e78 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/secure_string_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/secure_string_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .secret_base import SecretBase +from .secret_base_py3 import SecretBase class SecureString(SecretBase): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime_py3.py index 3715401fa8dd..2a6929ecdd15 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .integration_runtime import IntegrationRuntime +from .integration_runtime_py3 import IntegrationRuntime class SelfHostedIntegrationRuntime(IntegrationRuntime): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime_status_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime_status_py3.py index e88fa5fa019d..07362ed1835a 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime_status_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime_status_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .integration_runtime_status import IntegrationRuntimeStatus +from .integration_runtime_status_py3 import IntegrationRuntimeStatus class SelfHostedIntegrationRuntimeStatus(IntegrationRuntimeStatus): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_linked_service_py3.py index 5aed5a07e4fc..854468f281c0 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_linked_service_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .linked_service import LinkedService +from .linked_service_py3 import LinkedService class ServiceNowLinkedService(LinkedService): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_object_dataset_py3.py index 8d2abba32ec6..97a484fb8826 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_object_dataset_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .dataset import Dataset +from .dataset_py3 import Dataset class ServiceNowObjectDataset(Dataset): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_source_py3.py index 8010f2aea2fb..ffe72cb426e7 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_source_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_source_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .copy_source import CopySource +from .copy_source_py3 import CopySource class ServiceNowSource(CopySource): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sftp_server_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sftp_server_linked_service_py3.py index 2c2fcca6f001..581e8f2a0f8e 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sftp_server_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sftp_server_linked_service_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .linked_service import LinkedService +from .linked_service_py3 import LinkedService class SftpServerLinkedService(LinkedService): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_linked_service_py3.py index 4d1ac17b1ebe..8d7d659bebe1 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_linked_service_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .linked_service import LinkedService +from .linked_service_py3 import LinkedService class ShopifyLinkedService(LinkedService): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_object_dataset_py3.py index b3ccec372e44..f83ac708a93d 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_object_dataset_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .dataset import Dataset +from .dataset_py3 import Dataset class ShopifyObjectDataset(Dataset): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_source_py3.py index 1170ae521d47..953d9246d9a2 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_source_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_source_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .copy_source import CopySource +from .copy_source_py3 import CopySource class ShopifySource(CopySource): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/spark_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/spark_linked_service_py3.py index 832b8cb0c7da..c5e20deef8e8 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/spark_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/spark_linked_service_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .linked_service import LinkedService +from .linked_service_py3 import LinkedService class SparkLinkedService(LinkedService): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/spark_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/spark_object_dataset_py3.py index cee1552ac9ee..c12d454f68a7 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/spark_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/spark_object_dataset_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .dataset import Dataset +from .dataset_py3 import Dataset class SparkObjectDataset(Dataset): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/spark_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/spark_source_py3.py index 8fd55d21d579..ede7f9ed5e2b 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/spark_source_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/spark_source_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .copy_source import CopySource +from .copy_source_py3 import CopySource class SparkSource(CopySource): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_dw_sink_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_dw_sink_py3.py index 5380b32f61d7..2b2d44cf16c6 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_dw_sink_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_dw_sink_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .copy_sink import CopySink +from .copy_sink_py3 import CopySink class SqlDWSink(CopySink): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_dw_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_dw_source_py3.py index 3fb991a7bc64..b74c004141d1 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_dw_source_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_dw_source_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .copy_source import CopySource +from .copy_source_py3 import CopySource class SqlDWSource(CopySource): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_linked_service_py3.py index 0ec79b1eefe7..fb446a12f601 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_linked_service_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .linked_service import LinkedService +from .linked_service_py3 import LinkedService class SqlServerLinkedService(LinkedService): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_stored_procedure_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_stored_procedure_activity_py3.py index a03d1500425a..b1166e0de545 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_stored_procedure_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_stored_procedure_activity_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .execution_activity import ExecutionActivity +from .execution_activity_py3 import ExecutionActivity class SqlServerStoredProcedureActivity(ExecutionActivity): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_table_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_table_dataset_py3.py index fe75db9ab096..4a2f80144cda 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_table_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_table_dataset_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .dataset import Dataset +from .dataset_py3 import Dataset class SqlServerTableDataset(Dataset): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_sink_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_sink_py3.py index e12a17c18ece..5aa68f696f16 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_sink_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_sink_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .copy_sink import CopySink +from .copy_sink_py3 import CopySink class SqlSink(CopySink): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_source_py3.py index 34be7054e4e8..ff39b6768a9f 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_source_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_source_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .copy_source import CopySource +from .copy_source_py3 import CopySource class SqlSource(CopySource): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_linked_service_py3.py index 08cb36e5a54f..ffbe49d74ee8 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_linked_service_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .linked_service import LinkedService +from .linked_service_py3 import LinkedService class SquareLinkedService(LinkedService): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_object_dataset_py3.py index 0fd0dd4abbfd..879b34c83620 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_object_dataset_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .dataset import Dataset +from .dataset_py3 import Dataset class SquareObjectDataset(Dataset): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_source_py3.py index 81fbbeb9c150..a87b924949de 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_source_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_source_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .copy_source import CopySource +from .copy_source_py3 import CopySource class SquareSource(CopySource): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sybase_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sybase_linked_service_py3.py index b5c9c8788f25..59b20a5f73cd 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sybase_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sybase_linked_service_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .linked_service import LinkedService +from .linked_service_py3 import LinkedService class SybaseLinkedService(LinkedService): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/tabular_translator_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/tabular_translator_py3.py index eddbf396c0d6..45eae0147b20 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/tabular_translator_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/tabular_translator_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .copy_translator import CopyTranslator +from .copy_translator_py3 import CopyTranslator class TabularTranslator(CopyTranslator): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/teradata_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/teradata_linked_service.py index 0df87c4f1487..b3847d7dd9f4 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/teradata_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/teradata_linked_service.py @@ -36,9 +36,6 @@ class TeradataLinkedService(LinkedService): :param server: Required. Server name for connection. Type: string (or Expression with resultType string). :type server: object - :param schema: Schema name for connection. Type: string (or Expression - with resultType string). - :type schema: object :param authentication_type: AuthenticationType to be used for connection. Possible values include: 'Basic', 'Windows' :type authentication_type: str or @@ -67,7 +64,6 @@ class TeradataLinkedService(LinkedService): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'type': {'key': 'type', 'type': 'str'}, 'server': {'key': 'typeProperties.server', 'type': 'object'}, - 'schema': {'key': 'typeProperties.schema', 'type': 'object'}, 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, 'username': {'key': 'typeProperties.username', 'type': 'object'}, 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, @@ -77,7 +73,6 @@ class TeradataLinkedService(LinkedService): def __init__(self, **kwargs): super(TeradataLinkedService, self).__init__(**kwargs) self.server = kwargs.get('server', None) - self.schema = kwargs.get('schema', None) self.authentication_type = kwargs.get('authentication_type', None) self.username = kwargs.get('username', None) self.password = kwargs.get('password', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/teradata_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/teradata_linked_service_py3.py index 1fecbababcb2..236741422023 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/teradata_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/teradata_linked_service_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .linked_service import LinkedService +from .linked_service_py3 import LinkedService class TeradataLinkedService(LinkedService): @@ -36,9 +36,6 @@ class TeradataLinkedService(LinkedService): :param server: Required. Server name for connection. Type: string (or Expression with resultType string). :type server: object - :param schema: Schema name for connection. Type: string (or Expression - with resultType string). - :type schema: object :param authentication_type: AuthenticationType to be used for connection. Possible values include: 'Basic', 'Windows' :type authentication_type: str or @@ -67,17 +64,15 @@ class TeradataLinkedService(LinkedService): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'type': {'key': 'type', 'type': 'str'}, 'server': {'key': 'typeProperties.server', 'type': 'object'}, - 'schema': {'key': 'typeProperties.schema', 'type': 'object'}, 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, 'username': {'key': 'typeProperties.username', 'type': 'object'}, 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, *, server, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, schema=None, authentication_type=None, username=None, password=None, encrypted_credential=None, **kwargs) -> None: + def __init__(self, *, server, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, authentication_type=None, username=None, password=None, encrypted_credential=None, **kwargs) -> None: super(TeradataLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) self.server = server - self.schema = schema self.authentication_type = authentication_type self.username = username self.password = password diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/text_format_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/text_format_py3.py index 17a7b55327c4..0d876f62b112 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/text_format_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/text_format_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .dataset_storage_format import DatasetStorageFormat +from .dataset_storage_format_py3 import DatasetStorageFormat class TextFormat(DatasetStorageFormat): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger_resource_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger_resource_py3.py index 232374e962c5..ae6a04ac3128 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger_resource_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger_resource_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .sub_resource import SubResource +from .sub_resource_py3 import SubResource class TriggerResource(SubResource): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/tumbling_window_trigger_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/tumbling_window_trigger_py3.py index a436cc6fa2b4..9db68bc272d0 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/tumbling_window_trigger_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/tumbling_window_trigger_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .trigger import Trigger +from .trigger_py3 import Trigger class TumblingWindowTrigger(Trigger): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/until_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/until_activity_py3.py index 5d3c2739feef..a73467b7123b 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/until_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/until_activity_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .control_activity import ControlActivity +from .control_activity_py3 import ControlActivity class UntilActivity(ControlActivity): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/vertica_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/vertica_linked_service_py3.py index b3380cf3ccda..1f8d8d570205 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/vertica_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/vertica_linked_service_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .linked_service import LinkedService +from .linked_service_py3 import LinkedService class VerticaLinkedService(LinkedService): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/vertica_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/vertica_source_py3.py index 25eb02039b81..6be2edd35218 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/vertica_source_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/vertica_source_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .copy_source import CopySource +from .copy_source_py3 import CopySource class VerticaSource(CopySource): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/vertica_table_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/vertica_table_dataset_py3.py index 6681fca04f33..1dd56dfd46ee 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/vertica_table_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/vertica_table_dataset_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .dataset import Dataset +from .dataset_py3 import Dataset class VerticaTableDataset(Dataset): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/wait_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/wait_activity_py3.py index 7c7dc1d1f30f..ee3da26e5b83 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/wait_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/wait_activity_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .control_activity import ControlActivity +from .control_activity_py3 import ControlActivity class WaitActivity(ControlActivity): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_activity_py3.py index 5b3220509c34..c84a8e4d31d2 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_activity_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .execution_activity import ExecutionActivity +from .execution_activity_py3 import ExecutionActivity class WebActivity(ExecutionActivity): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_anonymous_authentication_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_anonymous_authentication_py3.py index c889c410c585..ee7a4e780a1f 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_anonymous_authentication_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_anonymous_authentication_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .web_linked_service_type_properties import WebLinkedServiceTypeProperties +from .web_linked_service_type_properties_py3 import WebLinkedServiceTypeProperties class WebAnonymousAuthentication(WebLinkedServiceTypeProperties): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_basic_authentication_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_basic_authentication_py3.py index a6ba3cc384cf..71577ec86565 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_basic_authentication_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_basic_authentication_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .web_linked_service_type_properties import WebLinkedServiceTypeProperties +from .web_linked_service_type_properties_py3 import WebLinkedServiceTypeProperties class WebBasicAuthentication(WebLinkedServiceTypeProperties): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_client_certificate_authentication_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_client_certificate_authentication_py3.py index 16d41490e2c2..7ac859b677a8 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_client_certificate_authentication_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_client_certificate_authentication_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .web_linked_service_type_properties import WebLinkedServiceTypeProperties +from .web_linked_service_type_properties_py3 import WebLinkedServiceTypeProperties class WebClientCertificateAuthentication(WebLinkedServiceTypeProperties): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_linked_service_py3.py index 0e49f955d436..3afa3a1bcb05 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_linked_service_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .linked_service import LinkedService +from .linked_service_py3 import LinkedService class WebLinkedService(LinkedService): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_source_py3.py index 97571eb096f8..7c5ce29d3d26 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_source_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_source_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .copy_source import CopySource +from .copy_source_py3 import CopySource class WebSource(CopySource): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_table_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_table_dataset_py3.py index a5780a7e480a..7a361289553c 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_table_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_table_dataset_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .dataset import Dataset +from .dataset_py3 import Dataset class WebTableDataset(Dataset): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_linked_service_py3.py index 104af987bddc..7a8a08d0bb22 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_linked_service_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .linked_service import LinkedService +from .linked_service_py3 import LinkedService class XeroLinkedService(LinkedService): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_object_dataset_py3.py index 80a74d953589..685be7b0dffa 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_object_dataset_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .dataset import Dataset +from .dataset_py3 import Dataset class XeroObjectDataset(Dataset): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_source_py3.py index 4a1c2d2bdbc3..e8c8e954eda6 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_source_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_source_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .copy_source import CopySource +from .copy_source_py3 import CopySource class XeroSource(CopySource): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/zoho_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/zoho_linked_service_py3.py index 4e5d0f5f31f3..c05d018146d6 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/zoho_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/zoho_linked_service_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .linked_service import LinkedService +from .linked_service_py3 import LinkedService class ZohoLinkedService(LinkedService): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/zoho_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/zoho_object_dataset_py3.py index 112d3bbd3952..40a579665206 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/zoho_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/zoho_object_dataset_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .dataset import Dataset +from .dataset_py3 import Dataset class ZohoObjectDataset(Dataset): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/zoho_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/zoho_source_py3.py index 8533b239ae13..5f0547d9465a 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/zoho_source_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/zoho_source_py3.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from .copy_source import CopySource +from .copy_source_py3 import CopySource class ZohoSource(CopySource): From 140c0d21241ed51072fc2938b900411bb16e3772 Mon Sep 17 00:00:00 2001 From: Azure SDK for Python bot Date: Tue, 22 May 2018 10:28:46 -0700 Subject: [PATCH 05/54] [AutoPR datafactory/resource-manager] add configure repo api swagger (#2392) * Generated from d7f5567289add61a8368ce9bf1a710264183366c add configure repo api swagger * Generated from 42b127de426d2e7e62893799af31cfac6a58b07d fixing autorest issue * Generated from 27d71a427c73a4adfd90f2347ba5c0a9e0403e7d change route * Generated from 2c233a59f11801e38916d47e13019b26ea4566ed fix per comments * Generated from fa5757e279b6732dedc799857b4b8cbaac22a663 change factoryId for factoryResourceId * Generated from 47d1905337fe73deee6e623add5eaafb62caf5d0 rename example * Generated from 47d1905337fe73deee6e623add5eaafb62caf5d0 rename example --- .../azure/mgmt/datafactory/models/__init__.py | 6 ++ .../azure/mgmt/datafactory/models/factory.py | 5 ++ .../mgmt/datafactory/models/factory_py3.py | 7 +- .../datafactory/models/factory_repo_update.py | 37 +++++++++++ .../models/factory_repo_update_py3.py | 37 +++++++++++ .../models/factory_vsts_configuration.py | 52 +++++++++++++++ .../models/factory_vsts_configuration_py3.py | 52 +++++++++++++++ .../operations/factories_operations.py | 65 +++++++++++++++++++ 8 files changed, 260 insertions(+), 1 deletion(-) create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_repo_update.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_repo_update_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_vsts_configuration.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_vsts_configuration_py3.py diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py index fdd54957b36f..ed790633bf18 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py @@ -18,6 +18,7 @@ from .azure_key_vault_secret_reference_py3 import AzureKeyVaultSecretReference from .secret_base_py3 import SecretBase from .factory_identity_py3 import FactoryIdentity + from .factory_vsts_configuration_py3 import FactoryVSTSConfiguration from .factory_py3 import Factory from .integration_runtime_py3 import IntegrationRuntime from .integration_runtime_resource_py3 import IntegrationRuntimeResource @@ -39,6 +40,7 @@ from .trigger_resource_py3 import TriggerResource from .create_run_response_py3 import CreateRunResponse from .error_response_py3 import ErrorResponse, ErrorResponseException + from .factory_repo_update_py3 import FactoryRepoUpdate from .pipeline_reference_py3 import PipelineReference from .trigger_pipeline_reference_py3 import TriggerPipelineReference from .factory_update_parameters_py3 import FactoryUpdateParameters @@ -343,6 +345,7 @@ from .azure_key_vault_secret_reference import AzureKeyVaultSecretReference from .secret_base import SecretBase from .factory_identity import FactoryIdentity + from .factory_vsts_configuration import FactoryVSTSConfiguration from .factory import Factory from .integration_runtime import IntegrationRuntime from .integration_runtime_resource import IntegrationRuntimeResource @@ -364,6 +367,7 @@ from .trigger_resource import TriggerResource from .create_run_response import CreateRunResponse from .error_response import ErrorResponse, ErrorResponseException + from .factory_repo_update import FactoryRepoUpdate from .pipeline_reference import PipelineReference from .trigger_pipeline_reference import TriggerPipelineReference from .factory_update_parameters import FactoryUpdateParameters @@ -737,6 +741,7 @@ 'AzureKeyVaultSecretReference', 'SecretBase', 'FactoryIdentity', + 'FactoryVSTSConfiguration', 'Factory', 'IntegrationRuntime', 'IntegrationRuntimeResource', @@ -758,6 +763,7 @@ 'TriggerResource', 'CreateRunResponse', 'ErrorResponse', 'ErrorResponseException', + 'FactoryRepoUpdate', 'PipelineReference', 'TriggerPipelineReference', 'FactoryUpdateParameters', diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory.py index 3c1fdbada034..eaeaa060605e 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory.py @@ -39,6 +39,9 @@ class Factory(Resource): :vartype create_time: datetime :ivar version: Version of the factory. :vartype version: str + :param vsts_configuration: VSTS repo information of the factory. + :type vsts_configuration: + ~azure.mgmt.datafactory.models.FactoryVSTSConfiguration """ _validation = { @@ -61,6 +64,7 @@ class Factory(Resource): 'provisioning_state': {'key': 'properties.provisioningState', 'type': 'str'}, 'create_time': {'key': 'properties.createTime', 'type': 'iso-8601'}, 'version': {'key': 'properties.version', 'type': 'str'}, + 'vsts_configuration': {'key': 'properties.vstsConfiguration', 'type': 'FactoryVSTSConfiguration'}, } def __init__(self, **kwargs): @@ -70,3 +74,4 @@ def __init__(self, **kwargs): self.provisioning_state = None self.create_time = None self.version = None + self.vsts_configuration = kwargs.get('vsts_configuration', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_py3.py index 5dd8f2e502ea..b718282c585a 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_py3.py @@ -39,6 +39,9 @@ class Factory(Resource): :vartype create_time: datetime :ivar version: Version of the factory. :vartype version: str + :param vsts_configuration: VSTS repo information of the factory. + :type vsts_configuration: + ~azure.mgmt.datafactory.models.FactoryVSTSConfiguration """ _validation = { @@ -61,12 +64,14 @@ class Factory(Resource): 'provisioning_state': {'key': 'properties.provisioningState', 'type': 'str'}, 'create_time': {'key': 'properties.createTime', 'type': 'iso-8601'}, 'version': {'key': 'properties.version', 'type': 'str'}, + 'vsts_configuration': {'key': 'properties.vstsConfiguration', 'type': 'FactoryVSTSConfiguration'}, } - def __init__(self, *, location: str=None, tags=None, additional_properties=None, identity=None, **kwargs) -> None: + def __init__(self, *, location: str=None, tags=None, additional_properties=None, identity=None, vsts_configuration=None, **kwargs) -> None: super(Factory, self).__init__(location=location, tags=tags, **kwargs) self.additional_properties = additional_properties self.identity = identity self.provisioning_state = None self.create_time = None self.version = None + self.vsts_configuration = vsts_configuration diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_repo_update.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_repo_update.py new file mode 100644 index 000000000000..7651442cd45c --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_repo_update.py @@ -0,0 +1,37 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class FactoryRepoUpdate(Model): + """Factory's VSTS repo information. + + :param factory_resource_id: The factory resource id. + :type factory_resource_id: str + :param resource_group_name: The resource group name. + :type resource_group_name: str + :param vsts_configuration: VSTS repo information of the factory. + :type vsts_configuration: + ~azure.mgmt.datafactory.models.FactoryVSTSConfiguration + """ + + _attribute_map = { + 'factory_resource_id': {'key': 'factoryResourceId', 'type': 'str'}, + 'resource_group_name': {'key': 'resourceGroupName', 'type': 'str'}, + 'vsts_configuration': {'key': 'vstsConfiguration', 'type': 'FactoryVSTSConfiguration'}, + } + + def __init__(self, **kwargs): + super(FactoryRepoUpdate, self).__init__(**kwargs) + self.factory_resource_id = kwargs.get('factory_resource_id', None) + self.resource_group_name = kwargs.get('resource_group_name', None) + self.vsts_configuration = kwargs.get('vsts_configuration', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_repo_update_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_repo_update_py3.py new file mode 100644 index 000000000000..160f3f80ea28 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_repo_update_py3.py @@ -0,0 +1,37 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class FactoryRepoUpdate(Model): + """Factory's VSTS repo information. + + :param factory_resource_id: The factory resource id. + :type factory_resource_id: str + :param resource_group_name: The resource group name. + :type resource_group_name: str + :param vsts_configuration: VSTS repo information of the factory. + :type vsts_configuration: + ~azure.mgmt.datafactory.models.FactoryVSTSConfiguration + """ + + _attribute_map = { + 'factory_resource_id': {'key': 'factoryResourceId', 'type': 'str'}, + 'resource_group_name': {'key': 'resourceGroupName', 'type': 'str'}, + 'vsts_configuration': {'key': 'vstsConfiguration', 'type': 'FactoryVSTSConfiguration'}, + } + + def __init__(self, *, factory_resource_id: str=None, resource_group_name: str=None, vsts_configuration=None, **kwargs) -> None: + super(FactoryRepoUpdate, self).__init__(**kwargs) + self.factory_resource_id = factory_resource_id + self.resource_group_name = resource_group_name + self.vsts_configuration = vsts_configuration diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_vsts_configuration.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_vsts_configuration.py new file mode 100644 index 000000000000..c2c017197398 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_vsts_configuration.py @@ -0,0 +1,52 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class FactoryVSTSConfiguration(Model): + """Factory's VSTS repo information. + + :param account_name: VSTS account name. + :type account_name: str + :param project_name: VSTS project name. + :type project_name: str + :param repository_name: VSTS repository name. + :type repository_name: str + :param collaboration_branch: VSTS collaboration branch. + :type collaboration_branch: str + :param root_folder: VSTS root folder. + :type root_folder: str + :param last_commit_id: VSTS last commit id. + :type last_commit_id: str + :param tenant_id: VSTS tenant id. + :type tenant_id: str + """ + + _attribute_map = { + 'account_name': {'key': 'accountName', 'type': 'str'}, + 'project_name': {'key': 'projectName', 'type': 'str'}, + 'repository_name': {'key': 'repositoryName', 'type': 'str'}, + 'collaboration_branch': {'key': 'collaborationBranch', 'type': 'str'}, + 'root_folder': {'key': 'rootFolder', 'type': 'str'}, + 'last_commit_id': {'key': 'lastCommitId', 'type': 'str'}, + 'tenant_id': {'key': 'tenantId', 'type': 'str'}, + } + + def __init__(self, **kwargs): + super(FactoryVSTSConfiguration, self).__init__(**kwargs) + self.account_name = kwargs.get('account_name', None) + self.project_name = kwargs.get('project_name', None) + self.repository_name = kwargs.get('repository_name', None) + self.collaboration_branch = kwargs.get('collaboration_branch', None) + self.root_folder = kwargs.get('root_folder', None) + self.last_commit_id = kwargs.get('last_commit_id', None) + self.tenant_id = kwargs.get('tenant_id', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_vsts_configuration_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_vsts_configuration_py3.py new file mode 100644 index 000000000000..49d6fe8b7a29 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_vsts_configuration_py3.py @@ -0,0 +1,52 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class FactoryVSTSConfiguration(Model): + """Factory's VSTS repo information. + + :param account_name: VSTS account name. + :type account_name: str + :param project_name: VSTS project name. + :type project_name: str + :param repository_name: VSTS repository name. + :type repository_name: str + :param collaboration_branch: VSTS collaboration branch. + :type collaboration_branch: str + :param root_folder: VSTS root folder. + :type root_folder: str + :param last_commit_id: VSTS last commit id. + :type last_commit_id: str + :param tenant_id: VSTS tenant id. + :type tenant_id: str + """ + + _attribute_map = { + 'account_name': {'key': 'accountName', 'type': 'str'}, + 'project_name': {'key': 'projectName', 'type': 'str'}, + 'repository_name': {'key': 'repositoryName', 'type': 'str'}, + 'collaboration_branch': {'key': 'collaborationBranch', 'type': 'str'}, + 'root_folder': {'key': 'rootFolder', 'type': 'str'}, + 'last_commit_id': {'key': 'lastCommitId', 'type': 'str'}, + 'tenant_id': {'key': 'tenantId', 'type': 'str'}, + } + + def __init__(self, *, account_name: str=None, project_name: str=None, repository_name: str=None, collaboration_branch: str=None, root_folder: str=None, last_commit_id: str=None, tenant_id: str=None, **kwargs) -> None: + super(FactoryVSTSConfiguration, self).__init__(**kwargs) + self.account_name = account_name + self.project_name = project_name + self.repository_name = repository_name + self.collaboration_branch = collaboration_branch + self.root_folder = root_folder + self.last_commit_id = last_commit_id + self.tenant_id = tenant_id diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/factories_operations.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/factories_operations.py index 2714690eda35..0a64dbac9083 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/factories_operations.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/factories_operations.py @@ -100,6 +100,71 @@ def internal_paging(next_link=None, raw=False): return deserialized list.metadata = {'url': '/subscriptions/{subscriptionId}/providers/Microsoft.DataFactory/factories'} + def configure_factory_repo( + self, location_id, factory_repo_update, custom_headers=None, raw=False, **operation_config): + """Updates a factory's repo information. + + :param location_id: The location identifier. + :type location_id: str + :param factory_repo_update: Update factory repo request definition. + :type factory_repo_update: + ~azure.mgmt.datafactory.models.FactoryRepoUpdate + :param dict custom_headers: headers that will be added to the request + :param bool raw: returns the direct response alongside the + deserialized response + :param operation_config: :ref:`Operation configuration + overrides`. + :return: Factory or ClientRawResponse if raw=true + :rtype: ~azure.mgmt.datafactory.models.Factory or + ~msrest.pipeline.ClientRawResponse + :raises: + :class:`ErrorResponseException` + """ + # Construct URL + url = self.configure_factory_repo.metadata['url'] + path_format_arguments = { + 'subscriptionId': self._serialize.url("self.config.subscription_id", self.config.subscription_id, 'str'), + 'locationId': self._serialize.url("location_id", location_id, 'str') + } + url = self._client.format_url(url, **path_format_arguments) + + # Construct parameters + query_parameters = {} + query_parameters['api-version'] = self._serialize.query("self.api_version", self.api_version, 'str') + + # Construct headers + header_parameters = {} + header_parameters['Content-Type'] = 'application/json; charset=utf-8' + if self.config.generate_client_request_id: + header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) + if custom_headers: + header_parameters.update(custom_headers) + if self.config.accept_language is not None: + header_parameters['accept-language'] = self._serialize.header("self.config.accept_language", self.config.accept_language, 'str') + + # Construct body + body_content = self._serialize.body(factory_repo_update, 'FactoryRepoUpdate') + + # Construct and send request + request = self._client.post(url, query_parameters) + response = self._client.send( + request, header_parameters, body_content, stream=False, **operation_config) + + if response.status_code not in [200]: + raise models.ErrorResponseException(self._deserialize, response) + + deserialized = None + + if response.status_code == 200: + deserialized = self._deserialize('Factory', response) + + if raw: + client_raw_response = ClientRawResponse(deserialized, response) + return client_raw_response + + return deserialized + configure_factory_repo.metadata = {'url': '/subscriptions/{subscriptionId}/providers/Microsoft.DataFactory/locations/{locationId}/configureFactoryRepo'} + def list_by_resource_group( self, resource_group_name, custom_headers=None, raw=False, **operation_config): """Lists factories. From 27bfa3534c4b81a1d6d901ac3c73dd1bd9d125b0 Mon Sep 17 00:00:00 2001 From: Azure SDK for Python bot Date: Fri, 8 Jun 2018 11:24:58 -0700 Subject: [PATCH 06/54] [AutoPR datafactory/resource-manager] [Data Factory]Add schema mapping for hierarchical data (#2431) * Generated from 219315700287d09055dec44446b34f261d15d2c1 Add schema mapping for hierarchical data * Generated from a3c2bff7ee233f1de44032b1fccc27450874c77f Add example for column mapping * Generated from efe8a46867a169c9cd530d4562c56c5c3dba4da8 Merge remote-tracking branch 'upstream/master' --- .../models/quick_books_linked_service.py | 21 ++++++++++++++--- .../models/quick_books_linked_service_py3.py | 23 +++++++++++++++---- .../models/service_now_linked_service.py | 2 +- .../models/service_now_linked_service_py3.py | 2 +- .../datafactory/models/tabular_translator.py | 12 ++++++++-- .../models/tabular_translator_py3.py | 14 ++++++++--- 6 files changed, 60 insertions(+), 14 deletions(-) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_linked_service.py index a199f9763f70..c2ca123e5409 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_linked_service.py @@ -39,10 +39,17 @@ class QuickBooksLinkedService(LinkedService): :param company_id: Required. The company ID of the QuickBooks company to authorize. :type company_id: object - :param access_token: The access token for OAuth 1.0 authentication. - :type access_token: ~azure.mgmt.datafactory.models.SecretBase - :param access_token_secret: The access token secret for OAuth 1.0 + :param consumer_key: Required. The consumer key for OAuth 1.0 + authentication. + :type consumer_key: object + :param consumer_secret: Required. The consumer secret for OAuth 1.0 authentication. + :type consumer_secret: ~azure.mgmt.datafactory.models.SecretBase + :param access_token: Required. The access token for OAuth 1.0 + authentication. + :type access_token: ~azure.mgmt.datafactory.models.SecretBase + :param access_token_secret: Required. The access token secret for OAuth + 1.0 authentication. :type access_token_secret: ~azure.mgmt.datafactory.models.SecretBase :param use_encrypted_endpoints: Specifies whether the data source endpoints are encrypted using HTTPS. The default value is true. @@ -57,6 +64,10 @@ class QuickBooksLinkedService(LinkedService): 'type': {'required': True}, 'endpoint': {'required': True}, 'company_id': {'required': True}, + 'consumer_key': {'required': True}, + 'consumer_secret': {'required': True}, + 'access_token': {'required': True}, + 'access_token_secret': {'required': True}, } _attribute_map = { @@ -68,6 +79,8 @@ class QuickBooksLinkedService(LinkedService): 'type': {'key': 'type', 'type': 'str'}, 'endpoint': {'key': 'typeProperties.endpoint', 'type': 'object'}, 'company_id': {'key': 'typeProperties.companyId', 'type': 'object'}, + 'consumer_key': {'key': 'typeProperties.consumerKey', 'type': 'object'}, + 'consumer_secret': {'key': 'typeProperties.consumerSecret', 'type': 'SecretBase'}, 'access_token': {'key': 'typeProperties.accessToken', 'type': 'SecretBase'}, 'access_token_secret': {'key': 'typeProperties.accessTokenSecret', 'type': 'SecretBase'}, 'use_encrypted_endpoints': {'key': 'typeProperties.useEncryptedEndpoints', 'type': 'object'}, @@ -78,6 +91,8 @@ def __init__(self, **kwargs): super(QuickBooksLinkedService, self).__init__(**kwargs) self.endpoint = kwargs.get('endpoint', None) self.company_id = kwargs.get('company_id', None) + self.consumer_key = kwargs.get('consumer_key', None) + self.consumer_secret = kwargs.get('consumer_secret', None) self.access_token = kwargs.get('access_token', None) self.access_token_secret = kwargs.get('access_token_secret', None) self.use_encrypted_endpoints = kwargs.get('use_encrypted_endpoints', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_linked_service_py3.py index 8c23934112bd..7ba9f145c26e 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_linked_service_py3.py @@ -39,10 +39,17 @@ class QuickBooksLinkedService(LinkedService): :param company_id: Required. The company ID of the QuickBooks company to authorize. :type company_id: object - :param access_token: The access token for OAuth 1.0 authentication. - :type access_token: ~azure.mgmt.datafactory.models.SecretBase - :param access_token_secret: The access token secret for OAuth 1.0 + :param consumer_key: Required. The consumer key for OAuth 1.0 + authentication. + :type consumer_key: object + :param consumer_secret: Required. The consumer secret for OAuth 1.0 authentication. + :type consumer_secret: ~azure.mgmt.datafactory.models.SecretBase + :param access_token: Required. The access token for OAuth 1.0 + authentication. + :type access_token: ~azure.mgmt.datafactory.models.SecretBase + :param access_token_secret: Required. The access token secret for OAuth + 1.0 authentication. :type access_token_secret: ~azure.mgmt.datafactory.models.SecretBase :param use_encrypted_endpoints: Specifies whether the data source endpoints are encrypted using HTTPS. The default value is true. @@ -57,6 +64,10 @@ class QuickBooksLinkedService(LinkedService): 'type': {'required': True}, 'endpoint': {'required': True}, 'company_id': {'required': True}, + 'consumer_key': {'required': True}, + 'consumer_secret': {'required': True}, + 'access_token': {'required': True}, + 'access_token_secret': {'required': True}, } _attribute_map = { @@ -68,16 +79,20 @@ class QuickBooksLinkedService(LinkedService): 'type': {'key': 'type', 'type': 'str'}, 'endpoint': {'key': 'typeProperties.endpoint', 'type': 'object'}, 'company_id': {'key': 'typeProperties.companyId', 'type': 'object'}, + 'consumer_key': {'key': 'typeProperties.consumerKey', 'type': 'object'}, + 'consumer_secret': {'key': 'typeProperties.consumerSecret', 'type': 'SecretBase'}, 'access_token': {'key': 'typeProperties.accessToken', 'type': 'SecretBase'}, 'access_token_secret': {'key': 'typeProperties.accessTokenSecret', 'type': 'SecretBase'}, 'use_encrypted_endpoints': {'key': 'typeProperties.useEncryptedEndpoints', 'type': 'object'}, 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, *, endpoint, company_id, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, access_token=None, access_token_secret=None, use_encrypted_endpoints=None, encrypted_credential=None, **kwargs) -> None: + def __init__(self, *, endpoint, company_id, consumer_key, consumer_secret, access_token, access_token_secret, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, use_encrypted_endpoints=None, encrypted_credential=None, **kwargs) -> None: super(QuickBooksLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) self.endpoint = endpoint self.company_id = company_id + self.consumer_key = consumer_key + self.consumer_secret = consumer_secret self.access_token = access_token self.access_token_secret = access_token_secret self.use_encrypted_endpoints = use_encrypted_endpoints diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_linked_service.py index 0e180ce83506..c433366826b8 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_linked_service.py @@ -34,7 +34,7 @@ class ServiceNowLinkedService(LinkedService): :param type: Required. Constant filled by server. :type type: str :param endpoint: Required. The endpoint of the ServiceNow server. (i.e. - ServiceNowData.com) + .service-now.com) :type endpoint: object :param authentication_type: Required. The authentication type to use. Possible values include: 'Basic', 'OAuth2' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_linked_service_py3.py index 854468f281c0..cdd9e8ebb718 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_linked_service_py3.py @@ -34,7 +34,7 @@ class ServiceNowLinkedService(LinkedService): :param type: Required. Constant filled by server. :type type: str :param endpoint: Required. The endpoint of the ServiceNow server. (i.e. - ServiceNowData.com) + .service-now.com) :type endpoint: object :param authentication_type: Required. The authentication type to use. Possible values include: 'Basic', 'OAuth2' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/tabular_translator.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/tabular_translator.py index ec499cceee24..fdd098ae9659 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/tabular_translator.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/tabular_translator.py @@ -22,9 +22,15 @@ class TabularTranslator(CopyTranslator): :type additional_properties: dict[str, object] :param type: Required. Constant filled by server. :type type: str - :param column_mappings: Column mappings. Type: string (or Expression with - resultType string). + :param column_mappings: Column mappings. Example: "UserId: MyUserId, + Group: MyGroup, Name: MyName" Type: string (or Expression with resultType + string). :type column_mappings: object + :param schema_mapping: The schema mapping to map between tabular data and + hierarchical data. Example: {"Column1": "$.Column1", "Column2": + "$.Column2.Property1", "Column3": "$.Column2.Property2"}. Type: object (or + Expression with resultType object). + :type schema_mapping: object """ _validation = { @@ -35,9 +41,11 @@ class TabularTranslator(CopyTranslator): 'additional_properties': {'key': '', 'type': '{object}'}, 'type': {'key': 'type', 'type': 'str'}, 'column_mappings': {'key': 'columnMappings', 'type': 'object'}, + 'schema_mapping': {'key': 'schemaMapping', 'type': 'object'}, } def __init__(self, **kwargs): super(TabularTranslator, self).__init__(**kwargs) self.column_mappings = kwargs.get('column_mappings', None) + self.schema_mapping = kwargs.get('schema_mapping', None) self.type = 'TabularTranslator' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/tabular_translator_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/tabular_translator_py3.py index 45eae0147b20..0bd2ce51a0f0 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/tabular_translator_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/tabular_translator_py3.py @@ -22,9 +22,15 @@ class TabularTranslator(CopyTranslator): :type additional_properties: dict[str, object] :param type: Required. Constant filled by server. :type type: str - :param column_mappings: Column mappings. Type: string (or Expression with - resultType string). + :param column_mappings: Column mappings. Example: "UserId: MyUserId, + Group: MyGroup, Name: MyName" Type: string (or Expression with resultType + string). :type column_mappings: object + :param schema_mapping: The schema mapping to map between tabular data and + hierarchical data. Example: {"Column1": "$.Column1", "Column2": + "$.Column2.Property1", "Column3": "$.Column2.Property2"}. Type: object (or + Expression with resultType object). + :type schema_mapping: object """ _validation = { @@ -35,9 +41,11 @@ class TabularTranslator(CopyTranslator): 'additional_properties': {'key': '', 'type': '{object}'}, 'type': {'key': 'type', 'type': 'str'}, 'column_mappings': {'key': 'columnMappings', 'type': 'object'}, + 'schema_mapping': {'key': 'schemaMapping', 'type': 'object'}, } - def __init__(self, *, additional_properties=None, column_mappings=None, **kwargs) -> None: + def __init__(self, *, additional_properties=None, column_mappings=None, schema_mapping=None, **kwargs) -> None: super(TabularTranslator, self).__init__(additional_properties=additional_properties, **kwargs) self.column_mappings = column_mappings + self.schema_mapping = schema_mapping self.type = 'TabularTranslator' From cfbe230c619db63940f54ef455cd14ab0b7ffbba Mon Sep 17 00:00:00 2001 From: Azure SDK for Python bot Date: Fri, 15 Jun 2018 11:19:37 -0700 Subject: [PATCH 07/54] Generated from 932f3adb13821639c173dc647dd024875372238a (#2748) Remove maximum maxParallelExecutionsPerNode validation --- .../models/integration_runtime_compute_properties.py | 2 +- .../models/integration_runtime_compute_properties_py3.py | 2 +- 2 files changed, 2 insertions(+), 2 deletions(-) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_compute_properties.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_compute_properties.py index 71723579c8aa..e387ef4077f2 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_compute_properties.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_compute_properties.py @@ -38,7 +38,7 @@ class IntegrationRuntimeComputeProperties(Model): _validation = { 'number_of_nodes': {'minimum': 1}, - 'max_parallel_executions_per_node': {'maximum': 8, 'minimum': 1}, + 'max_parallel_executions_per_node': {'minimum': 1}, } _attribute_map = { diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_compute_properties_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_compute_properties_py3.py index 6b9212633ee1..f47f339dd067 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_compute_properties_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_compute_properties_py3.py @@ -38,7 +38,7 @@ class IntegrationRuntimeComputeProperties(Model): _validation = { 'number_of_nodes': {'minimum': 1}, - 'max_parallel_executions_per_node': {'maximum': 8, 'minimum': 1}, + 'max_parallel_executions_per_node': {'minimum': 1}, } _attribute_map = { From f41d2095ac5819a1bfdb30c02c7bfb39551e08f5 Mon Sep 17 00:00:00 2001 From: Azure SDK for Python bot Date: Mon, 2 Jul 2018 11:47:17 -0700 Subject: [PATCH 08/54] [AutoPR datafactory/resource-manager] Updating the default package tag for ADF to latest API (#2849) * Generated from 8393549a4c4950496639ed890fb994a16ed1613b Updating default ADF package tag to GA * Generated from ad7c2a0bffc9aa83f9dca26810a8882131336c64 Python version --- .../data_factory_management_client.py | 7 +- .../azure/mgmt/datafactory/models/__init__.py | 88 ++++---- .../azure/mgmt/datafactory/models/activity.py | 4 + .../mgmt/datafactory/models/activity_py3.py | 6 +- .../datafactory/models/activity_run_paged.py | 27 --- .../models/activity_runs_query_response.py | 39 ++++ .../activity_runs_query_response_py3.py | 39 ++++ .../azure_ml_batch_execution_activity.py | 3 + .../azure_ml_batch_execution_activity_py3.py | 7 +- .../azure_ml_update_resource_activity.py | 3 + .../azure_ml_update_resource_activity_py3.py | 7 +- .../datafactory/models/control_activity.py | 3 + .../models/control_activity_py3.py | 7 +- .../mgmt/datafactory/models/copy_activity.py | 13 +- .../datafactory/models/copy_activity_py3.py | 17 +- .../datafactory/models/custom_activity.py | 3 + .../datafactory/models/custom_activity_py3.py | 7 +- .../data_factory_management_client_enums.py | 21 +- .../data_lake_analytics_usql_activity.py | 3 + .../data_lake_analytics_usql_activity_py3.py | 7 +- .../models/databricks_notebook_activity.py | 3 + .../databricks_notebook_activity_py3.py | 7 +- .../mgmt/datafactory/models/error_response.py | 60 ------ .../datafactory/models/error_response_py3.py | 60 ------ .../models/execute_pipeline_activity.py | 3 + .../models/execute_pipeline_activity_py3.py | 7 +- .../models/execute_ssis_package_activity.py | 3 + .../execute_ssis_package_activity_py3.py | 7 +- .../datafactory/models/execution_activity.py | 3 + .../models/execution_activity_py3.py | 7 +- .../azure/mgmt/datafactory/models/factory.py | 14 +- .../models/factory_git_hub_configuration.py | 61 ++++++ .../factory_git_hub_configuration_py3.py | 61 ++++++ .../mgmt/datafactory/models/factory_py3.py | 16 +- .../models/factory_repo_configuration.py | 65 ++++++ .../models/factory_repo_configuration_py3.py | 65 ++++++ .../datafactory/models/factory_repo_update.py | 16 +- .../models/factory_repo_update_py3.py | 18 +- .../models/factory_vsts_configuration.py | 40 ++-- .../models/factory_vsts_configuration_py3.py | 44 ++-- .../datafactory/models/filter_activity.py | 3 + .../datafactory/models/filter_activity_py3.py | 7 +- .../datafactory/models/for_each_activity.py | 3 + .../models/for_each_activity_py3.py | 7 +- .../models/get_metadata_activity.py | 3 + .../models/get_metadata_activity_py3.py | 7 +- .../models/hd_insight_hive_activity.py | 3 + .../models/hd_insight_hive_activity_py3.py | 7 +- .../models/hd_insight_map_reduce_activity.py | 3 + .../hd_insight_map_reduce_activity_py3.py | 7 +- .../models/hd_insight_pig_activity.py | 3 + .../models/hd_insight_pig_activity_py3.py | 7 +- .../models/hd_insight_spark_activity.py | 3 + .../models/hd_insight_spark_activity_py3.py | 7 +- .../models/hd_insight_streaming_activity.py | 3 + .../hd_insight_streaming_activity_py3.py | 7 +- .../models/if_condition_activity.py | 3 + .../models/if_condition_activity_py3.py | 7 +- ...ntegration_runtime_node_monitoring_data.py | 4 +- ...ration_runtime_node_monitoring_data_py3.py | 4 +- ...integration_runtime_remove_node_request.py | 33 --- ...gration_runtime_remove_node_request_py3.py | 33 --- .../models/linked_integration_runtime.py | 58 ------ .../models/linked_integration_runtime_key.py | 39 ---- .../linked_integration_runtime_key_py3.py | 39 ---- .../linked_integration_runtime_properties.py | 41 ---- ...nked_integration_runtime_properties_py3.py | 41 ---- .../models/linked_integration_runtime_py3.py | 58 ------ .../models/linked_integration_runtime_rbac.py | 40 ---- .../linked_integration_runtime_rbac_py3.py | 40 ---- .../datafactory/models/lookup_activity.py | 3 + .../datafactory/models/lookup_activity_py3.py | 7 +- .../models/operation_list_response.py | 34 ---- .../models/operation_list_response_py3.py | 34 ---- .../models/operation_metric_dimension.py | 37 ++++ .../models/operation_metric_dimension_py3.py | 37 ++++ .../models/operation_metric_specification.py | 5 + .../operation_metric_specification_py3.py | 7 +- ...rigger_run_paged.py => operation_paged.py} | 8 +- .../models/pipeline_run_invoked_by.py | 5 + .../models/pipeline_run_invoked_by_py3.py | 5 + ...nse.py => pipeline_runs_query_response.py} | 4 +- ...py => pipeline_runs_query_response_py3.py} | 4 +- .../azure/mgmt/datafactory/models/resource.py | 5 + .../mgmt/datafactory/models/resource_py3.py | 5 + ...parameters.py => run_filter_parameters.py} | 23 +-- ...rs_py3.py => run_filter_parameters_py3.py} | 23 +-- ...un_query_filter.py => run_query_filter.py} | 21 +- ..._filter_py3.py => run_query_filter_py3.py} | 21 +- ...uery_order_by.py => run_query_order_by.py} | 20 +- ...er_by_py3.py => run_query_order_by_py3.py} | 20 +- .../models/self_hosted_integration_runtime.py | 5 - .../self_hosted_integration_runtime_py3.py | 7 +- .../self_hosted_integration_runtime_status.py | 16 +- ...f_hosted_integration_runtime_status_py3.py | 18 +- .../sql_server_stored_procedure_activity.py | 3 + ...ql_server_stored_procedure_activity_py3.py | 7 +- .../models/trigger_runs_query_response.py | 39 ++++ .../models/trigger_runs_query_response_py3.py | 39 ++++ .../mgmt/datafactory/models/until_activity.py | 3 + .../datafactory/models/until_activity_py3.py | 7 +- .../mgmt/datafactory/models/wait_activity.py | 3 + .../datafactory/models/wait_activity_py3.py | 7 +- .../mgmt/datafactory/models/web_activity.py | 3 + .../datafactory/models/web_activity_py3.py | 7 +- .../mgmt/datafactory/operations/__init__.py | 2 + .../operations/activity_runs_operations.py | 130 ++++++------ .../operations/datasets_operations.py | 43 ++-- .../operations/factories_operations.py | 140 +++++-------- .../integration_runtime_nodes_operations.py | 26 ++- .../integration_runtimes_operations.py | 188 +++++++----------- .../operations/linked_services_operations.py | 43 ++-- .../mgmt/datafactory/operations/operations.py | 71 ++++--- .../operations/pipeline_runs_operations.py | 88 ++++++-- .../operations/pipelines_operations.py | 65 +++--- .../operations/trigger_runs_operations.py | 107 ++++++++++ .../operations/triggers_operations.py | 136 ++++--------- .../azure/mgmt/datafactory/version.py | 2 +- 118 files changed, 1530 insertions(+), 1394 deletions(-) delete mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/activity_run_paged.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/activity_runs_query_response.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/activity_runs_query_response_py3.py delete mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/error_response.py delete mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/error_response_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_git_hub_configuration.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_git_hub_configuration_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_repo_configuration.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_repo_configuration_py3.py delete mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_remove_node_request.py delete mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_remove_node_request_py3.py delete mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime.py delete mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_key.py delete mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_key_py3.py delete mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_properties.py delete mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_properties_py3.py delete mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_py3.py delete mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_rbac.py delete mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_rbac_py3.py delete mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_list_response.py delete mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_list_response_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_metric_dimension.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_metric_dimension_py3.py rename azure-mgmt-datafactory/azure/mgmt/datafactory/models/{trigger_run_paged.py => operation_paged.py} (70%) rename azure-mgmt-datafactory/azure/mgmt/datafactory/models/{pipeline_run_query_response.py => pipeline_runs_query_response.py} (92%) rename azure-mgmt-datafactory/azure/mgmt/datafactory/models/{pipeline_run_query_response_py3.py => pipeline_runs_query_response_py3.py} (92%) rename azure-mgmt-datafactory/azure/mgmt/datafactory/models/{pipeline_run_filter_parameters.py => run_filter_parameters.py} (74%) rename azure-mgmt-datafactory/azure/mgmt/datafactory/models/{pipeline_run_filter_parameters_py3.py => run_filter_parameters_py3.py} (74%) rename azure-mgmt-datafactory/azure/mgmt/datafactory/models/{pipeline_run_query_filter.py => run_query_filter.py} (65%) rename azure-mgmt-datafactory/azure/mgmt/datafactory/models/{pipeline_run_query_filter_py3.py => run_query_filter_py3.py} (65%) rename azure-mgmt-datafactory/azure/mgmt/datafactory/models/{pipeline_run_query_order_by.py => run_query_order_by.py} (60%) rename azure-mgmt-datafactory/azure/mgmt/datafactory/models/{pipeline_run_query_order_by_py3.py => run_query_order_by_py3.py} (60%) create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger_runs_query_response.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger_runs_query_response_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/operations/trigger_runs_operations.py diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/data_factory_management_client.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/data_factory_management_client.py index 86680dcf55dc..ab57aa1c0bfa 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/data_factory_management_client.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/data_factory_management_client.py @@ -23,6 +23,7 @@ from .operations.pipeline_runs_operations import PipelineRunsOperations from .operations.activity_runs_operations import ActivityRunsOperations from .operations.triggers_operations import TriggersOperations +from .operations.trigger_runs_operations import TriggerRunsOperations from . import models @@ -84,6 +85,8 @@ class DataFactoryManagementClient(SDKClient): :vartype activity_runs: azure.mgmt.datafactory.operations.ActivityRunsOperations :ivar triggers: Triggers operations :vartype triggers: azure.mgmt.datafactory.operations.TriggersOperations + :ivar trigger_runs: TriggerRuns operations + :vartype trigger_runs: azure.mgmt.datafactory.operations.TriggerRunsOperations :param credentials: Credentials needed for the client to connect to Azure. :type credentials: :mod:`A msrestazure Credentials @@ -100,7 +103,7 @@ def __init__( super(DataFactoryManagementClient, self).__init__(self.config.credentials, self.config) client_models = {k: v for k, v in models.__dict__.items() if isinstance(v, type)} - self.api_version = '2017-09-01-preview' + self.api_version = '2018-06-01' self._serialize = Serializer(client_models) self._deserialize = Deserializer(client_models) @@ -124,3 +127,5 @@ def __init__( self._client, self.config, self._serialize, self._deserialize) self.triggers = TriggersOperations( self._client, self.config, self._serialize, self._deserialize) + self.trigger_runs = TriggerRunsOperations( + self._client, self.config, self._serialize, self._deserialize) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py index ed790633bf18..99f5c25e46e5 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py @@ -18,7 +18,7 @@ from .azure_key_vault_secret_reference_py3 import AzureKeyVaultSecretReference from .secret_base_py3 import SecretBase from .factory_identity_py3 import FactoryIdentity - from .factory_vsts_configuration_py3 import FactoryVSTSConfiguration + from .factory_repo_configuration_py3 import FactoryRepoConfiguration from .factory_py3 import Factory from .integration_runtime_py3 import IntegrationRuntime from .integration_runtime_resource_py3 import IntegrationRuntimeResource @@ -39,27 +39,30 @@ from .trigger_py3 import Trigger from .trigger_resource_py3 import TriggerResource from .create_run_response_py3 import CreateRunResponse - from .error_response_py3 import ErrorResponse, ErrorResponseException + from .factory_vsts_configuration_py3 import FactoryVSTSConfiguration + from .factory_git_hub_configuration_py3 import FactoryGitHubConfiguration from .factory_repo_update_py3 import FactoryRepoUpdate from .pipeline_reference_py3 import PipelineReference from .trigger_pipeline_reference_py3 import TriggerPipelineReference from .factory_update_parameters_py3 import FactoryUpdateParameters from .dataset_reference_py3 import DatasetReference - from .pipeline_run_query_filter_py3 import PipelineRunQueryFilter - from .pipeline_run_query_order_by_py3 import PipelineRunQueryOrderBy - from .pipeline_run_filter_parameters_py3 import PipelineRunFilterParameters + from .run_query_filter_py3 import RunQueryFilter + from .run_query_order_by_py3 import RunQueryOrderBy + from .run_filter_parameters_py3 import RunFilterParameters from .pipeline_run_invoked_by_py3 import PipelineRunInvokedBy from .pipeline_run_py3 import PipelineRun - from .pipeline_run_query_response_py3 import PipelineRunQueryResponse + from .pipeline_runs_query_response_py3 import PipelineRunsQueryResponse from .activity_run_py3 import ActivityRun + from .activity_runs_query_response_py3 import ActivityRunsQueryResponse from .trigger_run_py3 import TriggerRun + from .trigger_runs_query_response_py3 import TriggerRunsQueryResponse from .operation_display_py3 import OperationDisplay from .operation_log_specification_py3 import OperationLogSpecification from .operation_metric_availability_py3 import OperationMetricAvailability + from .operation_metric_dimension_py3 import OperationMetricDimension from .operation_metric_specification_py3 import OperationMetricSpecification from .operation_service_specification_py3 import OperationServiceSpecification from .operation_py3 import Operation - from .operation_list_response_py3 import OperationListResponse from .responsys_linked_service_py3 import ResponsysLinkedService from .azure_databricks_linked_service_py3 import AzureDatabricksLinkedService from .azure_data_lake_analytics_linked_service_py3 import AzureDataLakeAnalyticsLinkedService @@ -312,16 +315,12 @@ from .if_condition_activity_py3 import IfConditionActivity from .execute_pipeline_activity_py3 import ExecutePipelineActivity from .control_activity_py3 import ControlActivity - from .linked_integration_runtime_py3 import LinkedIntegrationRuntime from .self_hosted_integration_runtime_node_py3 import SelfHostedIntegrationRuntimeNode from .self_hosted_integration_runtime_status_py3 import SelfHostedIntegrationRuntimeStatus from .managed_integration_runtime_operation_result_py3 import ManagedIntegrationRuntimeOperationResult from .managed_integration_runtime_error_py3 import ManagedIntegrationRuntimeError from .managed_integration_runtime_node_py3 import ManagedIntegrationRuntimeNode from .managed_integration_runtime_status_py3 import ManagedIntegrationRuntimeStatus - from .linked_integration_runtime_rbac_py3 import LinkedIntegrationRuntimeRbac - from .linked_integration_runtime_key_py3 import LinkedIntegrationRuntimeKey - from .linked_integration_runtime_properties_py3 import LinkedIntegrationRuntimeProperties from .self_hosted_integration_runtime_py3 import SelfHostedIntegrationRuntime from .integration_runtime_custom_setup_script_properties_py3 import IntegrationRuntimeCustomSetupScriptProperties from .integration_runtime_ssis_catalog_info_py3 import IntegrationRuntimeSsisCatalogInfo @@ -332,7 +331,6 @@ from .integration_runtime_node_ip_address_py3 import IntegrationRuntimeNodeIpAddress from .integration_runtime_node_monitoring_data_py3 import IntegrationRuntimeNodeMonitoringData from .integration_runtime_monitoring_data_py3 import IntegrationRuntimeMonitoringData - from .integration_runtime_remove_node_request_py3 import IntegrationRuntimeRemoveNodeRequest from .integration_runtime_auth_keys_py3 import IntegrationRuntimeAuthKeys from .integration_runtime_regenerate_key_parameters_py3 import IntegrationRuntimeRegenerateKeyParameters from .integration_runtime_connection_info_py3 import IntegrationRuntimeConnectionInfo @@ -345,7 +343,7 @@ from .azure_key_vault_secret_reference import AzureKeyVaultSecretReference from .secret_base import SecretBase from .factory_identity import FactoryIdentity - from .factory_vsts_configuration import FactoryVSTSConfiguration + from .factory_repo_configuration import FactoryRepoConfiguration from .factory import Factory from .integration_runtime import IntegrationRuntime from .integration_runtime_resource import IntegrationRuntimeResource @@ -366,27 +364,30 @@ from .trigger import Trigger from .trigger_resource import TriggerResource from .create_run_response import CreateRunResponse - from .error_response import ErrorResponse, ErrorResponseException + from .factory_vsts_configuration import FactoryVSTSConfiguration + from .factory_git_hub_configuration import FactoryGitHubConfiguration from .factory_repo_update import FactoryRepoUpdate from .pipeline_reference import PipelineReference from .trigger_pipeline_reference import TriggerPipelineReference from .factory_update_parameters import FactoryUpdateParameters from .dataset_reference import DatasetReference - from .pipeline_run_query_filter import PipelineRunQueryFilter - from .pipeline_run_query_order_by import PipelineRunQueryOrderBy - from .pipeline_run_filter_parameters import PipelineRunFilterParameters + from .run_query_filter import RunQueryFilter + from .run_query_order_by import RunQueryOrderBy + from .run_filter_parameters import RunFilterParameters from .pipeline_run_invoked_by import PipelineRunInvokedBy from .pipeline_run import PipelineRun - from .pipeline_run_query_response import PipelineRunQueryResponse + from .pipeline_runs_query_response import PipelineRunsQueryResponse from .activity_run import ActivityRun + from .activity_runs_query_response import ActivityRunsQueryResponse from .trigger_run import TriggerRun + from .trigger_runs_query_response import TriggerRunsQueryResponse from .operation_display import OperationDisplay from .operation_log_specification import OperationLogSpecification from .operation_metric_availability import OperationMetricAvailability + from .operation_metric_dimension import OperationMetricDimension from .operation_metric_specification import OperationMetricSpecification from .operation_service_specification import OperationServiceSpecification from .operation import Operation - from .operation_list_response import OperationListResponse from .responsys_linked_service import ResponsysLinkedService from .azure_databricks_linked_service import AzureDatabricksLinkedService from .azure_data_lake_analytics_linked_service import AzureDataLakeAnalyticsLinkedService @@ -639,16 +640,12 @@ from .if_condition_activity import IfConditionActivity from .execute_pipeline_activity import ExecutePipelineActivity from .control_activity import ControlActivity - from .linked_integration_runtime import LinkedIntegrationRuntime from .self_hosted_integration_runtime_node import SelfHostedIntegrationRuntimeNode from .self_hosted_integration_runtime_status import SelfHostedIntegrationRuntimeStatus from .managed_integration_runtime_operation_result import ManagedIntegrationRuntimeOperationResult from .managed_integration_runtime_error import ManagedIntegrationRuntimeError from .managed_integration_runtime_node import ManagedIntegrationRuntimeNode from .managed_integration_runtime_status import ManagedIntegrationRuntimeStatus - from .linked_integration_runtime_rbac import LinkedIntegrationRuntimeRbac - from .linked_integration_runtime_key import LinkedIntegrationRuntimeKey - from .linked_integration_runtime_properties import LinkedIntegrationRuntimeProperties from .self_hosted_integration_runtime import SelfHostedIntegrationRuntime from .integration_runtime_custom_setup_script_properties import IntegrationRuntimeCustomSetupScriptProperties from .integration_runtime_ssis_catalog_info import IntegrationRuntimeSsisCatalogInfo @@ -659,28 +656,26 @@ from .integration_runtime_node_ip_address import IntegrationRuntimeNodeIpAddress from .integration_runtime_node_monitoring_data import IntegrationRuntimeNodeMonitoringData from .integration_runtime_monitoring_data import IntegrationRuntimeMonitoringData - from .integration_runtime_remove_node_request import IntegrationRuntimeRemoveNodeRequest from .integration_runtime_auth_keys import IntegrationRuntimeAuthKeys from .integration_runtime_regenerate_key_parameters import IntegrationRuntimeRegenerateKeyParameters from .integration_runtime_connection_info import IntegrationRuntimeConnectionInfo +from .operation_paged import OperationPaged from .factory_paged import FactoryPaged from .integration_runtime_resource_paged import IntegrationRuntimeResourcePaged from .linked_service_resource_paged import LinkedServiceResourcePaged from .dataset_resource_paged import DatasetResourcePaged from .pipeline_resource_paged import PipelineResourcePaged -from .activity_run_paged import ActivityRunPaged from .trigger_resource_paged import TriggerResourcePaged -from .trigger_run_paged import TriggerRunPaged from .data_factory_management_client_enums import ( IntegrationRuntimeState, IntegrationRuntimeAutoUpdate, ParameterType, DependencyCondition, TriggerRuntimeState, - PipelineRunQueryFilterOperand, - PipelineRunQueryFilterOperator, - PipelineRunQueryOrderByField, - PipelineRunQueryOrder, + RunQueryFilterOperand, + RunQueryFilterOperator, + RunQueryOrderByField, + RunQueryOrder, TriggerRunStatus, SparkServerType, SparkThriftTransportProtocol, @@ -741,7 +736,7 @@ 'AzureKeyVaultSecretReference', 'SecretBase', 'FactoryIdentity', - 'FactoryVSTSConfiguration', + 'FactoryRepoConfiguration', 'Factory', 'IntegrationRuntime', 'IntegrationRuntimeResource', @@ -762,27 +757,30 @@ 'Trigger', 'TriggerResource', 'CreateRunResponse', - 'ErrorResponse', 'ErrorResponseException', + 'FactoryVSTSConfiguration', + 'FactoryGitHubConfiguration', 'FactoryRepoUpdate', 'PipelineReference', 'TriggerPipelineReference', 'FactoryUpdateParameters', 'DatasetReference', - 'PipelineRunQueryFilter', - 'PipelineRunQueryOrderBy', - 'PipelineRunFilterParameters', + 'RunQueryFilter', + 'RunQueryOrderBy', + 'RunFilterParameters', 'PipelineRunInvokedBy', 'PipelineRun', - 'PipelineRunQueryResponse', + 'PipelineRunsQueryResponse', 'ActivityRun', + 'ActivityRunsQueryResponse', 'TriggerRun', + 'TriggerRunsQueryResponse', 'OperationDisplay', 'OperationLogSpecification', 'OperationMetricAvailability', + 'OperationMetricDimension', 'OperationMetricSpecification', 'OperationServiceSpecification', 'Operation', - 'OperationListResponse', 'ResponsysLinkedService', 'AzureDatabricksLinkedService', 'AzureDataLakeAnalyticsLinkedService', @@ -1035,16 +1033,12 @@ 'IfConditionActivity', 'ExecutePipelineActivity', 'ControlActivity', - 'LinkedIntegrationRuntime', 'SelfHostedIntegrationRuntimeNode', 'SelfHostedIntegrationRuntimeStatus', 'ManagedIntegrationRuntimeOperationResult', 'ManagedIntegrationRuntimeError', 'ManagedIntegrationRuntimeNode', 'ManagedIntegrationRuntimeStatus', - 'LinkedIntegrationRuntimeRbac', - 'LinkedIntegrationRuntimeKey', - 'LinkedIntegrationRuntimeProperties', 'SelfHostedIntegrationRuntime', 'IntegrationRuntimeCustomSetupScriptProperties', 'IntegrationRuntimeSsisCatalogInfo', @@ -1055,27 +1049,25 @@ 'IntegrationRuntimeNodeIpAddress', 'IntegrationRuntimeNodeMonitoringData', 'IntegrationRuntimeMonitoringData', - 'IntegrationRuntimeRemoveNodeRequest', 'IntegrationRuntimeAuthKeys', 'IntegrationRuntimeRegenerateKeyParameters', 'IntegrationRuntimeConnectionInfo', + 'OperationPaged', 'FactoryPaged', 'IntegrationRuntimeResourcePaged', 'LinkedServiceResourcePaged', 'DatasetResourcePaged', 'PipelineResourcePaged', - 'ActivityRunPaged', 'TriggerResourcePaged', - 'TriggerRunPaged', 'IntegrationRuntimeState', 'IntegrationRuntimeAutoUpdate', 'ParameterType', 'DependencyCondition', 'TriggerRuntimeState', - 'PipelineRunQueryFilterOperand', - 'PipelineRunQueryFilterOperator', - 'PipelineRunQueryOrderByField', - 'PipelineRunQueryOrder', + 'RunQueryFilterOperand', + 'RunQueryFilterOperator', + 'RunQueryOrderByField', + 'RunQueryOrder', 'TriggerRunStatus', 'SparkServerType', 'SparkThriftTransportProtocol', diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/activity.py index 602b8d52a38b..438e4b23f7eb 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/activity.py @@ -29,6 +29,8 @@ class Activity(Model): :type description: str :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param user_properties: Activity user properties. + :type user_properties: dict[str, str] :param type: Required. Constant filled by server. :type type: str """ @@ -43,6 +45,7 @@ class Activity(Model): 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'user_properties': {'key': 'userProperties', 'type': '{str}'}, 'type': {'key': 'type', 'type': 'str'}, } @@ -56,4 +59,5 @@ def __init__(self, **kwargs): self.name = kwargs.get('name', None) self.description = kwargs.get('description', None) self.depends_on = kwargs.get('depends_on', None) + self.user_properties = kwargs.get('user_properties', None) self.type = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/activity_py3.py index c1283c260908..48d13e8e631d 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/activity_py3.py @@ -29,6 +29,8 @@ class Activity(Model): :type description: str :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param user_properties: Activity user properties. + :type user_properties: dict[str, str] :param type: Required. Constant filled by server. :type type: str """ @@ -43,6 +45,7 @@ class Activity(Model): 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'user_properties': {'key': 'userProperties', 'type': '{str}'}, 'type': {'key': 'type', 'type': 'str'}, } @@ -50,10 +53,11 @@ class Activity(Model): 'type': {'Execution': 'ExecutionActivity', 'Container': 'ControlActivity'} } - def __init__(self, *, name: str, additional_properties=None, description: str=None, depends_on=None, **kwargs) -> None: + def __init__(self, *, name: str, additional_properties=None, description: str=None, depends_on=None, user_properties=None, **kwargs) -> None: super(Activity, self).__init__(**kwargs) self.additional_properties = additional_properties self.name = name self.description = description self.depends_on = depends_on + self.user_properties = user_properties self.type = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/activity_run_paged.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/activity_run_paged.py deleted file mode 100644 index ac1ccd554d32..000000000000 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/activity_run_paged.py +++ /dev/null @@ -1,27 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for -# license information. -# -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is -# regenerated. -# -------------------------------------------------------------------------- - -from msrest.paging import Paged - - -class ActivityRunPaged(Paged): - """ - A paging container for iterating over a list of :class:`ActivityRun ` object - """ - - _attribute_map = { - 'next_link': {'key': 'nextLink', 'type': 'str'}, - 'current_page': {'key': 'value', 'type': '[ActivityRun]'} - } - - def __init__(self, *args, **kwargs): - - super(ActivityRunPaged, self).__init__(*args, **kwargs) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/activity_runs_query_response.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/activity_runs_query_response.py new file mode 100644 index 000000000000..2fcd25a5ced2 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/activity_runs_query_response.py @@ -0,0 +1,39 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class ActivityRunsQueryResponse(Model): + """A list activity runs. + + All required parameters must be populated in order to send to Azure. + + :param value: Required. List of activity runs. + :type value: list[~azure.mgmt.datafactory.models.ActivityRun] + :param continuation_token: The continuation token for getting the next + page of results, if any remaining results exist, null otherwise. + :type continuation_token: str + """ + + _validation = { + 'value': {'required': True}, + } + + _attribute_map = { + 'value': {'key': 'value', 'type': '[ActivityRun]'}, + 'continuation_token': {'key': 'continuationToken', 'type': 'str'}, + } + + def __init__(self, **kwargs): + super(ActivityRunsQueryResponse, self).__init__(**kwargs) + self.value = kwargs.get('value', None) + self.continuation_token = kwargs.get('continuation_token', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/activity_runs_query_response_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/activity_runs_query_response_py3.py new file mode 100644 index 000000000000..ee3eae141635 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/activity_runs_query_response_py3.py @@ -0,0 +1,39 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class ActivityRunsQueryResponse(Model): + """A list activity runs. + + All required parameters must be populated in order to send to Azure. + + :param value: Required. List of activity runs. + :type value: list[~azure.mgmt.datafactory.models.ActivityRun] + :param continuation_token: The continuation token for getting the next + page of results, if any remaining results exist, null otherwise. + :type continuation_token: str + """ + + _validation = { + 'value': {'required': True}, + } + + _attribute_map = { + 'value': {'key': 'value', 'type': '[ActivityRun]'}, + 'continuation_token': {'key': 'continuationToken', 'type': 'str'}, + } + + def __init__(self, *, value, continuation_token: str=None, **kwargs) -> None: + super(ActivityRunsQueryResponse, self).__init__(**kwargs) + self.value = value + self.continuation_token = continuation_token diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_batch_execution_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_batch_execution_activity.py index 826846634438..e9c935e65205 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_batch_execution_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_batch_execution_activity.py @@ -26,6 +26,8 @@ class AzureMLBatchExecutionActivity(ExecutionActivity): :type description: str :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param user_properties: Activity user properties. + :type user_properties: dict[str, str] :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. @@ -63,6 +65,7 @@ class AzureMLBatchExecutionActivity(ExecutionActivity): 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'user_properties': {'key': 'userProperties', 'type': '{str}'}, 'type': {'key': 'type', 'type': 'str'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_batch_execution_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_batch_execution_activity_py3.py index 67f0f605ca5d..e455622c0a89 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_batch_execution_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_batch_execution_activity_py3.py @@ -26,6 +26,8 @@ class AzureMLBatchExecutionActivity(ExecutionActivity): :type description: str :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param user_properties: Activity user properties. + :type user_properties: dict[str, str] :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. @@ -63,6 +65,7 @@ class AzureMLBatchExecutionActivity(ExecutionActivity): 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'user_properties': {'key': 'userProperties', 'type': '{str}'}, 'type': {'key': 'type', 'type': 'str'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, @@ -71,8 +74,8 @@ class AzureMLBatchExecutionActivity(ExecutionActivity): 'web_service_inputs': {'key': 'typeProperties.webServiceInputs', 'type': '{AzureMLWebServiceFile}'}, } - def __init__(self, *, name: str, additional_properties=None, description: str=None, depends_on=None, linked_service_name=None, policy=None, global_parameters=None, web_service_outputs=None, web_service_inputs=None, **kwargs) -> None: - super(AzureMLBatchExecutionActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, linked_service_name=linked_service_name, policy=policy, **kwargs) + def __init__(self, *, name: str, additional_properties=None, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, global_parameters=None, web_service_outputs=None, web_service_inputs=None, **kwargs) -> None: + super(AzureMLBatchExecutionActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) self.global_parameters = global_parameters self.web_service_outputs = web_service_outputs self.web_service_inputs = web_service_inputs diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_update_resource_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_update_resource_activity.py index 2a6860f5f41e..6c809dd27d4d 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_update_resource_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_update_resource_activity.py @@ -26,6 +26,8 @@ class AzureMLUpdateResourceActivity(ExecutionActivity): :type description: str :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param user_properties: Activity user properties. + :type user_properties: dict[str, str] :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. @@ -62,6 +64,7 @@ class AzureMLUpdateResourceActivity(ExecutionActivity): 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'user_properties': {'key': 'userProperties', 'type': '{str}'}, 'type': {'key': 'type', 'type': 'str'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_update_resource_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_update_resource_activity_py3.py index ff0dce23464f..13cbc43d5de5 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_update_resource_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_update_resource_activity_py3.py @@ -26,6 +26,8 @@ class AzureMLUpdateResourceActivity(ExecutionActivity): :type description: str :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param user_properties: Activity user properties. + :type user_properties: dict[str, str] :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. @@ -62,6 +64,7 @@ class AzureMLUpdateResourceActivity(ExecutionActivity): 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'user_properties': {'key': 'userProperties', 'type': '{str}'}, 'type': {'key': 'type', 'type': 'str'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, @@ -70,8 +73,8 @@ class AzureMLUpdateResourceActivity(ExecutionActivity): 'trained_model_file_path': {'key': 'typeProperties.trainedModelFilePath', 'type': 'object'}, } - def __init__(self, *, name: str, trained_model_name, trained_model_linked_service_name, trained_model_file_path, additional_properties=None, description: str=None, depends_on=None, linked_service_name=None, policy=None, **kwargs) -> None: - super(AzureMLUpdateResourceActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, linked_service_name=linked_service_name, policy=policy, **kwargs) + def __init__(self, *, name: str, trained_model_name, trained_model_linked_service_name, trained_model_file_path, additional_properties=None, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, **kwargs) -> None: + super(AzureMLUpdateResourceActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) self.trained_model_name = trained_model_name self.trained_model_linked_service_name = trained_model_linked_service_name self.trained_model_file_path = trained_model_file_path diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/control_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/control_activity.py index 200740cbe757..b1915fc2e454 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/control_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/control_activity.py @@ -30,6 +30,8 @@ class ControlActivity(Activity): :type description: str :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param user_properties: Activity user properties. + :type user_properties: dict[str, str] :param type: Required. Constant filled by server. :type type: str """ @@ -44,6 +46,7 @@ class ControlActivity(Activity): 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'user_properties': {'key': 'userProperties', 'type': '{str}'}, 'type': {'key': 'type', 'type': 'str'}, } diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/control_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/control_activity_py3.py index 8623e8d151e2..929788720bc4 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/control_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/control_activity_py3.py @@ -30,6 +30,8 @@ class ControlActivity(Activity): :type description: str :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param user_properties: Activity user properties. + :type user_properties: dict[str, str] :param type: Required. Constant filled by server. :type type: str """ @@ -44,6 +46,7 @@ class ControlActivity(Activity): 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'user_properties': {'key': 'userProperties', 'type': '{str}'}, 'type': {'key': 'type', 'type': 'str'}, } @@ -51,6 +54,6 @@ class ControlActivity(Activity): 'type': {'Filter': 'FilterActivity', 'Until': 'UntilActivity', 'Wait': 'WaitActivity', 'ForEach': 'ForEachActivity', 'IfCondition': 'IfConditionActivity', 'ExecutePipeline': 'ExecutePipelineActivity'} } - def __init__(self, *, name: str, additional_properties=None, description: str=None, depends_on=None, **kwargs) -> None: - super(ControlActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, **kwargs) + def __init__(self, *, name: str, additional_properties=None, description: str=None, depends_on=None, user_properties=None, **kwargs) -> None: + super(ControlActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, **kwargs) self.type = 'Container' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_activity.py index e600efe42d06..e26170d07c14 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_activity.py @@ -26,6 +26,8 @@ class CopyActivity(ExecutionActivity): :type description: str :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param user_properties: Activity user properties. + :type user_properties: dict[str, str] :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. @@ -51,10 +53,10 @@ class CopyActivity(ExecutionActivity): the source or sink to avoid overloading the data store. Type: integer (or Expression with resultType integer), minimum: 0. :type parallel_copies: object - :param cloud_data_movement_units: Maximum number of cloud data movement - units that can be used to perform this data movement. Type: integer (or + :param data_integration_units: Maximum number of data integration units + that can be used to perform this data movement. Type: integer (or Expression with resultType integer), minimum: 0. - :type cloud_data_movement_units: object + :type data_integration_units: object :param enable_skip_incompatible_row: Whether to skip incompatible row. Default value is false. Type: boolean (or Expression with resultType boolean). @@ -81,6 +83,7 @@ class CopyActivity(ExecutionActivity): 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'user_properties': {'key': 'userProperties', 'type': '{str}'}, 'type': {'key': 'type', 'type': 'str'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, @@ -90,7 +93,7 @@ class CopyActivity(ExecutionActivity): 'enable_staging': {'key': 'typeProperties.enableStaging', 'type': 'object'}, 'staging_settings': {'key': 'typeProperties.stagingSettings', 'type': 'StagingSettings'}, 'parallel_copies': {'key': 'typeProperties.parallelCopies', 'type': 'object'}, - 'cloud_data_movement_units': {'key': 'typeProperties.cloudDataMovementUnits', 'type': 'object'}, + 'data_integration_units': {'key': 'typeProperties.dataIntegrationUnits', 'type': 'object'}, 'enable_skip_incompatible_row': {'key': 'typeProperties.enableSkipIncompatibleRow', 'type': 'object'}, 'redirect_incompatible_row_settings': {'key': 'typeProperties.redirectIncompatibleRowSettings', 'type': 'RedirectIncompatibleRowSettings'}, 'inputs': {'key': 'inputs', 'type': '[DatasetReference]'}, @@ -105,7 +108,7 @@ def __init__(self, **kwargs): self.enable_staging = kwargs.get('enable_staging', None) self.staging_settings = kwargs.get('staging_settings', None) self.parallel_copies = kwargs.get('parallel_copies', None) - self.cloud_data_movement_units = kwargs.get('cloud_data_movement_units', None) + self.data_integration_units = kwargs.get('data_integration_units', None) self.enable_skip_incompatible_row = kwargs.get('enable_skip_incompatible_row', None) self.redirect_incompatible_row_settings = kwargs.get('redirect_incompatible_row_settings', None) self.inputs = kwargs.get('inputs', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_activity_py3.py index 30dd29e42139..bdbd626bb317 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_activity_py3.py @@ -26,6 +26,8 @@ class CopyActivity(ExecutionActivity): :type description: str :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param user_properties: Activity user properties. + :type user_properties: dict[str, str] :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. @@ -51,10 +53,10 @@ class CopyActivity(ExecutionActivity): the source or sink to avoid overloading the data store. Type: integer (or Expression with resultType integer), minimum: 0. :type parallel_copies: object - :param cloud_data_movement_units: Maximum number of cloud data movement - units that can be used to perform this data movement. Type: integer (or + :param data_integration_units: Maximum number of data integration units + that can be used to perform this data movement. Type: integer (or Expression with resultType integer), minimum: 0. - :type cloud_data_movement_units: object + :type data_integration_units: object :param enable_skip_incompatible_row: Whether to skip incompatible row. Default value is false. Type: boolean (or Expression with resultType boolean). @@ -81,6 +83,7 @@ class CopyActivity(ExecutionActivity): 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'user_properties': {'key': 'userProperties', 'type': '{str}'}, 'type': {'key': 'type', 'type': 'str'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, @@ -90,22 +93,22 @@ class CopyActivity(ExecutionActivity): 'enable_staging': {'key': 'typeProperties.enableStaging', 'type': 'object'}, 'staging_settings': {'key': 'typeProperties.stagingSettings', 'type': 'StagingSettings'}, 'parallel_copies': {'key': 'typeProperties.parallelCopies', 'type': 'object'}, - 'cloud_data_movement_units': {'key': 'typeProperties.cloudDataMovementUnits', 'type': 'object'}, + 'data_integration_units': {'key': 'typeProperties.dataIntegrationUnits', 'type': 'object'}, 'enable_skip_incompatible_row': {'key': 'typeProperties.enableSkipIncompatibleRow', 'type': 'object'}, 'redirect_incompatible_row_settings': {'key': 'typeProperties.redirectIncompatibleRowSettings', 'type': 'RedirectIncompatibleRowSettings'}, 'inputs': {'key': 'inputs', 'type': '[DatasetReference]'}, 'outputs': {'key': 'outputs', 'type': '[DatasetReference]'}, } - def __init__(self, *, name: str, source, sink, additional_properties=None, description: str=None, depends_on=None, linked_service_name=None, policy=None, translator=None, enable_staging=None, staging_settings=None, parallel_copies=None, cloud_data_movement_units=None, enable_skip_incompatible_row=None, redirect_incompatible_row_settings=None, inputs=None, outputs=None, **kwargs) -> None: - super(CopyActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, linked_service_name=linked_service_name, policy=policy, **kwargs) + def __init__(self, *, name: str, source, sink, additional_properties=None, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, translator=None, enable_staging=None, staging_settings=None, parallel_copies=None, data_integration_units=None, enable_skip_incompatible_row=None, redirect_incompatible_row_settings=None, inputs=None, outputs=None, **kwargs) -> None: + super(CopyActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) self.source = source self.sink = sink self.translator = translator self.enable_staging = enable_staging self.staging_settings = staging_settings self.parallel_copies = parallel_copies - self.cloud_data_movement_units = cloud_data_movement_units + self.data_integration_units = data_integration_units self.enable_skip_incompatible_row = enable_skip_incompatible_row self.redirect_incompatible_row_settings = redirect_incompatible_row_settings self.inputs = inputs diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_activity.py index 2ebe9c2bc401..9c0d65a18a16 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_activity.py @@ -26,6 +26,8 @@ class CustomActivity(ExecutionActivity): :type description: str :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param user_properties: Activity user properties. + :type user_properties: dict[str, str] :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. @@ -63,6 +65,7 @@ class CustomActivity(ExecutionActivity): 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'user_properties': {'key': 'userProperties', 'type': '{str}'}, 'type': {'key': 'type', 'type': 'str'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_activity_py3.py index 10542a55610f..abaee4d53210 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_activity_py3.py @@ -26,6 +26,8 @@ class CustomActivity(ExecutionActivity): :type description: str :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param user_properties: Activity user properties. + :type user_properties: dict[str, str] :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. @@ -63,6 +65,7 @@ class CustomActivity(ExecutionActivity): 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'user_properties': {'key': 'userProperties', 'type': '{str}'}, 'type': {'key': 'type', 'type': 'str'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, @@ -73,8 +76,8 @@ class CustomActivity(ExecutionActivity): 'extended_properties': {'key': 'typeProperties.extendedProperties', 'type': '{object}'}, } - def __init__(self, *, name: str, command, additional_properties=None, description: str=None, depends_on=None, linked_service_name=None, policy=None, resource_linked_service=None, folder_path=None, reference_objects=None, extended_properties=None, **kwargs) -> None: - super(CustomActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, linked_service_name=linked_service_name, policy=policy, **kwargs) + def __init__(self, *, name: str, command, additional_properties=None, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, resource_linked_service=None, folder_path=None, reference_objects=None, extended_properties=None, **kwargs) -> None: + super(CustomActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) self.command = command self.resource_linked_service = resource_linked_service self.folder_path = folder_path diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_factory_management_client_enums.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_factory_management_client_enums.py index fb948ae66a19..a7eb9a135489 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_factory_management_client_enums.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_factory_management_client_enums.py @@ -57,15 +57,21 @@ class TriggerRuntimeState(str, Enum): disabled = "Disabled" -class PipelineRunQueryFilterOperand(str, Enum): +class RunQueryFilterOperand(str, Enum): pipeline_name = "PipelineName" status = "Status" run_start = "RunStart" run_end = "RunEnd" + activity_name = "ActivityName" + activity_run_start = "ActivityRunStart" + activity_run_end = "ActivityRunEnd" + activity_type = "ActivityType" + trigger_name = "TriggerName" + trigger_run_timestamp = "TriggerRunTimestamp" -class PipelineRunQueryFilterOperator(str, Enum): +class RunQueryFilterOperator(str, Enum): equals = "Equals" not_equals = "NotEquals" @@ -73,13 +79,20 @@ class PipelineRunQueryFilterOperator(str, Enum): not_in = "NotIn" -class PipelineRunQueryOrderByField(str, Enum): +class RunQueryOrderByField(str, Enum): run_start = "RunStart" run_end = "RunEnd" + pipeline_name = "PipelineName" + status = "Status" + activity_name = "ActivityName" + activity_run_start = "ActivityRunStart" + activity_run_end = "ActivityRunEnd" + trigger_name = "TriggerName" + trigger_run_timestamp = "TriggerRunTimestamp" -class PipelineRunQueryOrder(str, Enum): +class RunQueryOrder(str, Enum): asc = "ASC" desc = "DESC" diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_lake_analytics_usql_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_lake_analytics_usql_activity.py index a4b8dcdc70ab..e35da3e25273 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_lake_analytics_usql_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_lake_analytics_usql_activity.py @@ -26,6 +26,8 @@ class DataLakeAnalyticsUSQLActivity(ExecutionActivity): :type description: str :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param user_properties: Activity user properties. + :type user_properties: dict[str, str] :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. @@ -71,6 +73,7 @@ class DataLakeAnalyticsUSQLActivity(ExecutionActivity): 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'user_properties': {'key': 'userProperties', 'type': '{str}'}, 'type': {'key': 'type', 'type': 'str'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_lake_analytics_usql_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_lake_analytics_usql_activity_py3.py index 3b0ff6dbe94f..5e4e4968b535 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_lake_analytics_usql_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_lake_analytics_usql_activity_py3.py @@ -26,6 +26,8 @@ class DataLakeAnalyticsUSQLActivity(ExecutionActivity): :type description: str :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param user_properties: Activity user properties. + :type user_properties: dict[str, str] :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. @@ -71,6 +73,7 @@ class DataLakeAnalyticsUSQLActivity(ExecutionActivity): 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'user_properties': {'key': 'userProperties', 'type': '{str}'}, 'type': {'key': 'type', 'type': 'str'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, @@ -83,8 +86,8 @@ class DataLakeAnalyticsUSQLActivity(ExecutionActivity): 'compilation_mode': {'key': 'typeProperties.compilationMode', 'type': 'object'}, } - def __init__(self, *, name: str, script_path, script_linked_service, additional_properties=None, description: str=None, depends_on=None, linked_service_name=None, policy=None, degree_of_parallelism=None, priority=None, parameters=None, runtime_version=None, compilation_mode=None, **kwargs) -> None: - super(DataLakeAnalyticsUSQLActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, linked_service_name=linked_service_name, policy=policy, **kwargs) + def __init__(self, *, name: str, script_path, script_linked_service, additional_properties=None, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, degree_of_parallelism=None, priority=None, parameters=None, runtime_version=None, compilation_mode=None, **kwargs) -> None: + super(DataLakeAnalyticsUSQLActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) self.script_path = script_path self.script_linked_service = script_linked_service self.degree_of_parallelism = degree_of_parallelism diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_notebook_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_notebook_activity.py index 136ee7debaca..fa0dfc114141 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_notebook_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_notebook_activity.py @@ -26,6 +26,8 @@ class DatabricksNotebookActivity(ExecutionActivity): :type description: str :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param user_properties: Activity user properties. + :type user_properties: dict[str, str] :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. @@ -54,6 +56,7 @@ class DatabricksNotebookActivity(ExecutionActivity): 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'user_properties': {'key': 'userProperties', 'type': '{str}'}, 'type': {'key': 'type', 'type': 'str'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_notebook_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_notebook_activity_py3.py index cefdc1681830..2b0ed28abd87 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_notebook_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_notebook_activity_py3.py @@ -26,6 +26,8 @@ class DatabricksNotebookActivity(ExecutionActivity): :type description: str :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param user_properties: Activity user properties. + :type user_properties: dict[str, str] :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. @@ -54,6 +56,7 @@ class DatabricksNotebookActivity(ExecutionActivity): 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'user_properties': {'key': 'userProperties', 'type': '{str}'}, 'type': {'key': 'type', 'type': 'str'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, @@ -61,8 +64,8 @@ class DatabricksNotebookActivity(ExecutionActivity): 'base_parameters': {'key': 'typeProperties.baseParameters', 'type': '{object}'}, } - def __init__(self, *, name: str, notebook_path, additional_properties=None, description: str=None, depends_on=None, linked_service_name=None, policy=None, base_parameters=None, **kwargs) -> None: - super(DatabricksNotebookActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, linked_service_name=linked_service_name, policy=policy, **kwargs) + def __init__(self, *, name: str, notebook_path, additional_properties=None, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, base_parameters=None, **kwargs) -> None: + super(DatabricksNotebookActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) self.notebook_path = notebook_path self.base_parameters = base_parameters self.type = 'DatabricksNotebook' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/error_response.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/error_response.py deleted file mode 100644 index 2a48122fa578..000000000000 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/error_response.py +++ /dev/null @@ -1,60 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for -# license information. -# -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is -# regenerated. -# -------------------------------------------------------------------------- - -from msrest.serialization import Model -from msrest.exceptions import HttpOperationError - - -class ErrorResponse(Model): - """The object that defines the structure of an Azure Data Factory response. - - All required parameters must be populated in order to send to Azure. - - :param code: Required. Error code. - :type code: str - :param message: Required. Error message. - :type message: str - :param target: Property name/path in request associated with error. - :type target: str - :param details: Array with additional error details. - :type details: list[~azure.mgmt.datafactory.models.ErrorResponse] - """ - - _validation = { - 'code': {'required': True}, - 'message': {'required': True}, - } - - _attribute_map = { - 'code': {'key': 'code', 'type': 'str'}, - 'message': {'key': 'message', 'type': 'str'}, - 'target': {'key': 'target', 'type': 'str'}, - 'details': {'key': 'details', 'type': '[ErrorResponse]'}, - } - - def __init__(self, **kwargs): - super(ErrorResponse, self).__init__(**kwargs) - self.code = kwargs.get('code', None) - self.message = kwargs.get('message', None) - self.target = kwargs.get('target', None) - self.details = kwargs.get('details', None) - - -class ErrorResponseException(HttpOperationError): - """Server responsed with exception of type: 'ErrorResponse'. - - :param deserialize: A deserializer - :param response: Server response to be deserialized. - """ - - def __init__(self, deserialize, response, *args): - - super(ErrorResponseException, self).__init__(deserialize, response, 'ErrorResponse', *args) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/error_response_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/error_response_py3.py deleted file mode 100644 index 6b36d51c90df..000000000000 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/error_response_py3.py +++ /dev/null @@ -1,60 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for -# license information. -# -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is -# regenerated. -# -------------------------------------------------------------------------- - -from msrest.serialization import Model -from msrest.exceptions import HttpOperationError - - -class ErrorResponse(Model): - """The object that defines the structure of an Azure Data Factory response. - - All required parameters must be populated in order to send to Azure. - - :param code: Required. Error code. - :type code: str - :param message: Required. Error message. - :type message: str - :param target: Property name/path in request associated with error. - :type target: str - :param details: Array with additional error details. - :type details: list[~azure.mgmt.datafactory.models.ErrorResponse] - """ - - _validation = { - 'code': {'required': True}, - 'message': {'required': True}, - } - - _attribute_map = { - 'code': {'key': 'code', 'type': 'str'}, - 'message': {'key': 'message', 'type': 'str'}, - 'target': {'key': 'target', 'type': 'str'}, - 'details': {'key': 'details', 'type': '[ErrorResponse]'}, - } - - def __init__(self, *, code: str, message: str, target: str=None, details=None, **kwargs) -> None: - super(ErrorResponse, self).__init__(**kwargs) - self.code = code - self.message = message - self.target = target - self.details = details - - -class ErrorResponseException(HttpOperationError): - """Server responsed with exception of type: 'ErrorResponse'. - - :param deserialize: A deserializer - :param response: Server response to be deserialized. - """ - - def __init__(self, deserialize, response, *args): - - super(ErrorResponseException, self).__init__(deserialize, response, 'ErrorResponse', *args) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_pipeline_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_pipeline_activity.py index 4e3b48e75bdf..beab510d3367 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_pipeline_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_pipeline_activity.py @@ -26,6 +26,8 @@ class ExecutePipelineActivity(ControlActivity): :type description: str :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param user_properties: Activity user properties. + :type user_properties: dict[str, str] :param type: Required. Constant filled by server. :type type: str :param pipeline: Required. Pipeline reference. @@ -48,6 +50,7 @@ class ExecutePipelineActivity(ControlActivity): 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'user_properties': {'key': 'userProperties', 'type': '{str}'}, 'type': {'key': 'type', 'type': 'str'}, 'pipeline': {'key': 'typeProperties.pipeline', 'type': 'PipelineReference'}, 'parameters': {'key': 'typeProperties.parameters', 'type': '{object}'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_pipeline_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_pipeline_activity_py3.py index 2550875243b6..ef878fec54fc 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_pipeline_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_pipeline_activity_py3.py @@ -26,6 +26,8 @@ class ExecutePipelineActivity(ControlActivity): :type description: str :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param user_properties: Activity user properties. + :type user_properties: dict[str, str] :param type: Required. Constant filled by server. :type type: str :param pipeline: Required. Pipeline reference. @@ -48,14 +50,15 @@ class ExecutePipelineActivity(ControlActivity): 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'user_properties': {'key': 'userProperties', 'type': '{str}'}, 'type': {'key': 'type', 'type': 'str'}, 'pipeline': {'key': 'typeProperties.pipeline', 'type': 'PipelineReference'}, 'parameters': {'key': 'typeProperties.parameters', 'type': '{object}'}, 'wait_on_completion': {'key': 'typeProperties.waitOnCompletion', 'type': 'bool'}, } - def __init__(self, *, name: str, pipeline, additional_properties=None, description: str=None, depends_on=None, parameters=None, wait_on_completion: bool=None, **kwargs) -> None: - super(ExecutePipelineActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, **kwargs) + def __init__(self, *, name: str, pipeline, additional_properties=None, description: str=None, depends_on=None, user_properties=None, parameters=None, wait_on_completion: bool=None, **kwargs) -> None: + super(ExecutePipelineActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, **kwargs) self.pipeline = pipeline self.parameters = parameters self.wait_on_completion = wait_on_completion diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_ssis_package_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_ssis_package_activity.py index f835b78bec55..66282d877881 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_ssis_package_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_ssis_package_activity.py @@ -26,6 +26,8 @@ class ExecuteSSISPackageActivity(ExecutionActivity): :type description: str :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param user_properties: Activity user properties. + :type user_properties: dict[str, str] :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. @@ -79,6 +81,7 @@ class ExecuteSSISPackageActivity(ExecutionActivity): 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'user_properties': {'key': 'userProperties', 'type': '{str}'}, 'type': {'key': 'type', 'type': 'str'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_ssis_package_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_ssis_package_activity_py3.py index 6b58bf34b57b..a1437141b785 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_ssis_package_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_ssis_package_activity_py3.py @@ -26,6 +26,8 @@ class ExecuteSSISPackageActivity(ExecutionActivity): :type description: str :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param user_properties: Activity user properties. + :type user_properties: dict[str, str] :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. @@ -79,6 +81,7 @@ class ExecuteSSISPackageActivity(ExecutionActivity): 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'user_properties': {'key': 'userProperties', 'type': '{str}'}, 'type': {'key': 'type', 'type': 'str'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, @@ -94,8 +97,8 @@ class ExecuteSSISPackageActivity(ExecutionActivity): 'property_overrides': {'key': 'typeProperties.propertyOverrides', 'type': '{SSISPropertyOverride}'}, } - def __init__(self, *, name: str, package_location, connect_via, additional_properties=None, description: str=None, depends_on=None, linked_service_name=None, policy=None, runtime=None, logging_level: str=None, environment_path: str=None, project_parameters=None, package_parameters=None, project_connection_managers=None, package_connection_managers=None, property_overrides=None, **kwargs) -> None: - super(ExecuteSSISPackageActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, linked_service_name=linked_service_name, policy=policy, **kwargs) + def __init__(self, *, name: str, package_location, connect_via, additional_properties=None, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, runtime=None, logging_level: str=None, environment_path: str=None, project_parameters=None, package_parameters=None, project_connection_managers=None, package_connection_managers=None, property_overrides=None, **kwargs) -> None: + super(ExecuteSSISPackageActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) self.package_location = package_location self.runtime = runtime self.logging_level = logging_level diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execution_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execution_activity.py index 736ece86c29c..8e88fec0e8bd 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execution_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execution_activity.py @@ -35,6 +35,8 @@ class ExecutionActivity(Activity): :type description: str :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param user_properties: Activity user properties. + :type user_properties: dict[str, str] :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. @@ -54,6 +56,7 @@ class ExecutionActivity(Activity): 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'user_properties': {'key': 'userProperties', 'type': '{str}'}, 'type': {'key': 'type', 'type': 'str'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execution_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execution_activity_py3.py index 0617ad31da47..d7290f78706a 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execution_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execution_activity_py3.py @@ -35,6 +35,8 @@ class ExecutionActivity(Activity): :type description: str :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param user_properties: Activity user properties. + :type user_properties: dict[str, str] :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. @@ -54,6 +56,7 @@ class ExecutionActivity(Activity): 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'user_properties': {'key': 'userProperties', 'type': '{str}'}, 'type': {'key': 'type', 'type': 'str'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, @@ -63,8 +66,8 @@ class ExecutionActivity(Activity): 'type': {'DatabricksNotebook': 'DatabricksNotebookActivity', 'DataLakeAnalyticsU-SQL': 'DataLakeAnalyticsUSQLActivity', 'AzureMLUpdateResource': 'AzureMLUpdateResourceActivity', 'AzureMLBatchExecution': 'AzureMLBatchExecutionActivity', 'GetMetadata': 'GetMetadataActivity', 'WebActivity': 'WebActivity', 'Lookup': 'LookupActivity', 'SqlServerStoredProcedure': 'SqlServerStoredProcedureActivity', 'Custom': 'CustomActivity', 'ExecuteSSISPackage': 'ExecuteSSISPackageActivity', 'HDInsightSpark': 'HDInsightSparkActivity', 'HDInsightStreaming': 'HDInsightStreamingActivity', 'HDInsightMapReduce': 'HDInsightMapReduceActivity', 'HDInsightPig': 'HDInsightPigActivity', 'HDInsightHive': 'HDInsightHiveActivity', 'Copy': 'CopyActivity'} } - def __init__(self, *, name: str, additional_properties=None, description: str=None, depends_on=None, linked_service_name=None, policy=None, **kwargs) -> None: - super(ExecutionActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, **kwargs) + def __init__(self, *, name: str, additional_properties=None, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, **kwargs) -> None: + super(ExecutionActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, **kwargs) self.linked_service_name = linked_service_name self.policy = policy self.type = 'Execution' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory.py index eaeaa060605e..614b3d7fc97a 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory.py @@ -28,6 +28,8 @@ class Factory(Resource): :type location: str :param tags: The resource tags. :type tags: dict[str, str] + :ivar e_tag: Etag identifies change in the resource. + :vartype e_tag: str :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -39,15 +41,16 @@ class Factory(Resource): :vartype create_time: datetime :ivar version: Version of the factory. :vartype version: str - :param vsts_configuration: VSTS repo information of the factory. - :type vsts_configuration: - ~azure.mgmt.datafactory.models.FactoryVSTSConfiguration + :param repo_configuration: Git repo information of the factory. + :type repo_configuration: + ~azure.mgmt.datafactory.models.FactoryRepoConfiguration """ _validation = { 'id': {'readonly': True}, 'name': {'readonly': True}, 'type': {'readonly': True}, + 'e_tag': {'readonly': True}, 'provisioning_state': {'readonly': True}, 'create_time': {'readonly': True}, 'version': {'readonly': True}, @@ -59,12 +62,13 @@ class Factory(Resource): 'type': {'key': 'type', 'type': 'str'}, 'location': {'key': 'location', 'type': 'str'}, 'tags': {'key': 'tags', 'type': '{str}'}, + 'e_tag': {'key': 'eTag', 'type': 'str'}, 'additional_properties': {'key': '', 'type': '{object}'}, 'identity': {'key': 'identity', 'type': 'FactoryIdentity'}, 'provisioning_state': {'key': 'properties.provisioningState', 'type': 'str'}, 'create_time': {'key': 'properties.createTime', 'type': 'iso-8601'}, 'version': {'key': 'properties.version', 'type': 'str'}, - 'vsts_configuration': {'key': 'properties.vstsConfiguration', 'type': 'FactoryVSTSConfiguration'}, + 'repo_configuration': {'key': 'properties.repoConfiguration', 'type': 'FactoryRepoConfiguration'}, } def __init__(self, **kwargs): @@ -74,4 +78,4 @@ def __init__(self, **kwargs): self.provisioning_state = None self.create_time = None self.version = None - self.vsts_configuration = kwargs.get('vsts_configuration', None) + self.repo_configuration = kwargs.get('repo_configuration', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_git_hub_configuration.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_git_hub_configuration.py new file mode 100644 index 000000000000..2c3f9b14840e --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_git_hub_configuration.py @@ -0,0 +1,61 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .factory_repo_configuration import FactoryRepoConfiguration + + +class FactoryGitHubConfiguration(FactoryRepoConfiguration): + """Factory's GitHub repo information. + + Variables are only populated by the server, and will be ignored when + sending a request. + + All required parameters must be populated in order to send to Azure. + + :param account_name: Required. Account name. + :type account_name: str + :param repository_name: Required. Rrepository name. + :type repository_name: str + :param collaboration_branch: Required. Collaboration branch. + :type collaboration_branch: str + :param root_folder: Required. Root folder. + :type root_folder: str + :param last_commit_id: Last commit id. + :type last_commit_id: str + :param type: Required. Constant filled by server. + :type type: str + :ivar host_name: GitHub repo host name. + :vartype host_name: str + """ + + _validation = { + 'account_name': {'required': True}, + 'repository_name': {'required': True}, + 'collaboration_branch': {'required': True}, + 'root_folder': {'required': True}, + 'type': {'required': True}, + 'host_name': {'readonly': True}, + } + + _attribute_map = { + 'account_name': {'key': 'accountName', 'type': 'str'}, + 'repository_name': {'key': 'repositoryName', 'type': 'str'}, + 'collaboration_branch': {'key': 'collaborationBranch', 'type': 'str'}, + 'root_folder': {'key': 'rootFolder', 'type': 'str'}, + 'last_commit_id': {'key': 'lastCommitId', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'str'}, + 'host_name': {'key': 'hostName', 'type': 'str'}, + } + + def __init__(self, **kwargs): + super(FactoryGitHubConfiguration, self).__init__(**kwargs) + self.host_name = None + self.type = 'FactoryGithubConfiguration' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_git_hub_configuration_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_git_hub_configuration_py3.py new file mode 100644 index 000000000000..1a71ff9b4476 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_git_hub_configuration_py3.py @@ -0,0 +1,61 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .factory_repo_configuration_py3 import FactoryRepoConfiguration + + +class FactoryGitHubConfiguration(FactoryRepoConfiguration): + """Factory's GitHub repo information. + + Variables are only populated by the server, and will be ignored when + sending a request. + + All required parameters must be populated in order to send to Azure. + + :param account_name: Required. Account name. + :type account_name: str + :param repository_name: Required. Rrepository name. + :type repository_name: str + :param collaboration_branch: Required. Collaboration branch. + :type collaboration_branch: str + :param root_folder: Required. Root folder. + :type root_folder: str + :param last_commit_id: Last commit id. + :type last_commit_id: str + :param type: Required. Constant filled by server. + :type type: str + :ivar host_name: GitHub repo host name. + :vartype host_name: str + """ + + _validation = { + 'account_name': {'required': True}, + 'repository_name': {'required': True}, + 'collaboration_branch': {'required': True}, + 'root_folder': {'required': True}, + 'type': {'required': True}, + 'host_name': {'readonly': True}, + } + + _attribute_map = { + 'account_name': {'key': 'accountName', 'type': 'str'}, + 'repository_name': {'key': 'repositoryName', 'type': 'str'}, + 'collaboration_branch': {'key': 'collaborationBranch', 'type': 'str'}, + 'root_folder': {'key': 'rootFolder', 'type': 'str'}, + 'last_commit_id': {'key': 'lastCommitId', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'str'}, + 'host_name': {'key': 'hostName', 'type': 'str'}, + } + + def __init__(self, *, account_name: str, repository_name: str, collaboration_branch: str, root_folder: str, last_commit_id: str=None, **kwargs) -> None: + super(FactoryGitHubConfiguration, self).__init__(account_name=account_name, repository_name=repository_name, collaboration_branch=collaboration_branch, root_folder=root_folder, last_commit_id=last_commit_id, **kwargs) + self.host_name = None + self.type = 'FactoryGithubConfiguration' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_py3.py index b718282c585a..0682aa5f8852 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_py3.py @@ -28,6 +28,8 @@ class Factory(Resource): :type location: str :param tags: The resource tags. :type tags: dict[str, str] + :ivar e_tag: Etag identifies change in the resource. + :vartype e_tag: str :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] @@ -39,15 +41,16 @@ class Factory(Resource): :vartype create_time: datetime :ivar version: Version of the factory. :vartype version: str - :param vsts_configuration: VSTS repo information of the factory. - :type vsts_configuration: - ~azure.mgmt.datafactory.models.FactoryVSTSConfiguration + :param repo_configuration: Git repo information of the factory. + :type repo_configuration: + ~azure.mgmt.datafactory.models.FactoryRepoConfiguration """ _validation = { 'id': {'readonly': True}, 'name': {'readonly': True}, 'type': {'readonly': True}, + 'e_tag': {'readonly': True}, 'provisioning_state': {'readonly': True}, 'create_time': {'readonly': True}, 'version': {'readonly': True}, @@ -59,19 +62,20 @@ class Factory(Resource): 'type': {'key': 'type', 'type': 'str'}, 'location': {'key': 'location', 'type': 'str'}, 'tags': {'key': 'tags', 'type': '{str}'}, + 'e_tag': {'key': 'eTag', 'type': 'str'}, 'additional_properties': {'key': '', 'type': '{object}'}, 'identity': {'key': 'identity', 'type': 'FactoryIdentity'}, 'provisioning_state': {'key': 'properties.provisioningState', 'type': 'str'}, 'create_time': {'key': 'properties.createTime', 'type': 'iso-8601'}, 'version': {'key': 'properties.version', 'type': 'str'}, - 'vsts_configuration': {'key': 'properties.vstsConfiguration', 'type': 'FactoryVSTSConfiguration'}, + 'repo_configuration': {'key': 'properties.repoConfiguration', 'type': 'FactoryRepoConfiguration'}, } - def __init__(self, *, location: str=None, tags=None, additional_properties=None, identity=None, vsts_configuration=None, **kwargs) -> None: + def __init__(self, *, location: str=None, tags=None, additional_properties=None, identity=None, repo_configuration=None, **kwargs) -> None: super(Factory, self).__init__(location=location, tags=tags, **kwargs) self.additional_properties = additional_properties self.identity = identity self.provisioning_state = None self.create_time = None self.version = None - self.vsts_configuration = vsts_configuration + self.repo_configuration = repo_configuration diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_repo_configuration.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_repo_configuration.py new file mode 100644 index 000000000000..2c7175f3a58c --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_repo_configuration.py @@ -0,0 +1,65 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class FactoryRepoConfiguration(Model): + """Factory's git repo information. + + You probably want to use the sub-classes and not this class directly. Known + sub-classes are: FactoryVSTSConfiguration, FactoryGitHubConfiguration + + All required parameters must be populated in order to send to Azure. + + :param account_name: Required. Account name. + :type account_name: str + :param repository_name: Required. Rrepository name. + :type repository_name: str + :param collaboration_branch: Required. Collaboration branch. + :type collaboration_branch: str + :param root_folder: Required. Root folder. + :type root_folder: str + :param last_commit_id: Last commit id. + :type last_commit_id: str + :param type: Required. Constant filled by server. + :type type: str + """ + + _validation = { + 'account_name': {'required': True}, + 'repository_name': {'required': True}, + 'collaboration_branch': {'required': True}, + 'root_folder': {'required': True}, + 'type': {'required': True}, + } + + _attribute_map = { + 'account_name': {'key': 'accountName', 'type': 'str'}, + 'repository_name': {'key': 'repositoryName', 'type': 'str'}, + 'collaboration_branch': {'key': 'collaborationBranch', 'type': 'str'}, + 'root_folder': {'key': 'rootFolder', 'type': 'str'}, + 'last_commit_id': {'key': 'lastCommitId', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + _subtype_map = { + 'type': {'FactoryVSTSConfiguration': 'FactoryVSTSConfiguration', 'FactoryGithubConfiguration': 'FactoryGitHubConfiguration'} + } + + def __init__(self, **kwargs): + super(FactoryRepoConfiguration, self).__init__(**kwargs) + self.account_name = kwargs.get('account_name', None) + self.repository_name = kwargs.get('repository_name', None) + self.collaboration_branch = kwargs.get('collaboration_branch', None) + self.root_folder = kwargs.get('root_folder', None) + self.last_commit_id = kwargs.get('last_commit_id', None) + self.type = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_repo_configuration_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_repo_configuration_py3.py new file mode 100644 index 000000000000..c974f2d5e7da --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_repo_configuration_py3.py @@ -0,0 +1,65 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class FactoryRepoConfiguration(Model): + """Factory's git repo information. + + You probably want to use the sub-classes and not this class directly. Known + sub-classes are: FactoryVSTSConfiguration, FactoryGitHubConfiguration + + All required parameters must be populated in order to send to Azure. + + :param account_name: Required. Account name. + :type account_name: str + :param repository_name: Required. Rrepository name. + :type repository_name: str + :param collaboration_branch: Required. Collaboration branch. + :type collaboration_branch: str + :param root_folder: Required. Root folder. + :type root_folder: str + :param last_commit_id: Last commit id. + :type last_commit_id: str + :param type: Required. Constant filled by server. + :type type: str + """ + + _validation = { + 'account_name': {'required': True}, + 'repository_name': {'required': True}, + 'collaboration_branch': {'required': True}, + 'root_folder': {'required': True}, + 'type': {'required': True}, + } + + _attribute_map = { + 'account_name': {'key': 'accountName', 'type': 'str'}, + 'repository_name': {'key': 'repositoryName', 'type': 'str'}, + 'collaboration_branch': {'key': 'collaborationBranch', 'type': 'str'}, + 'root_folder': {'key': 'rootFolder', 'type': 'str'}, + 'last_commit_id': {'key': 'lastCommitId', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + _subtype_map = { + 'type': {'FactoryVSTSConfiguration': 'FactoryVSTSConfiguration', 'FactoryGithubConfiguration': 'FactoryGitHubConfiguration'} + } + + def __init__(self, *, account_name: str, repository_name: str, collaboration_branch: str, root_folder: str, last_commit_id: str=None, **kwargs) -> None: + super(FactoryRepoConfiguration, self).__init__(**kwargs) + self.account_name = account_name + self.repository_name = repository_name + self.collaboration_branch = collaboration_branch + self.root_folder = root_folder + self.last_commit_id = last_commit_id + self.type = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_repo_update.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_repo_update.py index 7651442cd45c..44eac9d287ce 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_repo_update.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_repo_update.py @@ -13,25 +13,21 @@ class FactoryRepoUpdate(Model): - """Factory's VSTS repo information. + """Factory's git repo information. :param factory_resource_id: The factory resource id. :type factory_resource_id: str - :param resource_group_name: The resource group name. - :type resource_group_name: str - :param vsts_configuration: VSTS repo information of the factory. - :type vsts_configuration: - ~azure.mgmt.datafactory.models.FactoryVSTSConfiguration + :param repo_configuration: Git repo information of the factory. + :type repo_configuration: + ~azure.mgmt.datafactory.models.FactoryRepoConfiguration """ _attribute_map = { 'factory_resource_id': {'key': 'factoryResourceId', 'type': 'str'}, - 'resource_group_name': {'key': 'resourceGroupName', 'type': 'str'}, - 'vsts_configuration': {'key': 'vstsConfiguration', 'type': 'FactoryVSTSConfiguration'}, + 'repo_configuration': {'key': 'repoConfiguration', 'type': 'FactoryRepoConfiguration'}, } def __init__(self, **kwargs): super(FactoryRepoUpdate, self).__init__(**kwargs) self.factory_resource_id = kwargs.get('factory_resource_id', None) - self.resource_group_name = kwargs.get('resource_group_name', None) - self.vsts_configuration = kwargs.get('vsts_configuration', None) + self.repo_configuration = kwargs.get('repo_configuration', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_repo_update_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_repo_update_py3.py index 160f3f80ea28..68aca7a48db8 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_repo_update_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_repo_update_py3.py @@ -13,25 +13,21 @@ class FactoryRepoUpdate(Model): - """Factory's VSTS repo information. + """Factory's git repo information. :param factory_resource_id: The factory resource id. :type factory_resource_id: str - :param resource_group_name: The resource group name. - :type resource_group_name: str - :param vsts_configuration: VSTS repo information of the factory. - :type vsts_configuration: - ~azure.mgmt.datafactory.models.FactoryVSTSConfiguration + :param repo_configuration: Git repo information of the factory. + :type repo_configuration: + ~azure.mgmt.datafactory.models.FactoryRepoConfiguration """ _attribute_map = { 'factory_resource_id': {'key': 'factoryResourceId', 'type': 'str'}, - 'resource_group_name': {'key': 'resourceGroupName', 'type': 'str'}, - 'vsts_configuration': {'key': 'vstsConfiguration', 'type': 'FactoryVSTSConfiguration'}, + 'repo_configuration': {'key': 'repoConfiguration', 'type': 'FactoryRepoConfiguration'}, } - def __init__(self, *, factory_resource_id: str=None, resource_group_name: str=None, vsts_configuration=None, **kwargs) -> None: + def __init__(self, *, factory_resource_id: str=None, repo_configuration=None, **kwargs) -> None: super(FactoryRepoUpdate, self).__init__(**kwargs) self.factory_resource_id = factory_resource_id - self.resource_group_name = resource_group_name - self.vsts_configuration = vsts_configuration + self.repo_configuration = repo_configuration diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_vsts_configuration.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_vsts_configuration.py index c2c017197398..2d8d6d34861a 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_vsts_configuration.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_vsts_configuration.py @@ -9,44 +9,54 @@ # regenerated. # -------------------------------------------------------------------------- -from msrest.serialization import Model +from .factory_repo_configuration import FactoryRepoConfiguration -class FactoryVSTSConfiguration(Model): +class FactoryVSTSConfiguration(FactoryRepoConfiguration): """Factory's VSTS repo information. - :param account_name: VSTS account name. + All required parameters must be populated in order to send to Azure. + + :param account_name: Required. Account name. :type account_name: str - :param project_name: VSTS project name. - :type project_name: str - :param repository_name: VSTS repository name. + :param repository_name: Required. Rrepository name. :type repository_name: str - :param collaboration_branch: VSTS collaboration branch. + :param collaboration_branch: Required. Collaboration branch. :type collaboration_branch: str - :param root_folder: VSTS root folder. + :param root_folder: Required. Root folder. :type root_folder: str - :param last_commit_id: VSTS last commit id. + :param last_commit_id: Last commit id. :type last_commit_id: str + :param type: Required. Constant filled by server. + :type type: str + :param project_name: Required. VSTS project name. + :type project_name: str :param tenant_id: VSTS tenant id. :type tenant_id: str """ + _validation = { + 'account_name': {'required': True}, + 'repository_name': {'required': True}, + 'collaboration_branch': {'required': True}, + 'root_folder': {'required': True}, + 'type': {'required': True}, + 'project_name': {'required': True}, + } + _attribute_map = { 'account_name': {'key': 'accountName', 'type': 'str'}, - 'project_name': {'key': 'projectName', 'type': 'str'}, 'repository_name': {'key': 'repositoryName', 'type': 'str'}, 'collaboration_branch': {'key': 'collaborationBranch', 'type': 'str'}, 'root_folder': {'key': 'rootFolder', 'type': 'str'}, 'last_commit_id': {'key': 'lastCommitId', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'str'}, + 'project_name': {'key': 'projectName', 'type': 'str'}, 'tenant_id': {'key': 'tenantId', 'type': 'str'}, } def __init__(self, **kwargs): super(FactoryVSTSConfiguration, self).__init__(**kwargs) - self.account_name = kwargs.get('account_name', None) self.project_name = kwargs.get('project_name', None) - self.repository_name = kwargs.get('repository_name', None) - self.collaboration_branch = kwargs.get('collaboration_branch', None) - self.root_folder = kwargs.get('root_folder', None) - self.last_commit_id = kwargs.get('last_commit_id', None) self.tenant_id = kwargs.get('tenant_id', None) + self.type = 'FactoryVSTSConfiguration' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_vsts_configuration_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_vsts_configuration_py3.py index 49d6fe8b7a29..afa78e1b1f24 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_vsts_configuration_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_vsts_configuration_py3.py @@ -9,44 +9,54 @@ # regenerated. # -------------------------------------------------------------------------- -from msrest.serialization import Model +from .factory_repo_configuration_py3 import FactoryRepoConfiguration -class FactoryVSTSConfiguration(Model): +class FactoryVSTSConfiguration(FactoryRepoConfiguration): """Factory's VSTS repo information. - :param account_name: VSTS account name. + All required parameters must be populated in order to send to Azure. + + :param account_name: Required. Account name. :type account_name: str - :param project_name: VSTS project name. - :type project_name: str - :param repository_name: VSTS repository name. + :param repository_name: Required. Rrepository name. :type repository_name: str - :param collaboration_branch: VSTS collaboration branch. + :param collaboration_branch: Required. Collaboration branch. :type collaboration_branch: str - :param root_folder: VSTS root folder. + :param root_folder: Required. Root folder. :type root_folder: str - :param last_commit_id: VSTS last commit id. + :param last_commit_id: Last commit id. :type last_commit_id: str + :param type: Required. Constant filled by server. + :type type: str + :param project_name: Required. VSTS project name. + :type project_name: str :param tenant_id: VSTS tenant id. :type tenant_id: str """ + _validation = { + 'account_name': {'required': True}, + 'repository_name': {'required': True}, + 'collaboration_branch': {'required': True}, + 'root_folder': {'required': True}, + 'type': {'required': True}, + 'project_name': {'required': True}, + } + _attribute_map = { 'account_name': {'key': 'accountName', 'type': 'str'}, - 'project_name': {'key': 'projectName', 'type': 'str'}, 'repository_name': {'key': 'repositoryName', 'type': 'str'}, 'collaboration_branch': {'key': 'collaborationBranch', 'type': 'str'}, 'root_folder': {'key': 'rootFolder', 'type': 'str'}, 'last_commit_id': {'key': 'lastCommitId', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'str'}, + 'project_name': {'key': 'projectName', 'type': 'str'}, 'tenant_id': {'key': 'tenantId', 'type': 'str'}, } - def __init__(self, *, account_name: str=None, project_name: str=None, repository_name: str=None, collaboration_branch: str=None, root_folder: str=None, last_commit_id: str=None, tenant_id: str=None, **kwargs) -> None: - super(FactoryVSTSConfiguration, self).__init__(**kwargs) - self.account_name = account_name + def __init__(self, *, account_name: str, repository_name: str, collaboration_branch: str, root_folder: str, project_name: str, last_commit_id: str=None, tenant_id: str=None, **kwargs) -> None: + super(FactoryVSTSConfiguration, self).__init__(account_name=account_name, repository_name=repository_name, collaboration_branch=collaboration_branch, root_folder=root_folder, last_commit_id=last_commit_id, **kwargs) self.project_name = project_name - self.repository_name = repository_name - self.collaboration_branch = collaboration_branch - self.root_folder = root_folder - self.last_commit_id = last_commit_id self.tenant_id = tenant_id + self.type = 'FactoryVSTSConfiguration' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/filter_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/filter_activity.py index 409eff874d6d..29a386fdf0eb 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/filter_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/filter_activity.py @@ -26,6 +26,8 @@ class FilterActivity(ControlActivity): :type description: str :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param user_properties: Activity user properties. + :type user_properties: dict[str, str] :param type: Required. Constant filled by server. :type type: str :param items: Required. Input array on which filter should be applied. @@ -46,6 +48,7 @@ class FilterActivity(ControlActivity): 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'user_properties': {'key': 'userProperties', 'type': '{str}'}, 'type': {'key': 'type', 'type': 'str'}, 'items': {'key': 'typeProperties.items', 'type': 'Expression'}, 'condition': {'key': 'typeProperties.condition', 'type': 'Expression'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/filter_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/filter_activity_py3.py index acc4617a3f72..970767da8b32 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/filter_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/filter_activity_py3.py @@ -26,6 +26,8 @@ class FilterActivity(ControlActivity): :type description: str :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param user_properties: Activity user properties. + :type user_properties: dict[str, str] :param type: Required. Constant filled by server. :type type: str :param items: Required. Input array on which filter should be applied. @@ -46,13 +48,14 @@ class FilterActivity(ControlActivity): 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'user_properties': {'key': 'userProperties', 'type': '{str}'}, 'type': {'key': 'type', 'type': 'str'}, 'items': {'key': 'typeProperties.items', 'type': 'Expression'}, 'condition': {'key': 'typeProperties.condition', 'type': 'Expression'}, } - def __init__(self, *, name: str, items, condition, additional_properties=None, description: str=None, depends_on=None, **kwargs) -> None: - super(FilterActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, **kwargs) + def __init__(self, *, name: str, items, condition, additional_properties=None, description: str=None, depends_on=None, user_properties=None, **kwargs) -> None: + super(FilterActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, **kwargs) self.items = items self.condition = condition self.type = 'Filter' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/for_each_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/for_each_activity.py index f170cfa054ca..e5c37d53a17e 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/for_each_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/for_each_activity.py @@ -27,6 +27,8 @@ class ForEachActivity(ControlActivity): :type description: str :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param user_properties: Activity user properties. + :type user_properties: dict[str, str] :param type: Required. Constant filled by server. :type type: str :param is_sequential: Should the loop be executed in sequence or in @@ -54,6 +56,7 @@ class ForEachActivity(ControlActivity): 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'user_properties': {'key': 'userProperties', 'type': '{str}'}, 'type': {'key': 'type', 'type': 'str'}, 'is_sequential': {'key': 'typeProperties.isSequential', 'type': 'bool'}, 'batch_count': {'key': 'typeProperties.batchCount', 'type': 'int'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/for_each_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/for_each_activity_py3.py index e1ebd7ff426b..f33af7b8e254 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/for_each_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/for_each_activity_py3.py @@ -27,6 +27,8 @@ class ForEachActivity(ControlActivity): :type description: str :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param user_properties: Activity user properties. + :type user_properties: dict[str, str] :param type: Required. Constant filled by server. :type type: str :param is_sequential: Should the loop be executed in sequence or in @@ -54,6 +56,7 @@ class ForEachActivity(ControlActivity): 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'user_properties': {'key': 'userProperties', 'type': '{str}'}, 'type': {'key': 'type', 'type': 'str'}, 'is_sequential': {'key': 'typeProperties.isSequential', 'type': 'bool'}, 'batch_count': {'key': 'typeProperties.batchCount', 'type': 'int'}, @@ -61,8 +64,8 @@ class ForEachActivity(ControlActivity): 'activities': {'key': 'typeProperties.activities', 'type': '[Activity]'}, } - def __init__(self, *, name: str, items, activities, additional_properties=None, description: str=None, depends_on=None, is_sequential: bool=None, batch_count: int=None, **kwargs) -> None: - super(ForEachActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, **kwargs) + def __init__(self, *, name: str, items, activities, additional_properties=None, description: str=None, depends_on=None, user_properties=None, is_sequential: bool=None, batch_count: int=None, **kwargs) -> None: + super(ForEachActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, **kwargs) self.is_sequential = is_sequential self.batch_count = batch_count self.items = items diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/get_metadata_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/get_metadata_activity.py index c1bac6e451f1..2d5d2c99447e 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/get_metadata_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/get_metadata_activity.py @@ -26,6 +26,8 @@ class GetMetadataActivity(ExecutionActivity): :type description: str :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param user_properties: Activity user properties. + :type user_properties: dict[str, str] :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. @@ -50,6 +52,7 @@ class GetMetadataActivity(ExecutionActivity): 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'user_properties': {'key': 'userProperties', 'type': '{str}'}, 'type': {'key': 'type', 'type': 'str'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/get_metadata_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/get_metadata_activity_py3.py index 79e16f1a8c14..b992f462b6f3 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/get_metadata_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/get_metadata_activity_py3.py @@ -26,6 +26,8 @@ class GetMetadataActivity(ExecutionActivity): :type description: str :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param user_properties: Activity user properties. + :type user_properties: dict[str, str] :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. @@ -50,6 +52,7 @@ class GetMetadataActivity(ExecutionActivity): 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'user_properties': {'key': 'userProperties', 'type': '{str}'}, 'type': {'key': 'type', 'type': 'str'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, @@ -57,8 +60,8 @@ class GetMetadataActivity(ExecutionActivity): 'field_list': {'key': 'typeProperties.fieldList', 'type': '[object]'}, } - def __init__(self, *, name: str, dataset, additional_properties=None, description: str=None, depends_on=None, linked_service_name=None, policy=None, field_list=None, **kwargs) -> None: - super(GetMetadataActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, linked_service_name=linked_service_name, policy=policy, **kwargs) + def __init__(self, *, name: str, dataset, additional_properties=None, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, field_list=None, **kwargs) -> None: + super(GetMetadataActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) self.dataset = dataset self.field_list = field_list self.type = 'GetMetadata' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_hive_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_hive_activity.py index 21b78dbc4d24..a856c100b8b0 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_hive_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_hive_activity.py @@ -26,6 +26,8 @@ class HDInsightHiveActivity(ExecutionActivity): :type description: str :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param user_properties: Activity user properties. + :type user_properties: dict[str, str] :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. @@ -62,6 +64,7 @@ class HDInsightHiveActivity(ExecutionActivity): 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'user_properties': {'key': 'userProperties', 'type': '{str}'}, 'type': {'key': 'type', 'type': 'str'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_hive_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_hive_activity_py3.py index 0d3e900a4398..3ced50d708ac 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_hive_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_hive_activity_py3.py @@ -26,6 +26,8 @@ class HDInsightHiveActivity(ExecutionActivity): :type description: str :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param user_properties: Activity user properties. + :type user_properties: dict[str, str] :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. @@ -62,6 +64,7 @@ class HDInsightHiveActivity(ExecutionActivity): 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'user_properties': {'key': 'userProperties', 'type': '{str}'}, 'type': {'key': 'type', 'type': 'str'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, @@ -73,8 +76,8 @@ class HDInsightHiveActivity(ExecutionActivity): 'defines': {'key': 'typeProperties.defines', 'type': '{object}'}, } - def __init__(self, *, name: str, additional_properties=None, description: str=None, depends_on=None, linked_service_name=None, policy=None, storage_linked_services=None, arguments=None, get_debug_info=None, script_path=None, script_linked_service=None, defines=None, **kwargs) -> None: - super(HDInsightHiveActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, linked_service_name=linked_service_name, policy=policy, **kwargs) + def __init__(self, *, name: str, additional_properties=None, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, storage_linked_services=None, arguments=None, get_debug_info=None, script_path=None, script_linked_service=None, defines=None, **kwargs) -> None: + super(HDInsightHiveActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) self.storage_linked_services = storage_linked_services self.arguments = arguments self.get_debug_info = get_debug_info diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_map_reduce_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_map_reduce_activity.py index 2f7661fd79de..293013d73afc 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_map_reduce_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_map_reduce_activity.py @@ -26,6 +26,8 @@ class HDInsightMapReduceActivity(ExecutionActivity): :type description: str :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param user_properties: Activity user properties. + :type user_properties: dict[str, str] :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. @@ -70,6 +72,7 @@ class HDInsightMapReduceActivity(ExecutionActivity): 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'user_properties': {'key': 'userProperties', 'type': '{str}'}, 'type': {'key': 'type', 'type': 'str'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_map_reduce_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_map_reduce_activity_py3.py index 2ac97c2e8453..830df1d395c2 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_map_reduce_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_map_reduce_activity_py3.py @@ -26,6 +26,8 @@ class HDInsightMapReduceActivity(ExecutionActivity): :type description: str :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param user_properties: Activity user properties. + :type user_properties: dict[str, str] :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. @@ -70,6 +72,7 @@ class HDInsightMapReduceActivity(ExecutionActivity): 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'user_properties': {'key': 'userProperties', 'type': '{str}'}, 'type': {'key': 'type', 'type': 'str'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, @@ -83,8 +86,8 @@ class HDInsightMapReduceActivity(ExecutionActivity): 'defines': {'key': 'typeProperties.defines', 'type': '{object}'}, } - def __init__(self, *, name: str, class_name, jar_file_path, additional_properties=None, description: str=None, depends_on=None, linked_service_name=None, policy=None, storage_linked_services=None, arguments=None, get_debug_info=None, jar_linked_service=None, jar_libs=None, defines=None, **kwargs) -> None: - super(HDInsightMapReduceActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, linked_service_name=linked_service_name, policy=policy, **kwargs) + def __init__(self, *, name: str, class_name, jar_file_path, additional_properties=None, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, storage_linked_services=None, arguments=None, get_debug_info=None, jar_linked_service=None, jar_libs=None, defines=None, **kwargs) -> None: + super(HDInsightMapReduceActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) self.storage_linked_services = storage_linked_services self.arguments = arguments self.get_debug_info = get_debug_info diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_pig_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_pig_activity.py index a552b1d9ae52..9400060e9d12 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_pig_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_pig_activity.py @@ -26,6 +26,8 @@ class HDInsightPigActivity(ExecutionActivity): :type description: str :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param user_properties: Activity user properties. + :type user_properties: dict[str, str] :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. @@ -62,6 +64,7 @@ class HDInsightPigActivity(ExecutionActivity): 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'user_properties': {'key': 'userProperties', 'type': '{str}'}, 'type': {'key': 'type', 'type': 'str'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_pig_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_pig_activity_py3.py index b57d8da7b682..839d193c5059 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_pig_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_pig_activity_py3.py @@ -26,6 +26,8 @@ class HDInsightPigActivity(ExecutionActivity): :type description: str :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param user_properties: Activity user properties. + :type user_properties: dict[str, str] :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. @@ -62,6 +64,7 @@ class HDInsightPigActivity(ExecutionActivity): 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'user_properties': {'key': 'userProperties', 'type': '{str}'}, 'type': {'key': 'type', 'type': 'str'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, @@ -73,8 +76,8 @@ class HDInsightPigActivity(ExecutionActivity): 'defines': {'key': 'typeProperties.defines', 'type': '{object}'}, } - def __init__(self, *, name: str, additional_properties=None, description: str=None, depends_on=None, linked_service_name=None, policy=None, storage_linked_services=None, arguments=None, get_debug_info=None, script_path=None, script_linked_service=None, defines=None, **kwargs) -> None: - super(HDInsightPigActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, linked_service_name=linked_service_name, policy=policy, **kwargs) + def __init__(self, *, name: str, additional_properties=None, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, storage_linked_services=None, arguments=None, get_debug_info=None, script_path=None, script_linked_service=None, defines=None, **kwargs) -> None: + super(HDInsightPigActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) self.storage_linked_services = storage_linked_services self.arguments = arguments self.get_debug_info = get_debug_info diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_spark_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_spark_activity.py index 4984385f5d26..422177713af4 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_spark_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_spark_activity.py @@ -26,6 +26,8 @@ class HDInsightSparkActivity(ExecutionActivity): :type description: str :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param user_properties: Activity user properties. + :type user_properties: dict[str, str] :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. @@ -71,6 +73,7 @@ class HDInsightSparkActivity(ExecutionActivity): 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'user_properties': {'key': 'userProperties', 'type': '{str}'}, 'type': {'key': 'type', 'type': 'str'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_spark_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_spark_activity_py3.py index 322d99f7c063..b3a44e425368 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_spark_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_spark_activity_py3.py @@ -26,6 +26,8 @@ class HDInsightSparkActivity(ExecutionActivity): :type description: str :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param user_properties: Activity user properties. + :type user_properties: dict[str, str] :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. @@ -71,6 +73,7 @@ class HDInsightSparkActivity(ExecutionActivity): 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'user_properties': {'key': 'userProperties', 'type': '{str}'}, 'type': {'key': 'type', 'type': 'str'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, @@ -84,8 +87,8 @@ class HDInsightSparkActivity(ExecutionActivity): 'spark_config': {'key': 'typeProperties.sparkConfig', 'type': '{object}'}, } - def __init__(self, *, name: str, root_path, entry_file_path, additional_properties=None, description: str=None, depends_on=None, linked_service_name=None, policy=None, arguments=None, get_debug_info=None, spark_job_linked_service=None, class_name: str=None, proxy_user=None, spark_config=None, **kwargs) -> None: - super(HDInsightSparkActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, linked_service_name=linked_service_name, policy=policy, **kwargs) + def __init__(self, *, name: str, root_path, entry_file_path, additional_properties=None, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, arguments=None, get_debug_info=None, spark_job_linked_service=None, class_name: str=None, proxy_user=None, spark_config=None, **kwargs) -> None: + super(HDInsightSparkActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) self.root_path = root_path self.entry_file_path = entry_file_path self.arguments = arguments diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_streaming_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_streaming_activity.py index 80469577315f..fc2bcc3ac1b8 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_streaming_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_streaming_activity.py @@ -26,6 +26,8 @@ class HDInsightStreamingActivity(ExecutionActivity): :type description: str :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param user_properties: Activity user properties. + :type user_properties: dict[str, str] :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. @@ -85,6 +87,7 @@ class HDInsightStreamingActivity(ExecutionActivity): 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'user_properties': {'key': 'userProperties', 'type': '{str}'}, 'type': {'key': 'type', 'type': 'str'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_streaming_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_streaming_activity_py3.py index 10756ecd093d..40ae73bf18ec 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_streaming_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_streaming_activity_py3.py @@ -26,6 +26,8 @@ class HDInsightStreamingActivity(ExecutionActivity): :type description: str :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param user_properties: Activity user properties. + :type user_properties: dict[str, str] :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. @@ -85,6 +87,7 @@ class HDInsightStreamingActivity(ExecutionActivity): 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'user_properties': {'key': 'userProperties', 'type': '{str}'}, 'type': {'key': 'type', 'type': 'str'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, @@ -102,8 +105,8 @@ class HDInsightStreamingActivity(ExecutionActivity): 'defines': {'key': 'typeProperties.defines', 'type': '{object}'}, } - def __init__(self, *, name: str, mapper, reducer, input, output, file_paths, additional_properties=None, description: str=None, depends_on=None, linked_service_name=None, policy=None, storage_linked_services=None, arguments=None, get_debug_info=None, file_linked_service=None, combiner=None, command_environment=None, defines=None, **kwargs) -> None: - super(HDInsightStreamingActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, linked_service_name=linked_service_name, policy=policy, **kwargs) + def __init__(self, *, name: str, mapper, reducer, input, output, file_paths, additional_properties=None, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, storage_linked_services=None, arguments=None, get_debug_info=None, file_linked_service=None, combiner=None, command_environment=None, defines=None, **kwargs) -> None: + super(HDInsightStreamingActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) self.storage_linked_services = storage_linked_services self.arguments = arguments self.get_debug_info = get_debug_info diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/if_condition_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/if_condition_activity.py index 78eae5911550..1e66cc0a875b 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/if_condition_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/if_condition_activity.py @@ -28,6 +28,8 @@ class IfConditionActivity(ControlActivity): :type description: str :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param user_properties: Activity user properties. + :type user_properties: dict[str, str] :param type: Required. Constant filled by server. :type type: str :param expression: Required. An expression that would evaluate to Boolean. @@ -55,6 +57,7 @@ class IfConditionActivity(ControlActivity): 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'user_properties': {'key': 'userProperties', 'type': '{str}'}, 'type': {'key': 'type', 'type': 'str'}, 'expression': {'key': 'typeProperties.expression', 'type': 'Expression'}, 'if_true_activities': {'key': 'typeProperties.ifTrueActivities', 'type': '[Activity]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/if_condition_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/if_condition_activity_py3.py index cbef9a900284..00798ac67bff 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/if_condition_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/if_condition_activity_py3.py @@ -28,6 +28,8 @@ class IfConditionActivity(ControlActivity): :type description: str :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param user_properties: Activity user properties. + :type user_properties: dict[str, str] :param type: Required. Constant filled by server. :type type: str :param expression: Required. An expression that would evaluate to Boolean. @@ -55,14 +57,15 @@ class IfConditionActivity(ControlActivity): 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'user_properties': {'key': 'userProperties', 'type': '{str}'}, 'type': {'key': 'type', 'type': 'str'}, 'expression': {'key': 'typeProperties.expression', 'type': 'Expression'}, 'if_true_activities': {'key': 'typeProperties.ifTrueActivities', 'type': '[Activity]'}, 'if_false_activities': {'key': 'typeProperties.ifFalseActivities', 'type': '[Activity]'}, } - def __init__(self, *, name: str, expression, additional_properties=None, description: str=None, depends_on=None, if_true_activities=None, if_false_activities=None, **kwargs) -> None: - super(IfConditionActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, **kwargs) + def __init__(self, *, name: str, expression, additional_properties=None, description: str=None, depends_on=None, user_properties=None, if_true_activities=None, if_false_activities=None, **kwargs) -> None: + super(IfConditionActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, **kwargs) self.expression = expression self.if_true_activities = if_true_activities self.if_false_activities = if_false_activities diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_node_monitoring_data.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_node_monitoring_data.py index 93dc31785912..9d27bedf70aa 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_node_monitoring_data.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_node_monitoring_data.py @@ -27,7 +27,7 @@ class IntegrationRuntimeNodeMonitoringData(Model): runtime node. :vartype available_memory_in_mb: int :ivar cpu_utilization: CPU percentage on the integration runtime node. - :vartype cpu_utilization: float + :vartype cpu_utilization: int :ivar concurrent_jobs_limit: Maximum concurrent jobs on the integration runtime node. :vartype concurrent_jobs_limit: int @@ -58,7 +58,7 @@ class IntegrationRuntimeNodeMonitoringData(Model): 'additional_properties': {'key': '', 'type': '{object}'}, 'node_name': {'key': 'nodeName', 'type': 'str'}, 'available_memory_in_mb': {'key': 'availableMemoryInMB', 'type': 'int'}, - 'cpu_utilization': {'key': 'cpuUtilization', 'type': 'float'}, + 'cpu_utilization': {'key': 'cpuUtilization', 'type': 'int'}, 'concurrent_jobs_limit': {'key': 'concurrentJobsLimit', 'type': 'int'}, 'concurrent_jobs_running': {'key': 'concurrentJobsRunning', 'type': 'int'}, 'max_concurrent_jobs': {'key': 'maxConcurrentJobs', 'type': 'int'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_node_monitoring_data_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_node_monitoring_data_py3.py index 8c6482817e46..35c7e664b2ff 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_node_monitoring_data_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_node_monitoring_data_py3.py @@ -27,7 +27,7 @@ class IntegrationRuntimeNodeMonitoringData(Model): runtime node. :vartype available_memory_in_mb: int :ivar cpu_utilization: CPU percentage on the integration runtime node. - :vartype cpu_utilization: float + :vartype cpu_utilization: int :ivar concurrent_jobs_limit: Maximum concurrent jobs on the integration runtime node. :vartype concurrent_jobs_limit: int @@ -58,7 +58,7 @@ class IntegrationRuntimeNodeMonitoringData(Model): 'additional_properties': {'key': '', 'type': '{object}'}, 'node_name': {'key': 'nodeName', 'type': 'str'}, 'available_memory_in_mb': {'key': 'availableMemoryInMB', 'type': 'int'}, - 'cpu_utilization': {'key': 'cpuUtilization', 'type': 'float'}, + 'cpu_utilization': {'key': 'cpuUtilization', 'type': 'int'}, 'concurrent_jobs_limit': {'key': 'concurrentJobsLimit', 'type': 'int'}, 'concurrent_jobs_running': {'key': 'concurrentJobsRunning', 'type': 'int'}, 'max_concurrent_jobs': {'key': 'maxConcurrentJobs', 'type': 'int'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_remove_node_request.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_remove_node_request.py deleted file mode 100644 index 9080f93ffbdd..000000000000 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_remove_node_request.py +++ /dev/null @@ -1,33 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for -# license information. -# -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is -# regenerated. -# -------------------------------------------------------------------------- - -from msrest.serialization import Model - - -class IntegrationRuntimeRemoveNodeRequest(Model): - """Request to remove a node. - - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] - :param node_name: The name of the node to be removed. - :type node_name: str - """ - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'node_name': {'key': 'nodeName', 'type': 'str'}, - } - - def __init__(self, **kwargs): - super(IntegrationRuntimeRemoveNodeRequest, self).__init__(**kwargs) - self.additional_properties = kwargs.get('additional_properties', None) - self.node_name = kwargs.get('node_name', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_remove_node_request_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_remove_node_request_py3.py deleted file mode 100644 index 57453c633fc5..000000000000 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_remove_node_request_py3.py +++ /dev/null @@ -1,33 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for -# license information. -# -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is -# regenerated. -# -------------------------------------------------------------------------- - -from msrest.serialization import Model - - -class IntegrationRuntimeRemoveNodeRequest(Model): - """Request to remove a node. - - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] - :param node_name: The name of the node to be removed. - :type node_name: str - """ - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'node_name': {'key': 'nodeName', 'type': 'str'}, - } - - def __init__(self, *, additional_properties=None, node_name: str=None, **kwargs) -> None: - super(IntegrationRuntimeRemoveNodeRequest, self).__init__(**kwargs) - self.additional_properties = additional_properties - self.node_name = node_name diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime.py deleted file mode 100644 index f4a4e7eb8bf0..000000000000 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime.py +++ /dev/null @@ -1,58 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for -# license information. -# -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is -# regenerated. -# -------------------------------------------------------------------------- - -from msrest.serialization import Model - - -class LinkedIntegrationRuntime(Model): - """The linked integration runtime information. - - Variables are only populated by the server, and will be ignored when - sending a request. - - :ivar name: The name of the linked integration runtime. - :vartype name: str - :ivar subscription_id: The subscription ID for which the linked - integration runtime belong to. - :vartype subscription_id: str - :ivar data_factory_name: The name of the data factory for which the linked - integration runtime belong to. - :vartype data_factory_name: str - :ivar data_factory_location: The location of the data factory for which - the linked integration runtime belong to. - :vartype data_factory_location: str - :ivar create_time: The creating time of the linked integration runtime. - :vartype create_time: datetime - """ - - _validation = { - 'name': {'readonly': True}, - 'subscription_id': {'readonly': True}, - 'data_factory_name': {'readonly': True}, - 'data_factory_location': {'readonly': True}, - 'create_time': {'readonly': True}, - } - - _attribute_map = { - 'name': {'key': 'name', 'type': 'str'}, - 'subscription_id': {'key': 'subscriptionId', 'type': 'str'}, - 'data_factory_name': {'key': 'dataFactoryName', 'type': 'str'}, - 'data_factory_location': {'key': 'dataFactoryLocation', 'type': 'str'}, - 'create_time': {'key': 'createTime', 'type': 'iso-8601'}, - } - - def __init__(self, **kwargs): - super(LinkedIntegrationRuntime, self).__init__(**kwargs) - self.name = None - self.subscription_id = None - self.data_factory_name = None - self.data_factory_location = None - self.create_time = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_key.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_key.py deleted file mode 100644 index cbecfb9afae4..000000000000 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_key.py +++ /dev/null @@ -1,39 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for -# license information. -# -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is -# regenerated. -# -------------------------------------------------------------------------- - -from .linked_integration_runtime_properties import LinkedIntegrationRuntimeProperties - - -class LinkedIntegrationRuntimeKey(LinkedIntegrationRuntimeProperties): - """The base definition of a secret type. - - All required parameters must be populated in order to send to Azure. - - :param authorization_type: Required. Constant filled by server. - :type authorization_type: str - :param key: Required. Type of the secret. - :type key: ~azure.mgmt.datafactory.models.SecureString - """ - - _validation = { - 'authorization_type': {'required': True}, - 'key': {'required': True}, - } - - _attribute_map = { - 'authorization_type': {'key': 'authorizationType', 'type': 'str'}, - 'key': {'key': 'key', 'type': 'SecureString'}, - } - - def __init__(self, **kwargs): - super(LinkedIntegrationRuntimeKey, self).__init__(**kwargs) - self.key = kwargs.get('key', None) - self.authorization_type = 'Key' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_key_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_key_py3.py deleted file mode 100644 index c223b18281cd..000000000000 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_key_py3.py +++ /dev/null @@ -1,39 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for -# license information. -# -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is -# regenerated. -# -------------------------------------------------------------------------- - -from .linked_integration_runtime_properties_py3 import LinkedIntegrationRuntimeProperties - - -class LinkedIntegrationRuntimeKey(LinkedIntegrationRuntimeProperties): - """The base definition of a secret type. - - All required parameters must be populated in order to send to Azure. - - :param authorization_type: Required. Constant filled by server. - :type authorization_type: str - :param key: Required. Type of the secret. - :type key: ~azure.mgmt.datafactory.models.SecureString - """ - - _validation = { - 'authorization_type': {'required': True}, - 'key': {'required': True}, - } - - _attribute_map = { - 'authorization_type': {'key': 'authorizationType', 'type': 'str'}, - 'key': {'key': 'key', 'type': 'SecureString'}, - } - - def __init__(self, *, key, **kwargs) -> None: - super(LinkedIntegrationRuntimeKey, self).__init__(**kwargs) - self.key = key - self.authorization_type = 'Key' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_properties.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_properties.py deleted file mode 100644 index 20a80b23277c..000000000000 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_properties.py +++ /dev/null @@ -1,41 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for -# license information. -# -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is -# regenerated. -# -------------------------------------------------------------------------- - -from msrest.serialization import Model - - -class LinkedIntegrationRuntimeProperties(Model): - """The base definition of a secret type. - - You probably want to use the sub-classes and not this class directly. Known - sub-classes are: LinkedIntegrationRuntimeRbac, LinkedIntegrationRuntimeKey - - All required parameters must be populated in order to send to Azure. - - :param authorization_type: Required. Constant filled by server. - :type authorization_type: str - """ - - _validation = { - 'authorization_type': {'required': True}, - } - - _attribute_map = { - 'authorization_type': {'key': 'authorizationType', 'type': 'str'}, - } - - _subtype_map = { - 'authorization_type': {'RBAC': 'LinkedIntegrationRuntimeRbac', 'Key': 'LinkedIntegrationRuntimeKey'} - } - - def __init__(self, **kwargs): - super(LinkedIntegrationRuntimeProperties, self).__init__(**kwargs) - self.authorization_type = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_properties_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_properties_py3.py deleted file mode 100644 index da73c9b6b95a..000000000000 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_properties_py3.py +++ /dev/null @@ -1,41 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for -# license information. -# -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is -# regenerated. -# -------------------------------------------------------------------------- - -from msrest.serialization import Model - - -class LinkedIntegrationRuntimeProperties(Model): - """The base definition of a secret type. - - You probably want to use the sub-classes and not this class directly. Known - sub-classes are: LinkedIntegrationRuntimeRbac, LinkedIntegrationRuntimeKey - - All required parameters must be populated in order to send to Azure. - - :param authorization_type: Required. Constant filled by server. - :type authorization_type: str - """ - - _validation = { - 'authorization_type': {'required': True}, - } - - _attribute_map = { - 'authorization_type': {'key': 'authorizationType', 'type': 'str'}, - } - - _subtype_map = { - 'authorization_type': {'RBAC': 'LinkedIntegrationRuntimeRbac', 'Key': 'LinkedIntegrationRuntimeKey'} - } - - def __init__(self, **kwargs) -> None: - super(LinkedIntegrationRuntimeProperties, self).__init__(**kwargs) - self.authorization_type = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_py3.py deleted file mode 100644 index 6c831ab5f511..000000000000 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_py3.py +++ /dev/null @@ -1,58 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for -# license information. -# -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is -# regenerated. -# -------------------------------------------------------------------------- - -from msrest.serialization import Model - - -class LinkedIntegrationRuntime(Model): - """The linked integration runtime information. - - Variables are only populated by the server, and will be ignored when - sending a request. - - :ivar name: The name of the linked integration runtime. - :vartype name: str - :ivar subscription_id: The subscription ID for which the linked - integration runtime belong to. - :vartype subscription_id: str - :ivar data_factory_name: The name of the data factory for which the linked - integration runtime belong to. - :vartype data_factory_name: str - :ivar data_factory_location: The location of the data factory for which - the linked integration runtime belong to. - :vartype data_factory_location: str - :ivar create_time: The creating time of the linked integration runtime. - :vartype create_time: datetime - """ - - _validation = { - 'name': {'readonly': True}, - 'subscription_id': {'readonly': True}, - 'data_factory_name': {'readonly': True}, - 'data_factory_location': {'readonly': True}, - 'create_time': {'readonly': True}, - } - - _attribute_map = { - 'name': {'key': 'name', 'type': 'str'}, - 'subscription_id': {'key': 'subscriptionId', 'type': 'str'}, - 'data_factory_name': {'key': 'dataFactoryName', 'type': 'str'}, - 'data_factory_location': {'key': 'dataFactoryLocation', 'type': 'str'}, - 'create_time': {'key': 'createTime', 'type': 'iso-8601'}, - } - - def __init__(self, **kwargs) -> None: - super(LinkedIntegrationRuntime, self).__init__(**kwargs) - self.name = None - self.subscription_id = None - self.data_factory_name = None - self.data_factory_location = None - self.create_time = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_rbac.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_rbac.py deleted file mode 100644 index 3c7235ecc279..000000000000 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_rbac.py +++ /dev/null @@ -1,40 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for -# license information. -# -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is -# regenerated. -# -------------------------------------------------------------------------- - -from .linked_integration_runtime_properties import LinkedIntegrationRuntimeProperties - - -class LinkedIntegrationRuntimeRbac(LinkedIntegrationRuntimeProperties): - """The base definition of a secret type. - - All required parameters must be populated in order to send to Azure. - - :param authorization_type: Required. Constant filled by server. - :type authorization_type: str - :param resource_id: Required. The resource ID of the integration runtime - to be shared. - :type resource_id: str - """ - - _validation = { - 'authorization_type': {'required': True}, - 'resource_id': {'required': True}, - } - - _attribute_map = { - 'authorization_type': {'key': 'authorizationType', 'type': 'str'}, - 'resource_id': {'key': 'resourceId', 'type': 'str'}, - } - - def __init__(self, **kwargs): - super(LinkedIntegrationRuntimeRbac, self).__init__(**kwargs) - self.resource_id = kwargs.get('resource_id', None) - self.authorization_type = 'RBAC' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_rbac_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_rbac_py3.py deleted file mode 100644 index 24d2b946d97e..000000000000 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_rbac_py3.py +++ /dev/null @@ -1,40 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for -# license information. -# -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is -# regenerated. -# -------------------------------------------------------------------------- - -from .linked_integration_runtime_properties_py3 import LinkedIntegrationRuntimeProperties - - -class LinkedIntegrationRuntimeRbac(LinkedIntegrationRuntimeProperties): - """The base definition of a secret type. - - All required parameters must be populated in order to send to Azure. - - :param authorization_type: Required. Constant filled by server. - :type authorization_type: str - :param resource_id: Required. The resource ID of the integration runtime - to be shared. - :type resource_id: str - """ - - _validation = { - 'authorization_type': {'required': True}, - 'resource_id': {'required': True}, - } - - _attribute_map = { - 'authorization_type': {'key': 'authorizationType', 'type': 'str'}, - 'resource_id': {'key': 'resourceId', 'type': 'str'}, - } - - def __init__(self, *, resource_id: str, **kwargs) -> None: - super(LinkedIntegrationRuntimeRbac, self).__init__(**kwargs) - self.resource_id = resource_id - self.authorization_type = 'RBAC' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/lookup_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/lookup_activity.py index d5bfff330ee2..0cf215272adf 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/lookup_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/lookup_activity.py @@ -26,6 +26,8 @@ class LookupActivity(ExecutionActivity): :type description: str :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param user_properties: Activity user properties. + :type user_properties: dict[str, str] :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. @@ -55,6 +57,7 @@ class LookupActivity(ExecutionActivity): 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'user_properties': {'key': 'userProperties', 'type': '{str}'}, 'type': {'key': 'type', 'type': 'str'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/lookup_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/lookup_activity_py3.py index fdbcebcd0164..ef81802111c0 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/lookup_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/lookup_activity_py3.py @@ -26,6 +26,8 @@ class LookupActivity(ExecutionActivity): :type description: str :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param user_properties: Activity user properties. + :type user_properties: dict[str, str] :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. @@ -55,6 +57,7 @@ class LookupActivity(ExecutionActivity): 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'user_properties': {'key': 'userProperties', 'type': '{str}'}, 'type': {'key': 'type', 'type': 'str'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, @@ -63,8 +66,8 @@ class LookupActivity(ExecutionActivity): 'first_row_only': {'key': 'typeProperties.firstRowOnly', 'type': 'object'}, } - def __init__(self, *, name: str, source, dataset, additional_properties=None, description: str=None, depends_on=None, linked_service_name=None, policy=None, first_row_only=None, **kwargs) -> None: - super(LookupActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, linked_service_name=linked_service_name, policy=policy, **kwargs) + def __init__(self, *, name: str, source, dataset, additional_properties=None, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, first_row_only=None, **kwargs) -> None: + super(LookupActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) self.source = source self.dataset = dataset self.first_row_only = first_row_only diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_list_response.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_list_response.py deleted file mode 100644 index 5b2c7f8cb037..000000000000 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_list_response.py +++ /dev/null @@ -1,34 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for -# license information. -# -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is -# regenerated. -# -------------------------------------------------------------------------- - -from msrest.serialization import Model - - -class OperationListResponse(Model): - """A list of operations that can be performed by the Data Factory service. - - :param value: List of Data Factory operations supported by the Data - Factory resource provider. - :type value: list[~azure.mgmt.datafactory.models.Operation] - :param next_link: The link to the next page of results, if any remaining - results exist. - :type next_link: str - """ - - _attribute_map = { - 'value': {'key': 'value', 'type': '[Operation]'}, - 'next_link': {'key': 'nextLink', 'type': 'str'}, - } - - def __init__(self, **kwargs): - super(OperationListResponse, self).__init__(**kwargs) - self.value = kwargs.get('value', None) - self.next_link = kwargs.get('next_link', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_list_response_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_list_response_py3.py deleted file mode 100644 index fe5c1ce91260..000000000000 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_list_response_py3.py +++ /dev/null @@ -1,34 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for -# license information. -# -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is -# regenerated. -# -------------------------------------------------------------------------- - -from msrest.serialization import Model - - -class OperationListResponse(Model): - """A list of operations that can be performed by the Data Factory service. - - :param value: List of Data Factory operations supported by the Data - Factory resource provider. - :type value: list[~azure.mgmt.datafactory.models.Operation] - :param next_link: The link to the next page of results, if any remaining - results exist. - :type next_link: str - """ - - _attribute_map = { - 'value': {'key': 'value', 'type': '[Operation]'}, - 'next_link': {'key': 'nextLink', 'type': 'str'}, - } - - def __init__(self, *, value=None, next_link: str=None, **kwargs) -> None: - super(OperationListResponse, self).__init__(**kwargs) - self.value = value - self.next_link = next_link diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_metric_dimension.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_metric_dimension.py new file mode 100644 index 000000000000..24232e7b5470 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_metric_dimension.py @@ -0,0 +1,37 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class OperationMetricDimension(Model): + """Defines the metric dimension. + + :param name: The name of the dimension for the metric. + :type name: str + :param display_name: The display name of the metric dimension. + :type display_name: str + :param to_be_exported_for_shoebox: Whether the dimension should be + exported to Azure Monitor. + :type to_be_exported_for_shoebox: bool + """ + + _attribute_map = { + 'name': {'key': 'name', 'type': 'str'}, + 'display_name': {'key': 'displayName', 'type': 'str'}, + 'to_be_exported_for_shoebox': {'key': 'toBeExportedForShoebox', 'type': 'bool'}, + } + + def __init__(self, **kwargs): + super(OperationMetricDimension, self).__init__(**kwargs) + self.name = kwargs.get('name', None) + self.display_name = kwargs.get('display_name', None) + self.to_be_exported_for_shoebox = kwargs.get('to_be_exported_for_shoebox', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_metric_dimension_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_metric_dimension_py3.py new file mode 100644 index 000000000000..1d8610b7fab8 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_metric_dimension_py3.py @@ -0,0 +1,37 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class OperationMetricDimension(Model): + """Defines the metric dimension. + + :param name: The name of the dimension for the metric. + :type name: str + :param display_name: The display name of the metric dimension. + :type display_name: str + :param to_be_exported_for_shoebox: Whether the dimension should be + exported to Azure Monitor. + :type to_be_exported_for_shoebox: bool + """ + + _attribute_map = { + 'name': {'key': 'name', 'type': 'str'}, + 'display_name': {'key': 'displayName', 'type': 'str'}, + 'to_be_exported_for_shoebox': {'key': 'toBeExportedForShoebox', 'type': 'bool'}, + } + + def __init__(self, *, name: str=None, display_name: str=None, to_be_exported_for_shoebox: bool=None, **kwargs) -> None: + super(OperationMetricDimension, self).__init__(**kwargs) + self.name = name + self.display_name = display_name + self.to_be_exported_for_shoebox = to_be_exported_for_shoebox diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_metric_specification.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_metric_specification.py index e2a9c98580a8..77f533fdcebf 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_metric_specification.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_metric_specification.py @@ -36,6 +36,9 @@ class OperationMetricSpecification(Model): available. :type availabilities: list[~azure.mgmt.datafactory.models.OperationMetricAvailability] + :param dimensions: Defines the metric dimension. + :type dimensions: + list[~azure.mgmt.datafactory.models.OperationMetricDimension] """ _attribute_map = { @@ -48,6 +51,7 @@ class OperationMetricSpecification(Model): 'source_mdm_account': {'key': 'sourceMdmAccount', 'type': 'str'}, 'source_mdm_namespace': {'key': 'sourceMdmNamespace', 'type': 'str'}, 'availabilities': {'key': 'availabilities', 'type': '[OperationMetricAvailability]'}, + 'dimensions': {'key': 'dimensions', 'type': '[OperationMetricDimension]'}, } def __init__(self, **kwargs): @@ -61,3 +65,4 @@ def __init__(self, **kwargs): self.source_mdm_account = kwargs.get('source_mdm_account', None) self.source_mdm_namespace = kwargs.get('source_mdm_namespace', None) self.availabilities = kwargs.get('availabilities', None) + self.dimensions = kwargs.get('dimensions', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_metric_specification_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_metric_specification_py3.py index 788106b93dfd..c1cc4ad39e72 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_metric_specification_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_metric_specification_py3.py @@ -36,6 +36,9 @@ class OperationMetricSpecification(Model): available. :type availabilities: list[~azure.mgmt.datafactory.models.OperationMetricAvailability] + :param dimensions: Defines the metric dimension. + :type dimensions: + list[~azure.mgmt.datafactory.models.OperationMetricDimension] """ _attribute_map = { @@ -48,9 +51,10 @@ class OperationMetricSpecification(Model): 'source_mdm_account': {'key': 'sourceMdmAccount', 'type': 'str'}, 'source_mdm_namespace': {'key': 'sourceMdmNamespace', 'type': 'str'}, 'availabilities': {'key': 'availabilities', 'type': '[OperationMetricAvailability]'}, + 'dimensions': {'key': 'dimensions', 'type': '[OperationMetricDimension]'}, } - def __init__(self, *, name: str=None, display_name: str=None, display_description: str=None, unit: str=None, aggregation_type: str=None, enable_regional_mdm_account: str=None, source_mdm_account: str=None, source_mdm_namespace: str=None, availabilities=None, **kwargs) -> None: + def __init__(self, *, name: str=None, display_name: str=None, display_description: str=None, unit: str=None, aggregation_type: str=None, enable_regional_mdm_account: str=None, source_mdm_account: str=None, source_mdm_namespace: str=None, availabilities=None, dimensions=None, **kwargs) -> None: super(OperationMetricSpecification, self).__init__(**kwargs) self.name = name self.display_name = display_name @@ -61,3 +65,4 @@ def __init__(self, *, name: str=None, display_name: str=None, display_descriptio self.source_mdm_account = source_mdm_account self.source_mdm_namespace = source_mdm_namespace self.availabilities = availabilities + self.dimensions = dimensions diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger_run_paged.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_paged.py similarity index 70% rename from azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger_run_paged.py rename to azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_paged.py index c9ad99f8ced0..d6eea01bbdb9 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger_run_paged.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/operation_paged.py @@ -12,16 +12,16 @@ from msrest.paging import Paged -class TriggerRunPaged(Paged): +class OperationPaged(Paged): """ - A paging container for iterating over a list of :class:`TriggerRun ` object + A paging container for iterating over a list of :class:`Operation ` object """ _attribute_map = { 'next_link': {'key': 'nextLink', 'type': 'str'}, - 'current_page': {'key': 'value', 'type': '[TriggerRun]'} + 'current_page': {'key': 'value', 'type': '[Operation]'} } def __init__(self, *args, **kwargs): - super(TriggerRunPaged, self).__init__(*args, **kwargs) + super(OperationPaged, self).__init__(*args, **kwargs) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run_invoked_by.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run_invoked_by.py index c43b642159d8..acefb80fd078 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run_invoked_by.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run_invoked_by.py @@ -22,19 +22,24 @@ class PipelineRunInvokedBy(Model): :vartype name: str :ivar id: The ID of the entity that started the run. :vartype id: str + :ivar invoked_by_type: The type of the entity that started the run. + :vartype invoked_by_type: str """ _validation = { 'name': {'readonly': True}, 'id': {'readonly': True}, + 'invoked_by_type': {'readonly': True}, } _attribute_map = { 'name': {'key': 'name', 'type': 'str'}, 'id': {'key': 'id', 'type': 'str'}, + 'invoked_by_type': {'key': 'invokedByType', 'type': 'str'}, } def __init__(self, **kwargs): super(PipelineRunInvokedBy, self).__init__(**kwargs) self.name = None self.id = None + self.invoked_by_type = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run_invoked_by_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run_invoked_by_py3.py index d7540cdd0b0f..c954a18b8a67 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run_invoked_by_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run_invoked_by_py3.py @@ -22,19 +22,24 @@ class PipelineRunInvokedBy(Model): :vartype name: str :ivar id: The ID of the entity that started the run. :vartype id: str + :ivar invoked_by_type: The type of the entity that started the run. + :vartype invoked_by_type: str """ _validation = { 'name': {'readonly': True}, 'id': {'readonly': True}, + 'invoked_by_type': {'readonly': True}, } _attribute_map = { 'name': {'key': 'name', 'type': 'str'}, 'id': {'key': 'id', 'type': 'str'}, + 'invoked_by_type': {'key': 'invokedByType', 'type': 'str'}, } def __init__(self, **kwargs) -> None: super(PipelineRunInvokedBy, self).__init__(**kwargs) self.name = None self.id = None + self.invoked_by_type = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run_query_response.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_runs_query_response.py similarity index 92% rename from azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run_query_response.py rename to azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_runs_query_response.py index 50639429fc78..c4591c5467ba 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run_query_response.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_runs_query_response.py @@ -12,7 +12,7 @@ from msrest.serialization import Model -class PipelineRunQueryResponse(Model): +class PipelineRunsQueryResponse(Model): """A list pipeline runs. All required parameters must be populated in order to send to Azure. @@ -34,6 +34,6 @@ class PipelineRunQueryResponse(Model): } def __init__(self, **kwargs): - super(PipelineRunQueryResponse, self).__init__(**kwargs) + super(PipelineRunsQueryResponse, self).__init__(**kwargs) self.value = kwargs.get('value', None) self.continuation_token = kwargs.get('continuation_token', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run_query_response_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_runs_query_response_py3.py similarity index 92% rename from azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run_query_response_py3.py rename to azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_runs_query_response_py3.py index f0aef3ea8bc8..fbc689ec1632 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run_query_response_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_runs_query_response_py3.py @@ -12,7 +12,7 @@ from msrest.serialization import Model -class PipelineRunQueryResponse(Model): +class PipelineRunsQueryResponse(Model): """A list pipeline runs. All required parameters must be populated in order to send to Azure. @@ -34,6 +34,6 @@ class PipelineRunQueryResponse(Model): } def __init__(self, *, value, continuation_token: str=None, **kwargs) -> None: - super(PipelineRunQueryResponse, self).__init__(**kwargs) + super(PipelineRunsQueryResponse, self).__init__(**kwargs) self.value = value self.continuation_token = continuation_token diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/resource.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/resource.py index 38d454c351e6..f6b2d7d3b512 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/resource.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/resource.py @@ -28,12 +28,15 @@ class Resource(Model): :type location: str :param tags: The resource tags. :type tags: dict[str, str] + :ivar e_tag: Etag identifies change in the resource. + :vartype e_tag: str """ _validation = { 'id': {'readonly': True}, 'name': {'readonly': True}, 'type': {'readonly': True}, + 'e_tag': {'readonly': True}, } _attribute_map = { @@ -42,6 +45,7 @@ class Resource(Model): 'type': {'key': 'type', 'type': 'str'}, 'location': {'key': 'location', 'type': 'str'}, 'tags': {'key': 'tags', 'type': '{str}'}, + 'e_tag': {'key': 'eTag', 'type': 'str'}, } def __init__(self, **kwargs): @@ -51,3 +55,4 @@ def __init__(self, **kwargs): self.type = None self.location = kwargs.get('location', None) self.tags = kwargs.get('tags', None) + self.e_tag = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/resource_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/resource_py3.py index 47cbb7d6aa4b..cfc0e4b09aa5 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/resource_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/resource_py3.py @@ -28,12 +28,15 @@ class Resource(Model): :type location: str :param tags: The resource tags. :type tags: dict[str, str] + :ivar e_tag: Etag identifies change in the resource. + :vartype e_tag: str """ _validation = { 'id': {'readonly': True}, 'name': {'readonly': True}, 'type': {'readonly': True}, + 'e_tag': {'readonly': True}, } _attribute_map = { @@ -42,6 +45,7 @@ class Resource(Model): 'type': {'key': 'type', 'type': 'str'}, 'location': {'key': 'location', 'type': 'str'}, 'tags': {'key': 'tags', 'type': '{str}'}, + 'e_tag': {'key': 'eTag', 'type': 'str'}, } def __init__(self, *, location: str=None, tags=None, **kwargs) -> None: @@ -51,3 +55,4 @@ def __init__(self, *, location: str=None, tags=None, **kwargs) -> None: self.type = None self.location = location self.tags = tags + self.e_tag = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run_filter_parameters.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/run_filter_parameters.py similarity index 74% rename from azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run_filter_parameters.py rename to azure-mgmt-datafactory/azure/mgmt/datafactory/models/run_filter_parameters.py index 5e1341649f19..9271f7adf029 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run_filter_parameters.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/run_filter_parameters.py @@ -12,25 +12,24 @@ from msrest.serialization import Model -class PipelineRunFilterParameters(Model): - """Query parameters for listing pipeline runs. +class RunFilterParameters(Model): + """Query parameters for listing runs. All required parameters must be populated in order to send to Azure. :param continuation_token: The continuation token for getting the next page of results. Null for first page. :type continuation_token: str - :param last_updated_after: Required. The time at or after which the - pipeline run event was updated in 'ISO 8601' format. + :param last_updated_after: Required. The time at or after which the run + event was updated in 'ISO 8601' format. :type last_updated_after: datetime - :param last_updated_before: Required. The time at or before which the - pipeline run event was updated in 'ISO 8601' format. + :param last_updated_before: Required. The time at or before which the run + event was updated in 'ISO 8601' format. :type last_updated_before: datetime :param filters: List of filters. - :type filters: list[~azure.mgmt.datafactory.models.PipelineRunQueryFilter] + :type filters: list[~azure.mgmt.datafactory.models.RunQueryFilter] :param order_by: List of OrderBy option. - :type order_by: - list[~azure.mgmt.datafactory.models.PipelineRunQueryOrderBy] + :type order_by: list[~azure.mgmt.datafactory.models.RunQueryOrderBy] """ _validation = { @@ -42,12 +41,12 @@ class PipelineRunFilterParameters(Model): 'continuation_token': {'key': 'continuationToken', 'type': 'str'}, 'last_updated_after': {'key': 'lastUpdatedAfter', 'type': 'iso-8601'}, 'last_updated_before': {'key': 'lastUpdatedBefore', 'type': 'iso-8601'}, - 'filters': {'key': 'filters', 'type': '[PipelineRunQueryFilter]'}, - 'order_by': {'key': 'orderBy', 'type': '[PipelineRunQueryOrderBy]'}, + 'filters': {'key': 'filters', 'type': '[RunQueryFilter]'}, + 'order_by': {'key': 'orderBy', 'type': '[RunQueryOrderBy]'}, } def __init__(self, **kwargs): - super(PipelineRunFilterParameters, self).__init__(**kwargs) + super(RunFilterParameters, self).__init__(**kwargs) self.continuation_token = kwargs.get('continuation_token', None) self.last_updated_after = kwargs.get('last_updated_after', None) self.last_updated_before = kwargs.get('last_updated_before', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run_filter_parameters_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/run_filter_parameters_py3.py similarity index 74% rename from azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run_filter_parameters_py3.py rename to azure-mgmt-datafactory/azure/mgmt/datafactory/models/run_filter_parameters_py3.py index a436b8e4023c..c96e64eb63b3 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run_filter_parameters_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/run_filter_parameters_py3.py @@ -12,25 +12,24 @@ from msrest.serialization import Model -class PipelineRunFilterParameters(Model): - """Query parameters for listing pipeline runs. +class RunFilterParameters(Model): + """Query parameters for listing runs. All required parameters must be populated in order to send to Azure. :param continuation_token: The continuation token for getting the next page of results. Null for first page. :type continuation_token: str - :param last_updated_after: Required. The time at or after which the - pipeline run event was updated in 'ISO 8601' format. + :param last_updated_after: Required. The time at or after which the run + event was updated in 'ISO 8601' format. :type last_updated_after: datetime - :param last_updated_before: Required. The time at or before which the - pipeline run event was updated in 'ISO 8601' format. + :param last_updated_before: Required. The time at or before which the run + event was updated in 'ISO 8601' format. :type last_updated_before: datetime :param filters: List of filters. - :type filters: list[~azure.mgmt.datafactory.models.PipelineRunQueryFilter] + :type filters: list[~azure.mgmt.datafactory.models.RunQueryFilter] :param order_by: List of OrderBy option. - :type order_by: - list[~azure.mgmt.datafactory.models.PipelineRunQueryOrderBy] + :type order_by: list[~azure.mgmt.datafactory.models.RunQueryOrderBy] """ _validation = { @@ -42,12 +41,12 @@ class PipelineRunFilterParameters(Model): 'continuation_token': {'key': 'continuationToken', 'type': 'str'}, 'last_updated_after': {'key': 'lastUpdatedAfter', 'type': 'iso-8601'}, 'last_updated_before': {'key': 'lastUpdatedBefore', 'type': 'iso-8601'}, - 'filters': {'key': 'filters', 'type': '[PipelineRunQueryFilter]'}, - 'order_by': {'key': 'orderBy', 'type': '[PipelineRunQueryOrderBy]'}, + 'filters': {'key': 'filters', 'type': '[RunQueryFilter]'}, + 'order_by': {'key': 'orderBy', 'type': '[RunQueryOrderBy]'}, } def __init__(self, *, last_updated_after, last_updated_before, continuation_token: str=None, filters=None, order_by=None, **kwargs) -> None: - super(PipelineRunFilterParameters, self).__init__(**kwargs) + super(RunFilterParameters, self).__init__(**kwargs) self.continuation_token = continuation_token self.last_updated_after = last_updated_after self.last_updated_before = last_updated_before diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run_query_filter.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/run_query_filter.py similarity index 65% rename from azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run_query_filter.py rename to azure-mgmt-datafactory/azure/mgmt/datafactory/models/run_query_filter.py index bca4438a6ed6..63a4cddc063d 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run_query_filter.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/run_query_filter.py @@ -12,19 +12,24 @@ from msrest.serialization import Model -class PipelineRunQueryFilter(Model): - """Query filter option for listing pipeline runs. +class RunQueryFilter(Model): + """Query filter option for listing runs. All required parameters must be populated in order to send to Azure. - :param operand: Required. Parameter name to be used for filter. Possible - values include: 'PipelineName', 'Status', 'RunStart', 'RunEnd' - :type operand: str or - ~azure.mgmt.datafactory.models.PipelineRunQueryFilterOperand + :param operand: Required. Parameter name to be used for filter. The + allowed operands to query pipeline runs are PipelineName, RunStart, RunEnd + and Status; to query activity runs are ActivityName, ActivityRunStart, + ActivityRunEnd, ActivityType and Status, and to query trigger runs are + TriggerName, TriggerRunTimestamp and Status. Possible values include: + 'PipelineName', 'Status', 'RunStart', 'RunEnd', 'ActivityName', + 'ActivityRunStart', 'ActivityRunEnd', 'ActivityType', 'TriggerName', + 'TriggerRunTimestamp' + :type operand: str or ~azure.mgmt.datafactory.models.RunQueryFilterOperand :param operator: Required. Operator to be used for filter. Possible values include: 'Equals', 'NotEquals', 'In', 'NotIn' :type operator: str or - ~azure.mgmt.datafactory.models.PipelineRunQueryFilterOperator + ~azure.mgmt.datafactory.models.RunQueryFilterOperator :param values: Required. List of filter values. :type values: list[str] """ @@ -42,7 +47,7 @@ class PipelineRunQueryFilter(Model): } def __init__(self, **kwargs): - super(PipelineRunQueryFilter, self).__init__(**kwargs) + super(RunQueryFilter, self).__init__(**kwargs) self.operand = kwargs.get('operand', None) self.operator = kwargs.get('operator', None) self.values = kwargs.get('values', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run_query_filter_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/run_query_filter_py3.py similarity index 65% rename from azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run_query_filter_py3.py rename to azure-mgmt-datafactory/azure/mgmt/datafactory/models/run_query_filter_py3.py index 3e6d24a3ac73..fc95591801bd 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run_query_filter_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/run_query_filter_py3.py @@ -12,19 +12,24 @@ from msrest.serialization import Model -class PipelineRunQueryFilter(Model): - """Query filter option for listing pipeline runs. +class RunQueryFilter(Model): + """Query filter option for listing runs. All required parameters must be populated in order to send to Azure. - :param operand: Required. Parameter name to be used for filter. Possible - values include: 'PipelineName', 'Status', 'RunStart', 'RunEnd' - :type operand: str or - ~azure.mgmt.datafactory.models.PipelineRunQueryFilterOperand + :param operand: Required. Parameter name to be used for filter. The + allowed operands to query pipeline runs are PipelineName, RunStart, RunEnd + and Status; to query activity runs are ActivityName, ActivityRunStart, + ActivityRunEnd, ActivityType and Status, and to query trigger runs are + TriggerName, TriggerRunTimestamp and Status. Possible values include: + 'PipelineName', 'Status', 'RunStart', 'RunEnd', 'ActivityName', + 'ActivityRunStart', 'ActivityRunEnd', 'ActivityType', 'TriggerName', + 'TriggerRunTimestamp' + :type operand: str or ~azure.mgmt.datafactory.models.RunQueryFilterOperand :param operator: Required. Operator to be used for filter. Possible values include: 'Equals', 'NotEquals', 'In', 'NotIn' :type operator: str or - ~azure.mgmt.datafactory.models.PipelineRunQueryFilterOperator + ~azure.mgmt.datafactory.models.RunQueryFilterOperator :param values: Required. List of filter values. :type values: list[str] """ @@ -42,7 +47,7 @@ class PipelineRunQueryFilter(Model): } def __init__(self, *, operand, operator, values, **kwargs) -> None: - super(PipelineRunQueryFilter, self).__init__(**kwargs) + super(RunQueryFilter, self).__init__(**kwargs) self.operand = operand self.operator = operator self.values = values diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run_query_order_by.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/run_query_order_by.py similarity index 60% rename from azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run_query_order_by.py rename to azure-mgmt-datafactory/azure/mgmt/datafactory/models/run_query_order_by.py index 27da9656c88b..21afabcf215f 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run_query_order_by.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/run_query_order_by.py @@ -12,18 +12,22 @@ from msrest.serialization import Model -class PipelineRunQueryOrderBy(Model): - """An object to provide order by options for listing pipeline runs. +class RunQueryOrderBy(Model): + """An object to provide order by options for listing runs. All required parameters must be populated in order to send to Azure. - :param order_by: Required. Parameter name to be used for order by. - Possible values include: 'RunStart', 'RunEnd' - :type order_by: str or - ~azure.mgmt.datafactory.models.PipelineRunQueryOrderByField + :param order_by: Required. Parameter name to be used for order by. The + allowed parameters to order by for pipeline runs are PipelineName, + RunStart, RunEnd and Status; for activity runs are ActivityName, + ActivityRunStart, ActivityRunEnd and Status; for trigger runs are + TriggerName, TriggerRunTimestamp and Status. Possible values include: + 'RunStart', 'RunEnd', 'PipelineName', 'Status', 'ActivityName', + 'ActivityRunStart', 'ActivityRunEnd', 'TriggerName', 'TriggerRunTimestamp' + :type order_by: str or ~azure.mgmt.datafactory.models.RunQueryOrderByField :param order: Required. Sorting order of the parameter. Possible values include: 'ASC', 'DESC' - :type order: str or ~azure.mgmt.datafactory.models.PipelineRunQueryOrder + :type order: str or ~azure.mgmt.datafactory.models.RunQueryOrder """ _validation = { @@ -37,6 +41,6 @@ class PipelineRunQueryOrderBy(Model): } def __init__(self, **kwargs): - super(PipelineRunQueryOrderBy, self).__init__(**kwargs) + super(RunQueryOrderBy, self).__init__(**kwargs) self.order_by = kwargs.get('order_by', None) self.order = kwargs.get('order', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run_query_order_by_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/run_query_order_by_py3.py similarity index 60% rename from azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run_query_order_by_py3.py rename to azure-mgmt-datafactory/azure/mgmt/datafactory/models/run_query_order_by_py3.py index 4edaf9ba99de..a3ddc8854d47 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_run_query_order_by_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/run_query_order_by_py3.py @@ -12,18 +12,22 @@ from msrest.serialization import Model -class PipelineRunQueryOrderBy(Model): - """An object to provide order by options for listing pipeline runs. +class RunQueryOrderBy(Model): + """An object to provide order by options for listing runs. All required parameters must be populated in order to send to Azure. - :param order_by: Required. Parameter name to be used for order by. - Possible values include: 'RunStart', 'RunEnd' - :type order_by: str or - ~azure.mgmt.datafactory.models.PipelineRunQueryOrderByField + :param order_by: Required. Parameter name to be used for order by. The + allowed parameters to order by for pipeline runs are PipelineName, + RunStart, RunEnd and Status; for activity runs are ActivityName, + ActivityRunStart, ActivityRunEnd and Status; for trigger runs are + TriggerName, TriggerRunTimestamp and Status. Possible values include: + 'RunStart', 'RunEnd', 'PipelineName', 'Status', 'ActivityName', + 'ActivityRunStart', 'ActivityRunEnd', 'TriggerName', 'TriggerRunTimestamp' + :type order_by: str or ~azure.mgmt.datafactory.models.RunQueryOrderByField :param order: Required. Sorting order of the parameter. Possible values include: 'ASC', 'DESC' - :type order: str or ~azure.mgmt.datafactory.models.PipelineRunQueryOrder + :type order: str or ~azure.mgmt.datafactory.models.RunQueryOrder """ _validation = { @@ -37,6 +41,6 @@ class PipelineRunQueryOrderBy(Model): } def __init__(self, *, order_by, order, **kwargs) -> None: - super(PipelineRunQueryOrderBy, self).__init__(**kwargs) + super(RunQueryOrderBy, self).__init__(**kwargs) self.order_by = order_by self.order = order diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime.py index d2b334ee11b5..b9c6c2870fff 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime.py @@ -24,9 +24,6 @@ class SelfHostedIntegrationRuntime(IntegrationRuntime): :type description: str :param type: Required. Constant filled by server. :type type: str - :param linked_info: - :type linked_info: - ~azure.mgmt.datafactory.models.LinkedIntegrationRuntimeProperties """ _validation = { @@ -37,10 +34,8 @@ class SelfHostedIntegrationRuntime(IntegrationRuntime): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'type': {'key': 'type', 'type': 'str'}, - 'linked_info': {'key': 'typeProperties.linkedInfo', 'type': 'LinkedIntegrationRuntimeProperties'}, } def __init__(self, **kwargs): super(SelfHostedIntegrationRuntime, self).__init__(**kwargs) - self.linked_info = kwargs.get('linked_info', None) self.type = 'SelfHosted' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime_py3.py index 2a6929ecdd15..744d92d28a5a 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime_py3.py @@ -24,9 +24,6 @@ class SelfHostedIntegrationRuntime(IntegrationRuntime): :type description: str :param type: Required. Constant filled by server. :type type: str - :param linked_info: - :type linked_info: - ~azure.mgmt.datafactory.models.LinkedIntegrationRuntimeProperties """ _validation = { @@ -37,10 +34,8 @@ class SelfHostedIntegrationRuntime(IntegrationRuntime): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'type': {'key': 'type', 'type': 'str'}, - 'linked_info': {'key': 'typeProperties.linkedInfo', 'type': 'LinkedIntegrationRuntimeProperties'}, } - def __init__(self, *, additional_properties=None, description: str=None, linked_info=None, **kwargs) -> None: + def __init__(self, *, additional_properties=None, description: str=None, **kwargs) -> None: super(SelfHostedIntegrationRuntime, self).__init__(additional_properties=additional_properties, description=description, **kwargs) - self.linked_info = linked_info self.type = 'SelfHosted' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime_status.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime_status.py index 7bd67718e078..3a07410bb21e 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime_status.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime_status.py @@ -69,9 +69,11 @@ class SelfHostedIntegrationRuntimeStatus(IntegrationRuntimeStatus): ~azure.mgmt.datafactory.models.IntegrationRuntimeAutoUpdate :ivar version_status: Status of the integration runtime version. :vartype version_status: str - :param links: The list of linked integration runtimes that are created to - share with this integration runtime. - :type links: list[~azure.mgmt.datafactory.models.LinkedIntegrationRuntime] + :ivar pushed_version: The version that the integration runtime is going to + update to. + :vartype pushed_version: str + :ivar latest_version: The latest version on download center. + :vartype latest_version: str """ _validation = { @@ -89,6 +91,8 @@ class SelfHostedIntegrationRuntimeStatus(IntegrationRuntimeStatus): 'service_urls': {'readonly': True}, 'auto_update': {'readonly': True}, 'version_status': {'readonly': True}, + 'pushed_version': {'readonly': True}, + 'latest_version': {'readonly': True}, } _attribute_map = { @@ -108,7 +112,8 @@ class SelfHostedIntegrationRuntimeStatus(IntegrationRuntimeStatus): 'service_urls': {'key': 'typeProperties.serviceUrls', 'type': '[str]'}, 'auto_update': {'key': 'typeProperties.autoUpdate', 'type': 'str'}, 'version_status': {'key': 'typeProperties.versionStatus', 'type': 'str'}, - 'links': {'key': 'typeProperties.links', 'type': '[LinkedIntegrationRuntime]'}, + 'pushed_version': {'key': 'typeProperties.pushedVersion', 'type': 'str'}, + 'latest_version': {'key': 'typeProperties.latestVersion', 'type': 'str'}, } def __init__(self, **kwargs): @@ -125,5 +130,6 @@ def __init__(self, **kwargs): self.service_urls = None self.auto_update = None self.version_status = None - self.links = kwargs.get('links', None) + self.pushed_version = None + self.latest_version = None self.type = 'SelfHosted' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime_status_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime_status_py3.py index 07362ed1835a..0714c35ca388 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime_status_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime_status_py3.py @@ -69,9 +69,11 @@ class SelfHostedIntegrationRuntimeStatus(IntegrationRuntimeStatus): ~azure.mgmt.datafactory.models.IntegrationRuntimeAutoUpdate :ivar version_status: Status of the integration runtime version. :vartype version_status: str - :param links: The list of linked integration runtimes that are created to - share with this integration runtime. - :type links: list[~azure.mgmt.datafactory.models.LinkedIntegrationRuntime] + :ivar pushed_version: The version that the integration runtime is going to + update to. + :vartype pushed_version: str + :ivar latest_version: The latest version on download center. + :vartype latest_version: str """ _validation = { @@ -89,6 +91,8 @@ class SelfHostedIntegrationRuntimeStatus(IntegrationRuntimeStatus): 'service_urls': {'readonly': True}, 'auto_update': {'readonly': True}, 'version_status': {'readonly': True}, + 'pushed_version': {'readonly': True}, + 'latest_version': {'readonly': True}, } _attribute_map = { @@ -108,10 +112,11 @@ class SelfHostedIntegrationRuntimeStatus(IntegrationRuntimeStatus): 'service_urls': {'key': 'typeProperties.serviceUrls', 'type': '[str]'}, 'auto_update': {'key': 'typeProperties.autoUpdate', 'type': 'str'}, 'version_status': {'key': 'typeProperties.versionStatus', 'type': 'str'}, - 'links': {'key': 'typeProperties.links', 'type': '[LinkedIntegrationRuntime]'}, + 'pushed_version': {'key': 'typeProperties.pushedVersion', 'type': 'str'}, + 'latest_version': {'key': 'typeProperties.latestVersion', 'type': 'str'}, } - def __init__(self, *, additional_properties=None, nodes=None, links=None, **kwargs) -> None: + def __init__(self, *, additional_properties=None, nodes=None, **kwargs) -> None: super(SelfHostedIntegrationRuntimeStatus, self).__init__(additional_properties=additional_properties, **kwargs) self.create_time = None self.task_queue_id = None @@ -125,5 +130,6 @@ def __init__(self, *, additional_properties=None, nodes=None, links=None, **kwar self.service_urls = None self.auto_update = None self.version_status = None - self.links = links + self.pushed_version = None + self.latest_version = None self.type = 'SelfHosted' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_stored_procedure_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_stored_procedure_activity.py index 4d39b9d806f1..5a80087427fa 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_stored_procedure_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_stored_procedure_activity.py @@ -26,6 +26,8 @@ class SqlServerStoredProcedureActivity(ExecutionActivity): :type description: str :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param user_properties: Activity user properties. + :type user_properties: dict[str, str] :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. @@ -53,6 +55,7 @@ class SqlServerStoredProcedureActivity(ExecutionActivity): 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'user_properties': {'key': 'userProperties', 'type': '{str}'}, 'type': {'key': 'type', 'type': 'str'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_stored_procedure_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_stored_procedure_activity_py3.py index b1166e0de545..f7200f8f5776 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_stored_procedure_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_stored_procedure_activity_py3.py @@ -26,6 +26,8 @@ class SqlServerStoredProcedureActivity(ExecutionActivity): :type description: str :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param user_properties: Activity user properties. + :type user_properties: dict[str, str] :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. @@ -53,6 +55,7 @@ class SqlServerStoredProcedureActivity(ExecutionActivity): 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'user_properties': {'key': 'userProperties', 'type': '{str}'}, 'type': {'key': 'type', 'type': 'str'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, @@ -60,8 +63,8 @@ class SqlServerStoredProcedureActivity(ExecutionActivity): 'stored_procedure_parameters': {'key': 'typeProperties.storedProcedureParameters', 'type': '{StoredProcedureParameter}'}, } - def __init__(self, *, name: str, stored_procedure_name, additional_properties=None, description: str=None, depends_on=None, linked_service_name=None, policy=None, stored_procedure_parameters=None, **kwargs) -> None: - super(SqlServerStoredProcedureActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, linked_service_name=linked_service_name, policy=policy, **kwargs) + def __init__(self, *, name: str, stored_procedure_name, additional_properties=None, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, stored_procedure_parameters=None, **kwargs) -> None: + super(SqlServerStoredProcedureActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) self.stored_procedure_name = stored_procedure_name self.stored_procedure_parameters = stored_procedure_parameters self.type = 'SqlServerStoredProcedure' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger_runs_query_response.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger_runs_query_response.py new file mode 100644 index 000000000000..7684fe7eb7dc --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger_runs_query_response.py @@ -0,0 +1,39 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class TriggerRunsQueryResponse(Model): + """A list of trigger runs. + + All required parameters must be populated in order to send to Azure. + + :param value: Required. List of trigger runs. + :type value: list[~azure.mgmt.datafactory.models.TriggerRun] + :param continuation_token: The continuation token for getting the next + page of results, if any remaining results exist, null otherwise. + :type continuation_token: str + """ + + _validation = { + 'value': {'required': True}, + } + + _attribute_map = { + 'value': {'key': 'value', 'type': '[TriggerRun]'}, + 'continuation_token': {'key': 'continuationToken', 'type': 'str'}, + } + + def __init__(self, **kwargs): + super(TriggerRunsQueryResponse, self).__init__(**kwargs) + self.value = kwargs.get('value', None) + self.continuation_token = kwargs.get('continuation_token', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger_runs_query_response_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger_runs_query_response_py3.py new file mode 100644 index 000000000000..391a2441b3d1 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger_runs_query_response_py3.py @@ -0,0 +1,39 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class TriggerRunsQueryResponse(Model): + """A list of trigger runs. + + All required parameters must be populated in order to send to Azure. + + :param value: Required. List of trigger runs. + :type value: list[~azure.mgmt.datafactory.models.TriggerRun] + :param continuation_token: The continuation token for getting the next + page of results, if any remaining results exist, null otherwise. + :type continuation_token: str + """ + + _validation = { + 'value': {'required': True}, + } + + _attribute_map = { + 'value': {'key': 'value', 'type': '[TriggerRun]'}, + 'continuation_token': {'key': 'continuationToken', 'type': 'str'}, + } + + def __init__(self, *, value, continuation_token: str=None, **kwargs) -> None: + super(TriggerRunsQueryResponse, self).__init__(**kwargs) + self.value = value + self.continuation_token = continuation_token diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/until_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/until_activity.py index 7cf4fc07594b..707c02e538c5 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/until_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/until_activity.py @@ -27,6 +27,8 @@ class UntilActivity(ControlActivity): :type description: str :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param user_properties: Activity user properties. + :type user_properties: dict[str, str] :param type: Required. Constant filled by server. :type type: str :param expression: Required. An expression that would evaluate to Boolean. @@ -55,6 +57,7 @@ class UntilActivity(ControlActivity): 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'user_properties': {'key': 'userProperties', 'type': '{str}'}, 'type': {'key': 'type', 'type': 'str'}, 'expression': {'key': 'typeProperties.expression', 'type': 'Expression'}, 'timeout': {'key': 'typeProperties.timeout', 'type': 'object'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/until_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/until_activity_py3.py index a73467b7123b..ac23fd6d4e92 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/until_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/until_activity_py3.py @@ -27,6 +27,8 @@ class UntilActivity(ControlActivity): :type description: str :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param user_properties: Activity user properties. + :type user_properties: dict[str, str] :param type: Required. Constant filled by server. :type type: str :param expression: Required. An expression that would evaluate to Boolean. @@ -55,14 +57,15 @@ class UntilActivity(ControlActivity): 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'user_properties': {'key': 'userProperties', 'type': '{str}'}, 'type': {'key': 'type', 'type': 'str'}, 'expression': {'key': 'typeProperties.expression', 'type': 'Expression'}, 'timeout': {'key': 'typeProperties.timeout', 'type': 'object'}, 'activities': {'key': 'typeProperties.activities', 'type': '[Activity]'}, } - def __init__(self, *, name: str, expression, activities, additional_properties=None, description: str=None, depends_on=None, timeout=None, **kwargs) -> None: - super(UntilActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, **kwargs) + def __init__(self, *, name: str, expression, activities, additional_properties=None, description: str=None, depends_on=None, user_properties=None, timeout=None, **kwargs) -> None: + super(UntilActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, **kwargs) self.expression = expression self.timeout = timeout self.activities = activities diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/wait_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/wait_activity.py index 0ed8536f0afc..e2000f3f588c 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/wait_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/wait_activity.py @@ -26,6 +26,8 @@ class WaitActivity(ControlActivity): :type description: str :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param user_properties: Activity user properties. + :type user_properties: dict[str, str] :param type: Required. Constant filled by server. :type type: str :param wait_time_in_seconds: Required. Duration in seconds. @@ -43,6 +45,7 @@ class WaitActivity(ControlActivity): 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'user_properties': {'key': 'userProperties', 'type': '{str}'}, 'type': {'key': 'type', 'type': 'str'}, 'wait_time_in_seconds': {'key': 'typeProperties.waitTimeInSeconds', 'type': 'int'}, } diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/wait_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/wait_activity_py3.py index ee3da26e5b83..4170198d8d3b 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/wait_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/wait_activity_py3.py @@ -26,6 +26,8 @@ class WaitActivity(ControlActivity): :type description: str :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param user_properties: Activity user properties. + :type user_properties: dict[str, str] :param type: Required. Constant filled by server. :type type: str :param wait_time_in_seconds: Required. Duration in seconds. @@ -43,11 +45,12 @@ class WaitActivity(ControlActivity): 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'user_properties': {'key': 'userProperties', 'type': '{str}'}, 'type': {'key': 'type', 'type': 'str'}, 'wait_time_in_seconds': {'key': 'typeProperties.waitTimeInSeconds', 'type': 'int'}, } - def __init__(self, *, name: str, wait_time_in_seconds: int, additional_properties=None, description: str=None, depends_on=None, **kwargs) -> None: - super(WaitActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, **kwargs) + def __init__(self, *, name: str, wait_time_in_seconds: int, additional_properties=None, description: str=None, depends_on=None, user_properties=None, **kwargs) -> None: + super(WaitActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, **kwargs) self.wait_time_in_seconds = wait_time_in_seconds self.type = 'Wait' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_activity.py index 65831d837ae9..017aea067c90 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_activity.py @@ -26,6 +26,8 @@ class WebActivity(ExecutionActivity): :type description: str :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param user_properties: Activity user properties. + :type user_properties: dict[str, str] :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. @@ -71,6 +73,7 @@ class WebActivity(ExecutionActivity): 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'user_properties': {'key': 'userProperties', 'type': '{str}'}, 'type': {'key': 'type', 'type': 'str'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_activity_py3.py index c84a8e4d31d2..5bdf1c4a0ac8 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_activity_py3.py @@ -26,6 +26,8 @@ class WebActivity(ExecutionActivity): :type description: str :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param user_properties: Activity user properties. + :type user_properties: dict[str, str] :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. @@ -71,6 +73,7 @@ class WebActivity(ExecutionActivity): 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'user_properties': {'key': 'userProperties', 'type': '{str}'}, 'type': {'key': 'type', 'type': 'str'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, @@ -83,8 +86,8 @@ class WebActivity(ExecutionActivity): 'linked_services': {'key': 'typeProperties.linkedServices', 'type': '[LinkedServiceReference]'}, } - def __init__(self, *, name: str, method, url, additional_properties=None, description: str=None, depends_on=None, linked_service_name=None, policy=None, headers=None, body=None, authentication=None, datasets=None, linked_services=None, **kwargs) -> None: - super(WebActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, linked_service_name=linked_service_name, policy=policy, **kwargs) + def __init__(self, *, name: str, method, url, additional_properties=None, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, headers=None, body=None, authentication=None, datasets=None, linked_services=None, **kwargs) -> None: + super(WebActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) self.method = method self.url = url self.headers = headers diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/__init__.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/__init__.py index e074c1ecd77a..b6b9497ae922 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/__init__.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/__init__.py @@ -19,6 +19,7 @@ from .pipeline_runs_operations import PipelineRunsOperations from .activity_runs_operations import ActivityRunsOperations from .triggers_operations import TriggersOperations +from .trigger_runs_operations import TriggerRunsOperations __all__ = [ 'Operations', @@ -31,4 +32,5 @@ 'PipelineRunsOperations', 'ActivityRunsOperations', 'TriggersOperations', + 'TriggerRunsOperations', ] diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/activity_runs_operations.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/activity_runs_operations.py index 3cf6354d05f6..ec8288ad562a 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/activity_runs_operations.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/activity_runs_operations.py @@ -11,6 +11,7 @@ import uuid from msrest.pipeline import ClientRawResponse +from msrestazure.azure_exceptions import CloudError from .. import models @@ -22,7 +23,7 @@ class ActivityRunsOperations(object): :param config: Configuration of service client. :param serializer: An object model serializer. :param deserializer: An object model deserializer. - :ivar api_version: The API version. Constant value: "2017-09-01-preview". + :ivar api_version: The API version. Constant value: "2018-06-01". """ models = models @@ -32,13 +33,13 @@ def __init__(self, client, config, serializer, deserializer): self._client = client self._serialize = serializer self._deserialize = deserializer - self.api_version = "2017-09-01-preview" + self.api_version = "2018-06-01" self.config = config - def list_by_pipeline_run( - self, resource_group_name, factory_name, run_id, start_time, end_time, status=None, activity_name=None, linked_service_name=None, custom_headers=None, raw=False, **operation_config): - """List activity runs based on input filter conditions. + def query_by_pipeline_run( + self, resource_group_name, factory_name, run_id, filter_parameters, custom_headers=None, raw=False, **operation_config): + """Query activity runs based on input filter conditions. :param resource_group_name: The resource group name. :type resource_group_name: str @@ -46,83 +47,64 @@ def list_by_pipeline_run( :type factory_name: str :param run_id: The pipeline run identifier. :type run_id: str - :param start_time: The start time of activity runs in ISO8601 format. - :type start_time: datetime - :param end_time: The end time of activity runs in ISO8601 format. - :type end_time: datetime - :param status: The status of the pipeline run. - :type status: str - :param activity_name: The name of the activity. - :type activity_name: str - :param linked_service_name: The linked service name. - :type linked_service_name: str + :param filter_parameters: Parameters to filter the activity runs. + :type filter_parameters: + ~azure.mgmt.datafactory.models.RunFilterParameters :param dict custom_headers: headers that will be added to the request :param bool raw: returns the direct response alongside the deserialized response :param operation_config: :ref:`Operation configuration overrides`. - :return: An iterator like instance of ActivityRun - :rtype: - ~azure.mgmt.datafactory.models.ActivityRunPaged[~azure.mgmt.datafactory.models.ActivityRun] - :raises: - :class:`ErrorResponseException` + :return: ActivityRunsQueryResponse or ClientRawResponse if raw=true + :rtype: ~azure.mgmt.datafactory.models.ActivityRunsQueryResponse or + ~msrest.pipeline.ClientRawResponse + :raises: :class:`CloudError` """ - def internal_paging(next_link=None, raw=False): - - if not next_link: - # Construct URL - url = self.list_by_pipeline_run.metadata['url'] - path_format_arguments = { - 'subscriptionId': self._serialize.url("self.config.subscription_id", self.config.subscription_id, 'str'), - 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str', max_length=90, min_length=1, pattern=r'^[-\w\._\(\)]+$'), - 'factoryName': self._serialize.url("factory_name", factory_name, 'str', max_length=63, min_length=3, pattern=r'^[A-Za-z0-9]+(?:-[A-Za-z0-9]+)*$'), - 'runId': self._serialize.url("run_id", run_id, 'str') - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} - query_parameters['api-version'] = self._serialize.query("self.api_version", self.api_version, 'str') - query_parameters['startTime'] = self._serialize.query("start_time", start_time, 'iso-8601') - query_parameters['endTime'] = self._serialize.query("end_time", end_time, 'iso-8601') - if status is not None: - query_parameters['status'] = self._serialize.query("status", status, 'str') - if activity_name is not None: - query_parameters['activityName'] = self._serialize.query("activity_name", activity_name, 'str') - if linked_service_name is not None: - query_parameters['linkedServiceName'] = self._serialize.query("linked_service_name", linked_service_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$') - - else: - url = next_link - query_parameters = {} - - # Construct headers - header_parameters = {} - header_parameters['Content-Type'] = 'application/json; charset=utf-8' - if self.config.generate_client_request_id: - header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) - if custom_headers: - header_parameters.update(custom_headers) - if self.config.accept_language is not None: - header_parameters['accept-language'] = self._serialize.header("self.config.accept_language", self.config.accept_language, 'str') - - # Construct and send request - request = self._client.get(url, query_parameters) - response = self._client.send( - request, header_parameters, stream=False, **operation_config) - - if response.status_code not in [200]: - raise models.ErrorResponseException(self._deserialize, response) - - return response - - # Deserialize response - deserialized = models.ActivityRunPaged(internal_paging, self._deserialize.dependencies) + # Construct URL + url = self.query_by_pipeline_run.metadata['url'] + path_format_arguments = { + 'subscriptionId': self._serialize.url("self.config.subscription_id", self.config.subscription_id, 'str'), + 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str', max_length=90, min_length=1, pattern=r'^[-\w\._\(\)]+$'), + 'factoryName': self._serialize.url("factory_name", factory_name, 'str', max_length=63, min_length=3, pattern=r'^[A-Za-z0-9]+(?:-[A-Za-z0-9]+)*$'), + 'runId': self._serialize.url("run_id", run_id, 'str') + } + url = self._client.format_url(url, **path_format_arguments) + + # Construct parameters + query_parameters = {} + query_parameters['api-version'] = self._serialize.query("self.api_version", self.api_version, 'str') + + # Construct headers + header_parameters = {} + header_parameters['Content-Type'] = 'application/json; charset=utf-8' + if self.config.generate_client_request_id: + header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) + if custom_headers: + header_parameters.update(custom_headers) + if self.config.accept_language is not None: + header_parameters['accept-language'] = self._serialize.header("self.config.accept_language", self.config.accept_language, 'str') + + # Construct body + body_content = self._serialize.body(filter_parameters, 'RunFilterParameters') + + # Construct and send request + request = self._client.post(url, query_parameters) + response = self._client.send( + request, header_parameters, body_content, stream=False, **operation_config) + + if response.status_code not in [200]: + exp = CloudError(response) + exp.request_id = response.headers.get('x-ms-request-id') + raise exp + + deserialized = None + + if response.status_code == 200: + deserialized = self._deserialize('ActivityRunsQueryResponse', response) if raw: - header_dict = {} - client_raw_response = models.ActivityRunPaged(internal_paging, self._deserialize.dependencies, header_dict) + client_raw_response = ClientRawResponse(deserialized, response) return client_raw_response return deserialized - list_by_pipeline_run.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/pipelineruns/{runId}/activityruns'} + query_by_pipeline_run.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/pipelineruns/{runId}/queryActivityruns'} diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/datasets_operations.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/datasets_operations.py index 9179f1e9dc95..d2768372d196 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/datasets_operations.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/datasets_operations.py @@ -11,6 +11,7 @@ import uuid from msrest.pipeline import ClientRawResponse +from msrestazure.azure_exceptions import CloudError from .. import models @@ -22,7 +23,7 @@ class DatasetsOperations(object): :param config: Configuration of service client. :param serializer: An object model serializer. :param deserializer: An object model deserializer. - :ivar api_version: The API version. Constant value: "2017-09-01-preview". + :ivar api_version: The API version. Constant value: "2018-06-01". """ models = models @@ -32,7 +33,7 @@ def __init__(self, client, config, serializer, deserializer): self._client = client self._serialize = serializer self._deserialize = deserializer - self.api_version = "2017-09-01-preview" + self.api_version = "2018-06-01" self.config = config @@ -52,8 +53,7 @@ def list_by_factory( :return: An iterator like instance of DatasetResource :rtype: ~azure.mgmt.datafactory.models.DatasetResourcePaged[~azure.mgmt.datafactory.models.DatasetResource] - :raises: - :class:`ErrorResponseException` + :raises: :class:`CloudError` """ def internal_paging(next_link=None, raw=False): @@ -91,7 +91,9 @@ def internal_paging(next_link=None, raw=False): request, header_parameters, stream=False, **operation_config) if response.status_code not in [200]: - raise models.ErrorResponseException(self._deserialize, response) + exp = CloudError(response) + exp.request_id = response.headers.get('x-ms-request-id') + raise exp return response @@ -130,8 +132,7 @@ def create_or_update( :return: DatasetResource or ClientRawResponse if raw=true :rtype: ~azure.mgmt.datafactory.models.DatasetResource or ~msrest.pipeline.ClientRawResponse - :raises: - :class:`ErrorResponseException` + :raises: :class:`CloudError` """ dataset = models.DatasetResource(properties=properties) @@ -170,7 +171,9 @@ def create_or_update( request, header_parameters, body_content, stream=False, **operation_config) if response.status_code not in [200]: - raise models.ErrorResponseException(self._deserialize, response) + exp = CloudError(response) + exp.request_id = response.headers.get('x-ms-request-id') + raise exp deserialized = None @@ -185,7 +188,7 @@ def create_or_update( create_or_update.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/datasets/{datasetName}'} def get( - self, resource_group_name, factory_name, dataset_name, custom_headers=None, raw=False, **operation_config): + self, resource_group_name, factory_name, dataset_name, if_none_match=None, custom_headers=None, raw=False, **operation_config): """Gets a dataset. :param resource_group_name: The resource group name. @@ -194,6 +197,10 @@ def get( :type factory_name: str :param dataset_name: The dataset name. :type dataset_name: str + :param if_none_match: ETag of the dataset entity. Should only be + specified for get. If the ETag matches the existing entity tag, or if + * was provided, then no content will be returned. + :type if_none_match: str :param dict custom_headers: headers that will be added to the request :param bool raw: returns the direct response alongside the deserialized response @@ -202,8 +209,7 @@ def get( :return: DatasetResource or ClientRawResponse if raw=true :rtype: ~azure.mgmt.datafactory.models.DatasetResource or ~msrest.pipeline.ClientRawResponse - :raises: - :class:`ErrorResponseException` + :raises: :class:`CloudError` """ # Construct URL url = self.get.metadata['url'] @@ -226,6 +232,8 @@ def get( header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) if custom_headers: header_parameters.update(custom_headers) + if if_none_match is not None: + header_parameters['If-None-Match'] = self._serialize.header("if_none_match", if_none_match, 'str') if self.config.accept_language is not None: header_parameters['accept-language'] = self._serialize.header("self.config.accept_language", self.config.accept_language, 'str') @@ -233,8 +241,10 @@ def get( request = self._client.get(url, query_parameters) response = self._client.send(request, header_parameters, stream=False, **operation_config) - if response.status_code not in [200]: - raise models.ErrorResponseException(self._deserialize, response) + if response.status_code not in [200, 304]: + exp = CloudError(response) + exp.request_id = response.headers.get('x-ms-request-id') + raise exp deserialized = None @@ -265,8 +275,7 @@ def delete( overrides`. :return: None or ClientRawResponse if raw=true :rtype: None or ~msrest.pipeline.ClientRawResponse - :raises: - :class:`ErrorResponseException` + :raises: :class:`CloudError` """ # Construct URL url = self.delete.metadata['url'] @@ -297,7 +306,9 @@ def delete( response = self._client.send(request, header_parameters, stream=False, **operation_config) if response.status_code not in [200, 204]: - raise models.ErrorResponseException(self._deserialize, response) + exp = CloudError(response) + exp.request_id = response.headers.get('x-ms-request-id') + raise exp if raw: client_raw_response = ClientRawResponse(None, response) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/factories_operations.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/factories_operations.py index 0a64dbac9083..d77c7fa31166 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/factories_operations.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/factories_operations.py @@ -11,6 +11,7 @@ import uuid from msrest.pipeline import ClientRawResponse +from msrestazure.azure_exceptions import CloudError from .. import models @@ -22,7 +23,7 @@ class FactoriesOperations(object): :param config: Configuration of service client. :param serializer: An object model serializer. :param deserializer: An object model deserializer. - :ivar api_version: The API version. Constant value: "2017-09-01-preview". + :ivar api_version: The API version. Constant value: "2018-06-01". """ models = models @@ -32,7 +33,7 @@ def __init__(self, client, config, serializer, deserializer): self._client = client self._serialize = serializer self._deserialize = deserializer - self.api_version = "2017-09-01-preview" + self.api_version = "2018-06-01" self.config = config @@ -48,8 +49,7 @@ def list( :return: An iterator like instance of Factory :rtype: ~azure.mgmt.datafactory.models.FactoryPaged[~azure.mgmt.datafactory.models.Factory] - :raises: - :class:`ErrorResponseException` + :raises: :class:`CloudError` """ def internal_paging(next_link=None, raw=False): @@ -85,7 +85,9 @@ def internal_paging(next_link=None, raw=False): request, header_parameters, stream=False, **operation_config) if response.status_code not in [200]: - raise models.ErrorResponseException(self._deserialize, response) + exp = CloudError(response) + exp.request_id = response.headers.get('x-ms-request-id') + raise exp return response @@ -101,14 +103,16 @@ def internal_paging(next_link=None, raw=False): list.metadata = {'url': '/subscriptions/{subscriptionId}/providers/Microsoft.DataFactory/factories'} def configure_factory_repo( - self, location_id, factory_repo_update, custom_headers=None, raw=False, **operation_config): + self, location_id, factory_resource_id=None, repo_configuration=None, custom_headers=None, raw=False, **operation_config): """Updates a factory's repo information. :param location_id: The location identifier. :type location_id: str - :param factory_repo_update: Update factory repo request definition. - :type factory_repo_update: - ~azure.mgmt.datafactory.models.FactoryRepoUpdate + :param factory_resource_id: The factory resource id. + :type factory_resource_id: str + :param repo_configuration: Git repo information of the factory. + :type repo_configuration: + ~azure.mgmt.datafactory.models.FactoryRepoConfiguration :param dict custom_headers: headers that will be added to the request :param bool raw: returns the direct response alongside the deserialized response @@ -117,9 +121,10 @@ def configure_factory_repo( :return: Factory or ClientRawResponse if raw=true :rtype: ~azure.mgmt.datafactory.models.Factory or ~msrest.pipeline.ClientRawResponse - :raises: - :class:`ErrorResponseException` + :raises: :class:`CloudError` """ + factory_repo_update = models.FactoryRepoUpdate(factory_resource_id=factory_resource_id, repo_configuration=repo_configuration) + # Construct URL url = self.configure_factory_repo.metadata['url'] path_format_arguments = { @@ -151,7 +156,9 @@ def configure_factory_repo( request, header_parameters, body_content, stream=False, **operation_config) if response.status_code not in [200]: - raise models.ErrorResponseException(self._deserialize, response) + exp = CloudError(response) + exp.request_id = response.headers.get('x-ms-request-id') + raise exp deserialized = None @@ -179,8 +186,7 @@ def list_by_resource_group( :return: An iterator like instance of Factory :rtype: ~azure.mgmt.datafactory.models.FactoryPaged[~azure.mgmt.datafactory.models.Factory] - :raises: - :class:`ErrorResponseException` + :raises: :class:`CloudError` """ def internal_paging(next_link=None, raw=False): @@ -217,7 +223,9 @@ def internal_paging(next_link=None, raw=False): request, header_parameters, stream=False, **operation_config) if response.status_code not in [200]: - raise models.ErrorResponseException(self._deserialize, response) + exp = CloudError(response) + exp.request_id = response.headers.get('x-ms-request-id') + raise exp return response @@ -233,7 +241,7 @@ def internal_paging(next_link=None, raw=False): list_by_resource_group.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories'} def create_or_update( - self, resource_group_name, factory_name, factory, custom_headers=None, raw=False, **operation_config): + self, resource_group_name, factory_name, factory, if_match=None, custom_headers=None, raw=False, **operation_config): """Creates or updates a factory. :param resource_group_name: The resource group name. @@ -242,6 +250,10 @@ def create_or_update( :type factory_name: str :param factory: Factory resource definition. :type factory: ~azure.mgmt.datafactory.models.Factory + :param if_match: ETag of the factory entity. Should only be specified + for update, for which it should match existing entity or can be * for + unconditional update. + :type if_match: str :param dict custom_headers: headers that will be added to the request :param bool raw: returns the direct response alongside the deserialized response @@ -250,8 +262,7 @@ def create_or_update( :return: Factory or ClientRawResponse if raw=true :rtype: ~azure.mgmt.datafactory.models.Factory or ~msrest.pipeline.ClientRawResponse - :raises: - :class:`ErrorResponseException` + :raises: :class:`CloudError` """ # Construct URL url = self.create_or_update.metadata['url'] @@ -273,6 +284,8 @@ def create_or_update( header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) if custom_headers: header_parameters.update(custom_headers) + if if_match is not None: + header_parameters['If-Match'] = self._serialize.header("if_match", if_match, 'str') if self.config.accept_language is not None: header_parameters['accept-language'] = self._serialize.header("self.config.accept_language", self.config.accept_language, 'str') @@ -285,7 +298,9 @@ def create_or_update( request, header_parameters, body_content, stream=False, **operation_config) if response.status_code not in [200]: - raise models.ErrorResponseException(self._deserialize, response) + exp = CloudError(response) + exp.request_id = response.headers.get('x-ms-request-id') + raise exp deserialized = None @@ -319,8 +334,7 @@ def update( :return: Factory or ClientRawResponse if raw=true :rtype: ~azure.mgmt.datafactory.models.Factory or ~msrest.pipeline.ClientRawResponse - :raises: - :class:`ErrorResponseException` + :raises: :class:`CloudError` """ factory_update_parameters = models.FactoryUpdateParameters(tags=tags, identity=identity) @@ -356,7 +370,9 @@ def update( request, header_parameters, body_content, stream=False, **operation_config) if response.status_code not in [200]: - raise models.ErrorResponseException(self._deserialize, response) + exp = CloudError(response) + exp.request_id = response.headers.get('x-ms-request-id') + raise exp deserialized = None @@ -371,13 +387,17 @@ def update( update.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}'} def get( - self, resource_group_name, factory_name, custom_headers=None, raw=False, **operation_config): + self, resource_group_name, factory_name, if_none_match=None, custom_headers=None, raw=False, **operation_config): """Gets a factory. :param resource_group_name: The resource group name. :type resource_group_name: str :param factory_name: The factory name. :type factory_name: str + :param if_none_match: ETag of the factory entity. Should only be + specified for get. If the ETag matches the existing entity tag, or if + * was provided, then no content will be returned. + :type if_none_match: str :param dict custom_headers: headers that will be added to the request :param bool raw: returns the direct response alongside the deserialized response @@ -386,8 +406,7 @@ def get( :return: Factory or ClientRawResponse if raw=true :rtype: ~azure.mgmt.datafactory.models.Factory or ~msrest.pipeline.ClientRawResponse - :raises: - :class:`ErrorResponseException` + :raises: :class:`CloudError` """ # Construct URL url = self.get.metadata['url'] @@ -409,6 +428,8 @@ def get( header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) if custom_headers: header_parameters.update(custom_headers) + if if_none_match is not None: + header_parameters['If-None-Match'] = self._serialize.header("if_none_match", if_none_match, 'str') if self.config.accept_language is not None: header_parameters['accept-language'] = self._serialize.header("self.config.accept_language", self.config.accept_language, 'str') @@ -416,8 +437,10 @@ def get( request = self._client.get(url, query_parameters) response = self._client.send(request, header_parameters, stream=False, **operation_config) - if response.status_code not in [200]: - raise models.ErrorResponseException(self._deserialize, response) + if response.status_code not in [200, 304]: + exp = CloudError(response) + exp.request_id = response.headers.get('x-ms-request-id') + raise exp deserialized = None @@ -446,8 +469,7 @@ def delete( overrides`. :return: None or ClientRawResponse if raw=true :rtype: None or ~msrest.pipeline.ClientRawResponse - :raises: - :class:`ErrorResponseException` + :raises: :class:`CloudError` """ # Construct URL url = self.delete.metadata['url'] @@ -477,65 +499,11 @@ def delete( response = self._client.send(request, header_parameters, stream=False, **operation_config) if response.status_code not in [200, 204]: - raise models.ErrorResponseException(self._deserialize, response) + exp = CloudError(response) + exp.request_id = response.headers.get('x-ms-request-id') + raise exp if raw: client_raw_response = ClientRawResponse(None, response) return client_raw_response delete.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}'} - - def cancel_pipeline_run( - self, resource_group_name, factory_name, run_id, custom_headers=None, raw=False, **operation_config): - """Cancel a pipeline run by its run ID. - - :param resource_group_name: The resource group name. - :type resource_group_name: str - :param factory_name: The factory name. - :type factory_name: str - :param run_id: The pipeline run identifier. - :type run_id: str - :param dict custom_headers: headers that will be added to the request - :param bool raw: returns the direct response alongside the - deserialized response - :param operation_config: :ref:`Operation configuration - overrides`. - :return: None or ClientRawResponse if raw=true - :rtype: None or ~msrest.pipeline.ClientRawResponse - :raises: - :class:`ErrorResponseException` - """ - # Construct URL - url = self.cancel_pipeline_run.metadata['url'] - path_format_arguments = { - 'subscriptionId': self._serialize.url("self.config.subscription_id", self.config.subscription_id, 'str'), - 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str', max_length=90, min_length=1, pattern=r'^[-\w\._\(\)]+$'), - 'factoryName': self._serialize.url("factory_name", factory_name, 'str', max_length=63, min_length=3, pattern=r'^[A-Za-z0-9]+(?:-[A-Za-z0-9]+)*$'), - 'runId': self._serialize.url("run_id", run_id, 'str') - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} - query_parameters['api-version'] = self._serialize.query("self.api_version", self.api_version, 'str') - - # Construct headers - header_parameters = {} - header_parameters['Content-Type'] = 'application/json; charset=utf-8' - if self.config.generate_client_request_id: - header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) - if custom_headers: - header_parameters.update(custom_headers) - if self.config.accept_language is not None: - header_parameters['accept-language'] = self._serialize.header("self.config.accept_language", self.config.accept_language, 'str') - - # Construct and send request - request = self._client.post(url, query_parameters) - response = self._client.send(request, header_parameters, stream=False, **operation_config) - - if response.status_code not in [200]: - raise models.ErrorResponseException(self._deserialize, response) - - if raw: - client_raw_response = ClientRawResponse(None, response) - return client_raw_response - cancel_pipeline_run.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/cancelpipelinerun/{runId}'} diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/integration_runtime_nodes_operations.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/integration_runtime_nodes_operations.py index d1cc12ef640d..e909517a9cbc 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/integration_runtime_nodes_operations.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/integration_runtime_nodes_operations.py @@ -11,6 +11,7 @@ import uuid from msrest.pipeline import ClientRawResponse +from msrestazure.azure_exceptions import CloudError from .. import models @@ -22,7 +23,7 @@ class IntegrationRuntimeNodesOperations(object): :param config: Configuration of service client. :param serializer: An object model serializer. :param deserializer: An object model deserializer. - :ivar api_version: The API version. Constant value: "2017-09-01-preview". + :ivar api_version: The API version. Constant value: "2018-06-01". """ models = models @@ -32,7 +33,7 @@ def __init__(self, client, config, serializer, deserializer): self._client = client self._serialize = serializer self._deserialize = deserializer - self.api_version = "2017-09-01-preview" + self.api_version = "2018-06-01" self.config = config @@ -55,8 +56,7 @@ def delete( overrides`. :return: None or ClientRawResponse if raw=true :rtype: None or ~msrest.pipeline.ClientRawResponse - :raises: - :class:`ErrorResponseException` + :raises: :class:`CloudError` """ # Construct URL url = self.delete.metadata['url'] @@ -88,7 +88,9 @@ def delete( response = self._client.send(request, header_parameters, stream=False, **operation_config) if response.status_code not in [200, 204]: - raise models.ErrorResponseException(self._deserialize, response) + exp = CloudError(response) + exp.request_id = response.headers.get('x-ms-request-id') + raise exp if raw: client_raw_response = ClientRawResponse(None, response) @@ -121,8 +123,7 @@ def update( :rtype: ~azure.mgmt.datafactory.models.SelfHostedIntegrationRuntimeNode or ~msrest.pipeline.ClientRawResponse - :raises: - :class:`ErrorResponseException` + :raises: :class:`CloudError` """ update_integration_runtime_node_request = models.UpdateIntegrationRuntimeNodeRequest(concurrent_jobs_limit=concurrent_jobs_limit) @@ -160,7 +161,9 @@ def update( request, header_parameters, body_content, stream=False, **operation_config) if response.status_code not in [200]: - raise models.ErrorResponseException(self._deserialize, response) + exp = CloudError(response) + exp.request_id = response.headers.get('x-ms-request-id') + raise exp deserialized = None @@ -195,8 +198,7 @@ def get_ip_address( raw=true :rtype: ~azure.mgmt.datafactory.models.IntegrationRuntimeNodeIpAddress or ~msrest.pipeline.ClientRawResponse - :raises: - :class:`ErrorResponseException` + :raises: :class:`CloudError` """ # Construct URL url = self.get_ip_address.metadata['url'] @@ -228,7 +230,9 @@ def get_ip_address( response = self._client.send(request, header_parameters, stream=False, **operation_config) if response.status_code not in [200]: - raise models.ErrorResponseException(self._deserialize, response) + exp = CloudError(response) + exp.request_id = response.headers.get('x-ms-request-id') + raise exp deserialized = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/integration_runtimes_operations.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/integration_runtimes_operations.py index 6790452bfc6f..6fe48b40a33c 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/integration_runtimes_operations.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/integration_runtimes_operations.py @@ -11,6 +11,7 @@ import uuid from msrest.pipeline import ClientRawResponse +from msrestazure.azure_exceptions import CloudError from msrest.polling import LROPoller, NoPolling from msrestazure.polling.arm_polling import ARMPolling @@ -24,7 +25,7 @@ class IntegrationRuntimesOperations(object): :param config: Configuration of service client. :param serializer: An object model serializer. :param deserializer: An object model deserializer. - :ivar api_version: The API version. Constant value: "2017-09-01-preview". + :ivar api_version: The API version. Constant value: "2018-06-01". """ models = models @@ -34,7 +35,7 @@ def __init__(self, client, config, serializer, deserializer): self._client = client self._serialize = serializer self._deserialize = deserializer - self.api_version = "2017-09-01-preview" + self.api_version = "2018-06-01" self.config = config @@ -54,8 +55,7 @@ def list_by_factory( :return: An iterator like instance of IntegrationRuntimeResource :rtype: ~azure.mgmt.datafactory.models.IntegrationRuntimeResourcePaged[~azure.mgmt.datafactory.models.IntegrationRuntimeResource] - :raises: - :class:`ErrorResponseException` + :raises: :class:`CloudError` """ def internal_paging(next_link=None, raw=False): @@ -93,7 +93,9 @@ def internal_paging(next_link=None, raw=False): request, header_parameters, stream=False, **operation_config) if response.status_code not in [200]: - raise models.ErrorResponseException(self._deserialize, response) + exp = CloudError(response) + exp.request_id = response.headers.get('x-ms-request-id') + raise exp return response @@ -132,8 +134,7 @@ def create_or_update( :return: IntegrationRuntimeResource or ClientRawResponse if raw=true :rtype: ~azure.mgmt.datafactory.models.IntegrationRuntimeResource or ~msrest.pipeline.ClientRawResponse - :raises: - :class:`ErrorResponseException` + :raises: :class:`CloudError` """ integration_runtime = models.IntegrationRuntimeResource(properties=properties) @@ -172,7 +173,9 @@ def create_or_update( request, header_parameters, body_content, stream=False, **operation_config) if response.status_code not in [200]: - raise models.ErrorResponseException(self._deserialize, response) + exp = CloudError(response) + exp.request_id = response.headers.get('x-ms-request-id') + raise exp deserialized = None @@ -187,7 +190,7 @@ def create_or_update( create_or_update.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/integrationRuntimes/{integrationRuntimeName}'} def get( - self, resource_group_name, factory_name, integration_runtime_name, custom_headers=None, raw=False, **operation_config): + self, resource_group_name, factory_name, integration_runtime_name, if_none_match=None, custom_headers=None, raw=False, **operation_config): """Gets an integration runtime. :param resource_group_name: The resource group name. @@ -196,6 +199,10 @@ def get( :type factory_name: str :param integration_runtime_name: The integration runtime name. :type integration_runtime_name: str + :param if_none_match: ETag of the integration runtime entity. Should + only be specified for get. If the ETag matches the existing entity + tag, or if * was provided, then no content will be returned. + :type if_none_match: str :param dict custom_headers: headers that will be added to the request :param bool raw: returns the direct response alongside the deserialized response @@ -204,8 +211,7 @@ def get( :return: IntegrationRuntimeResource or ClientRawResponse if raw=true :rtype: ~azure.mgmt.datafactory.models.IntegrationRuntimeResource or ~msrest.pipeline.ClientRawResponse - :raises: - :class:`ErrorResponseException` + :raises: :class:`CloudError` """ # Construct URL url = self.get.metadata['url'] @@ -228,6 +234,8 @@ def get( header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) if custom_headers: header_parameters.update(custom_headers) + if if_none_match is not None: + header_parameters['If-None-Match'] = self._serialize.header("if_none_match", if_none_match, 'str') if self.config.accept_language is not None: header_parameters['accept-language'] = self._serialize.header("self.config.accept_language", self.config.accept_language, 'str') @@ -235,8 +243,10 @@ def get( request = self._client.get(url, query_parameters) response = self._client.send(request, header_parameters, stream=False, **operation_config) - if response.status_code not in [200]: - raise models.ErrorResponseException(self._deserialize, response) + if response.status_code not in [200, 304]: + exp = CloudError(response) + exp.request_id = response.headers.get('x-ms-request-id') + raise exp deserialized = None @@ -275,13 +285,10 @@ def update( deserialized response :param operation_config: :ref:`Operation configuration overrides`. - :return: IntegrationRuntimeStatusResponse or ClientRawResponse if - raw=true - :rtype: - ~azure.mgmt.datafactory.models.IntegrationRuntimeStatusResponse or + :return: IntegrationRuntimeResource or ClientRawResponse if raw=true + :rtype: ~azure.mgmt.datafactory.models.IntegrationRuntimeResource or ~msrest.pipeline.ClientRawResponse - :raises: - :class:`ErrorResponseException` + :raises: :class:`CloudError` """ update_integration_runtime_request = models.UpdateIntegrationRuntimeRequest(auto_update=auto_update, update_delay_offset=update_delay_offset) @@ -318,12 +325,14 @@ def update( request, header_parameters, body_content, stream=False, **operation_config) if response.status_code not in [200]: - raise models.ErrorResponseException(self._deserialize, response) + exp = CloudError(response) + exp.request_id = response.headers.get('x-ms-request-id') + raise exp deserialized = None if response.status_code == 200: - deserialized = self._deserialize('IntegrationRuntimeStatusResponse', response) + deserialized = self._deserialize('IntegrationRuntimeResource', response) if raw: client_raw_response = ClientRawResponse(deserialized, response) @@ -349,8 +358,7 @@ def delete( overrides`. :return: None or ClientRawResponse if raw=true :rtype: None or ~msrest.pipeline.ClientRawResponse - :raises: - :class:`ErrorResponseException` + :raises: :class:`CloudError` """ # Construct URL url = self.delete.metadata['url'] @@ -381,7 +389,9 @@ def delete( response = self._client.send(request, header_parameters, stream=False, **operation_config) if response.status_code not in [200, 204]: - raise models.ErrorResponseException(self._deserialize, response) + exp = CloudError(response) + exp.request_id = response.headers.get('x-ms-request-id') + raise exp if raw: client_raw_response = ClientRawResponse(None, response) @@ -408,8 +418,7 @@ def get_status( :rtype: ~azure.mgmt.datafactory.models.IntegrationRuntimeStatusResponse or ~msrest.pipeline.ClientRawResponse - :raises: - :class:`ErrorResponseException` + :raises: :class:`CloudError` """ # Construct URL url = self.get_status.metadata['url'] @@ -440,7 +449,9 @@ def get_status( response = self._client.send(request, header_parameters, stream=False, **operation_config) if response.status_code not in [200]: - raise models.ErrorResponseException(self._deserialize, response) + exp = CloudError(response) + exp.request_id = response.headers.get('x-ms-request-id') + raise exp deserialized = None @@ -475,8 +486,7 @@ def get_connection_info( :rtype: ~azure.mgmt.datafactory.models.IntegrationRuntimeConnectionInfo or ~msrest.pipeline.ClientRawResponse - :raises: - :class:`ErrorResponseException` + :raises: :class:`CloudError` """ # Construct URL url = self.get_connection_info.metadata['url'] @@ -507,7 +517,9 @@ def get_connection_info( response = self._client.send(request, header_parameters, stream=False, **operation_config) if response.status_code not in [200]: - raise models.ErrorResponseException(self._deserialize, response) + exp = CloudError(response) + exp.request_id = response.headers.get('x-ms-request-id') + raise exp deserialized = None @@ -543,8 +555,7 @@ def regenerate_auth_key( :return: IntegrationRuntimeAuthKeys or ClientRawResponse if raw=true :rtype: ~azure.mgmt.datafactory.models.IntegrationRuntimeAuthKeys or ~msrest.pipeline.ClientRawResponse - :raises: - :class:`ErrorResponseException` + :raises: :class:`CloudError` """ regenerate_key_parameters = models.IntegrationRuntimeRegenerateKeyParameters(key_name=key_name) @@ -581,7 +592,9 @@ def regenerate_auth_key( request, header_parameters, body_content, stream=False, **operation_config) if response.status_code not in [200]: - raise models.ErrorResponseException(self._deserialize, response) + exp = CloudError(response) + exp.request_id = response.headers.get('x-ms-request-id') + raise exp deserialized = None @@ -613,8 +626,7 @@ def list_auth_keys( :return: IntegrationRuntimeAuthKeys or ClientRawResponse if raw=true :rtype: ~azure.mgmt.datafactory.models.IntegrationRuntimeAuthKeys or ~msrest.pipeline.ClientRawResponse - :raises: - :class:`ErrorResponseException` + :raises: :class:`CloudError` """ # Construct URL url = self.list_auth_keys.metadata['url'] @@ -645,7 +657,9 @@ def list_auth_keys( response = self._client.send(request, header_parameters, stream=False, **operation_config) if response.status_code not in [200]: - raise models.ErrorResponseException(self._deserialize, response) + exp = CloudError(response) + exp.request_id = response.headers.get('x-ms-request-id') + raise exp deserialized = None @@ -691,7 +705,9 @@ def _start_initial( response = self._client.send(request, header_parameters, stream=False, **operation_config) if response.status_code not in [200, 202]: - raise models.ErrorResponseException(self._deserialize, response) + exp = CloudError(response) + exp.request_id = response.headers.get('x-ms-request-id') + raise exp deserialized = None @@ -726,8 +742,7 @@ def start( ~msrestazure.azure_operation.AzureOperationPoller[~azure.mgmt.datafactory.models.IntegrationRuntimeStatusResponse] or ~msrestazure.azure_operation.AzureOperationPoller[~msrest.pipeline.ClientRawResponse[~azure.mgmt.datafactory.models.IntegrationRuntimeStatusResponse]] - :raises: - :class:`ErrorResponseException` + :raises: :class:`CloudError` """ raw_result = self._start_initial( resource_group_name=resource_group_name, @@ -788,7 +803,9 @@ def _stop_initial( response = self._client.send(request, header_parameters, stream=False, **operation_config) if response.status_code not in [200, 202]: - raise models.ErrorResponseException(self._deserialize, response) + exp = CloudError(response) + exp.request_id = response.headers.get('x-ms-request-id') + raise exp if raw: client_raw_response = ClientRawResponse(None, response) @@ -813,8 +830,7 @@ def stop( ClientRawResponse if raw==True :rtype: ~msrestazure.azure_operation.AzureOperationPoller[None] or ~msrestazure.azure_operation.AzureOperationPoller[~msrest.pipeline.ClientRawResponse[None]] - :raises: - :class:`ErrorResponseException` + :raises: :class:`CloudError` """ raw_result = self._stop_initial( resource_group_name=resource_group_name, @@ -839,73 +855,6 @@ def get_long_running_output(response): return LROPoller(self._client, raw_result, get_long_running_output, polling_method) stop.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/integrationRuntimes/{integrationRuntimeName}/stop'} - def remove_node( - self, resource_group_name, factory_name, integration_runtime_name, additional_properties=None, node_name=None, custom_headers=None, raw=False, **operation_config): - """Remove a node from integration runtime. - - :param resource_group_name: The resource group name. - :type resource_group_name: str - :param factory_name: The factory name. - :type factory_name: str - :param integration_runtime_name: The integration runtime name. - :type integration_runtime_name: str - :param additional_properties: Unmatched properties from the message - are deserialized this collection - :type additional_properties: dict[str, object] - :param node_name: The name of the node to be removed. - :type node_name: str - :param dict custom_headers: headers that will be added to the request - :param bool raw: returns the direct response alongside the - deserialized response - :param operation_config: :ref:`Operation configuration - overrides`. - :return: None or ClientRawResponse if raw=true - :rtype: None or ~msrest.pipeline.ClientRawResponse - :raises: - :class:`ErrorResponseException` - """ - remove_node_parameters = models.IntegrationRuntimeRemoveNodeRequest(additional_properties=additional_properties, node_name=node_name) - - # Construct URL - url = self.remove_node.metadata['url'] - path_format_arguments = { - 'subscriptionId': self._serialize.url("self.config.subscription_id", self.config.subscription_id, 'str'), - 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str', max_length=90, min_length=1, pattern=r'^[-\w\._\(\)]+$'), - 'factoryName': self._serialize.url("factory_name", factory_name, 'str', max_length=63, min_length=3, pattern=r'^[A-Za-z0-9]+(?:-[A-Za-z0-9]+)*$'), - 'integrationRuntimeName': self._serialize.url("integration_runtime_name", integration_runtime_name, 'str', max_length=63, min_length=3, pattern=r'^[A-Za-z0-9]+(?:-[A-Za-z0-9]+)*$') - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} - query_parameters['api-version'] = self._serialize.query("self.api_version", self.api_version, 'str') - - # Construct headers - header_parameters = {} - header_parameters['Content-Type'] = 'application/json; charset=utf-8' - if self.config.generate_client_request_id: - header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) - if custom_headers: - header_parameters.update(custom_headers) - if self.config.accept_language is not None: - header_parameters['accept-language'] = self._serialize.header("self.config.accept_language", self.config.accept_language, 'str') - - # Construct body - body_content = self._serialize.body(remove_node_parameters, 'IntegrationRuntimeRemoveNodeRequest') - - # Construct and send request - request = self._client.post(url, query_parameters) - response = self._client.send( - request, header_parameters, body_content, stream=False, **operation_config) - - if response.status_code not in [200, 204]: - raise models.ErrorResponseException(self._deserialize, response) - - if raw: - client_raw_response = ClientRawResponse(None, response) - return client_raw_response - remove_node.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/integrationRuntimes/{integrationRuntimeName}/removeNode'} - def sync_credentials( self, resource_group_name, factory_name, integration_runtime_name, custom_headers=None, raw=False, **operation_config): """Force the integration runtime to synchronize credentials across @@ -928,8 +877,7 @@ def sync_credentials( overrides`. :return: None or ClientRawResponse if raw=true :rtype: None or ~msrest.pipeline.ClientRawResponse - :raises: - :class:`ErrorResponseException` + :raises: :class:`CloudError` """ # Construct URL url = self.sync_credentials.metadata['url'] @@ -960,7 +908,9 @@ def sync_credentials( response = self._client.send(request, header_parameters, stream=False, **operation_config) if response.status_code not in [200]: - raise models.ErrorResponseException(self._deserialize, response) + exp = CloudError(response) + exp.request_id = response.headers.get('x-ms-request-id') + raise exp if raw: client_raw_response = ClientRawResponse(None, response) @@ -988,8 +938,7 @@ def get_monitoring_data( :rtype: ~azure.mgmt.datafactory.models.IntegrationRuntimeMonitoringData or ~msrest.pipeline.ClientRawResponse - :raises: - :class:`ErrorResponseException` + :raises: :class:`CloudError` """ # Construct URL url = self.get_monitoring_data.metadata['url'] @@ -1020,7 +969,9 @@ def get_monitoring_data( response = self._client.send(request, header_parameters, stream=False, **operation_config) if response.status_code not in [200]: - raise models.ErrorResponseException(self._deserialize, response) + exp = CloudError(response) + exp.request_id = response.headers.get('x-ms-request-id') + raise exp deserialized = None @@ -1051,8 +1002,7 @@ def upgrade( overrides`. :return: None or ClientRawResponse if raw=true :rtype: None or ~msrest.pipeline.ClientRawResponse - :raises: - :class:`ErrorResponseException` + :raises: :class:`CloudError` """ # Construct URL url = self.upgrade.metadata['url'] @@ -1083,7 +1033,9 @@ def upgrade( response = self._client.send(request, header_parameters, stream=False, **operation_config) if response.status_code not in [200]: - raise models.ErrorResponseException(self._deserialize, response) + exp = CloudError(response) + exp.request_id = response.headers.get('x-ms-request-id') + raise exp if raw: client_raw_response = ClientRawResponse(None, response) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/linked_services_operations.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/linked_services_operations.py index 93a9eb3dd631..98ef184787ba 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/linked_services_operations.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/linked_services_operations.py @@ -11,6 +11,7 @@ import uuid from msrest.pipeline import ClientRawResponse +from msrestazure.azure_exceptions import CloudError from .. import models @@ -22,7 +23,7 @@ class LinkedServicesOperations(object): :param config: Configuration of service client. :param serializer: An object model serializer. :param deserializer: An object model deserializer. - :ivar api_version: The API version. Constant value: "2017-09-01-preview". + :ivar api_version: The API version. Constant value: "2018-06-01". """ models = models @@ -32,7 +33,7 @@ def __init__(self, client, config, serializer, deserializer): self._client = client self._serialize = serializer self._deserialize = deserializer - self.api_version = "2017-09-01-preview" + self.api_version = "2018-06-01" self.config = config @@ -52,8 +53,7 @@ def list_by_factory( :return: An iterator like instance of LinkedServiceResource :rtype: ~azure.mgmt.datafactory.models.LinkedServiceResourcePaged[~azure.mgmt.datafactory.models.LinkedServiceResource] - :raises: - :class:`ErrorResponseException` + :raises: :class:`CloudError` """ def internal_paging(next_link=None, raw=False): @@ -91,7 +91,9 @@ def internal_paging(next_link=None, raw=False): request, header_parameters, stream=False, **operation_config) if response.status_code not in [200]: - raise models.ErrorResponseException(self._deserialize, response) + exp = CloudError(response) + exp.request_id = response.headers.get('x-ms-request-id') + raise exp return response @@ -130,8 +132,7 @@ def create_or_update( :return: LinkedServiceResource or ClientRawResponse if raw=true :rtype: ~azure.mgmt.datafactory.models.LinkedServiceResource or ~msrest.pipeline.ClientRawResponse - :raises: - :class:`ErrorResponseException` + :raises: :class:`CloudError` """ linked_service = models.LinkedServiceResource(properties=properties) @@ -170,7 +171,9 @@ def create_or_update( request, header_parameters, body_content, stream=False, **operation_config) if response.status_code not in [200]: - raise models.ErrorResponseException(self._deserialize, response) + exp = CloudError(response) + exp.request_id = response.headers.get('x-ms-request-id') + raise exp deserialized = None @@ -185,7 +188,7 @@ def create_or_update( create_or_update.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/linkedservices/{linkedServiceName}'} def get( - self, resource_group_name, factory_name, linked_service_name, custom_headers=None, raw=False, **operation_config): + self, resource_group_name, factory_name, linked_service_name, if_none_match=None, custom_headers=None, raw=False, **operation_config): """Gets a linked service. :param resource_group_name: The resource group name. @@ -194,6 +197,10 @@ def get( :type factory_name: str :param linked_service_name: The linked service name. :type linked_service_name: str + :param if_none_match: ETag of the linked service entity. Should only + be specified for get. If the ETag matches the existing entity tag, or + if * was provided, then no content will be returned. + :type if_none_match: str :param dict custom_headers: headers that will be added to the request :param bool raw: returns the direct response alongside the deserialized response @@ -202,8 +209,7 @@ def get( :return: LinkedServiceResource or ClientRawResponse if raw=true :rtype: ~azure.mgmt.datafactory.models.LinkedServiceResource or ~msrest.pipeline.ClientRawResponse - :raises: - :class:`ErrorResponseException` + :raises: :class:`CloudError` """ # Construct URL url = self.get.metadata['url'] @@ -226,6 +232,8 @@ def get( header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) if custom_headers: header_parameters.update(custom_headers) + if if_none_match is not None: + header_parameters['If-None-Match'] = self._serialize.header("if_none_match", if_none_match, 'str') if self.config.accept_language is not None: header_parameters['accept-language'] = self._serialize.header("self.config.accept_language", self.config.accept_language, 'str') @@ -233,8 +241,10 @@ def get( request = self._client.get(url, query_parameters) response = self._client.send(request, header_parameters, stream=False, **operation_config) - if response.status_code not in [200]: - raise models.ErrorResponseException(self._deserialize, response) + if response.status_code not in [200, 304]: + exp = CloudError(response) + exp.request_id = response.headers.get('x-ms-request-id') + raise exp deserialized = None @@ -265,8 +275,7 @@ def delete( overrides`. :return: None or ClientRawResponse if raw=true :rtype: None or ~msrest.pipeline.ClientRawResponse - :raises: - :class:`ErrorResponseException` + :raises: :class:`CloudError` """ # Construct URL url = self.delete.metadata['url'] @@ -297,7 +306,9 @@ def delete( response = self._client.send(request, header_parameters, stream=False, **operation_config) if response.status_code not in [200, 204]: - raise models.ErrorResponseException(self._deserialize, response) + exp = CloudError(response) + exp.request_id = response.headers.get('x-ms-request-id') + raise exp if raw: client_raw_response = ClientRawResponse(None, response) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/operations.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/operations.py index 5e3dbd283530..8af242714177 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/operations.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/operations.py @@ -11,6 +11,7 @@ import uuid from msrest.pipeline import ClientRawResponse +from msrestazure.azure_exceptions import CloudError from .. import models @@ -22,7 +23,7 @@ class Operations(object): :param config: Configuration of service client. :param serializer: An object model serializer. :param deserializer: An object model deserializer. - :ivar api_version: The API version. Constant value: "2017-09-01-preview". + :ivar api_version: The API version. Constant value: "2018-06-01". """ models = models @@ -32,7 +33,7 @@ def __init__(self, client, config, serializer, deserializer): self._client = client self._serialize = serializer self._deserialize = deserializer - self.api_version = "2017-09-01-preview" + self.api_version = "2018-06-01" self.config = config @@ -45,43 +46,53 @@ def list( deserialized response :param operation_config: :ref:`Operation configuration overrides`. - :return: OperationListResponse or ClientRawResponse if raw=true - :rtype: ~azure.mgmt.datafactory.models.OperationListResponse or - ~msrest.pipeline.ClientRawResponse - :raises: - :class:`ErrorResponseException` + :return: An iterator like instance of Operation + :rtype: + ~azure.mgmt.datafactory.models.OperationPaged[~azure.mgmt.datafactory.models.Operation] + :raises: :class:`CloudError` """ - # Construct URL - url = self.list.metadata['url'] + def internal_paging(next_link=None, raw=False): - # Construct parameters - query_parameters = {} - query_parameters['api-version'] = self._serialize.query("self.api_version", self.api_version, 'str') + if not next_link: + # Construct URL + url = self.list.metadata['url'] - # Construct headers - header_parameters = {} - header_parameters['Content-Type'] = 'application/json; charset=utf-8' - if self.config.generate_client_request_id: - header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) - if custom_headers: - header_parameters.update(custom_headers) - if self.config.accept_language is not None: - header_parameters['accept-language'] = self._serialize.header("self.config.accept_language", self.config.accept_language, 'str') + # Construct parameters + query_parameters = {} + query_parameters['api-version'] = self._serialize.query("self.api_version", self.api_version, 'str') - # Construct and send request - request = self._client.get(url, query_parameters) - response = self._client.send(request, header_parameters, stream=False, **operation_config) + else: + url = next_link + query_parameters = {} - if response.status_code not in [200]: - raise models.ErrorResponseException(self._deserialize, response) + # Construct headers + header_parameters = {} + header_parameters['Content-Type'] = 'application/json; charset=utf-8' + if self.config.generate_client_request_id: + header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) + if custom_headers: + header_parameters.update(custom_headers) + if self.config.accept_language is not None: + header_parameters['accept-language'] = self._serialize.header("self.config.accept_language", self.config.accept_language, 'str') - deserialized = None + # Construct and send request + request = self._client.get(url, query_parameters) + response = self._client.send( + request, header_parameters, stream=False, **operation_config) - if response.status_code == 200: - deserialized = self._deserialize('OperationListResponse', response) + if response.status_code not in [200]: + exp = CloudError(response) + exp.request_id = response.headers.get('x-ms-request-id') + raise exp + + return response + + # Deserialize response + deserialized = models.OperationPaged(internal_paging, self._deserialize.dependencies) if raw: - client_raw_response = ClientRawResponse(deserialized, response) + header_dict = {} + client_raw_response = models.OperationPaged(internal_paging, self._deserialize.dependencies, header_dict) return client_raw_response return deserialized diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/pipeline_runs_operations.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/pipeline_runs_operations.py index 70daad3a45f8..6fcc2316ce16 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/pipeline_runs_operations.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/pipeline_runs_operations.py @@ -11,6 +11,7 @@ import uuid from msrest.pipeline import ClientRawResponse +from msrestazure.azure_exceptions import CloudError from .. import models @@ -22,7 +23,7 @@ class PipelineRunsOperations(object): :param config: Configuration of service client. :param serializer: An object model serializer. :param deserializer: An object model deserializer. - :ivar api_version: The API version. Constant value: "2017-09-01-preview". + :ivar api_version: The API version. Constant value: "2018-06-01". """ models = models @@ -32,7 +33,7 @@ def __init__(self, client, config, serializer, deserializer): self._client = client self._serialize = serializer self._deserialize = deserializer - self.api_version = "2017-09-01-preview" + self.api_version = "2018-06-01" self.config = config @@ -46,17 +47,16 @@ def query_by_factory( :type factory_name: str :param filter_parameters: Parameters to filter the pipeline run. :type filter_parameters: - ~azure.mgmt.datafactory.models.PipelineRunFilterParameters + ~azure.mgmt.datafactory.models.RunFilterParameters :param dict custom_headers: headers that will be added to the request :param bool raw: returns the direct response alongside the deserialized response :param operation_config: :ref:`Operation configuration overrides`. - :return: PipelineRunQueryResponse or ClientRawResponse if raw=true - :rtype: ~azure.mgmt.datafactory.models.PipelineRunQueryResponse or + :return: PipelineRunsQueryResponse or ClientRawResponse if raw=true + :rtype: ~azure.mgmt.datafactory.models.PipelineRunsQueryResponse or ~msrest.pipeline.ClientRawResponse - :raises: - :class:`ErrorResponseException` + :raises: :class:`CloudError` """ # Construct URL url = self.query_by_factory.metadata['url'] @@ -82,7 +82,7 @@ def query_by_factory( header_parameters['accept-language'] = self._serialize.header("self.config.accept_language", self.config.accept_language, 'str') # Construct body - body_content = self._serialize.body(filter_parameters, 'PipelineRunFilterParameters') + body_content = self._serialize.body(filter_parameters, 'RunFilterParameters') # Construct and send request request = self._client.post(url, query_parameters) @@ -90,19 +90,21 @@ def query_by_factory( request, header_parameters, body_content, stream=False, **operation_config) if response.status_code not in [200]: - raise models.ErrorResponseException(self._deserialize, response) + exp = CloudError(response) + exp.request_id = response.headers.get('x-ms-request-id') + raise exp deserialized = None if response.status_code == 200: - deserialized = self._deserialize('PipelineRunQueryResponse', response) + deserialized = self._deserialize('PipelineRunsQueryResponse', response) if raw: client_raw_response = ClientRawResponse(deserialized, response) return client_raw_response return deserialized - query_by_factory.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/pipelineruns'} + query_by_factory.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/queryPipelineRuns'} def get( self, resource_group_name, factory_name, run_id, custom_headers=None, raw=False, **operation_config): @@ -122,8 +124,7 @@ def get( :return: PipelineRun or ClientRawResponse if raw=true :rtype: ~azure.mgmt.datafactory.models.PipelineRun or ~msrest.pipeline.ClientRawResponse - :raises: - :class:`ErrorResponseException` + :raises: :class:`CloudError` """ # Construct URL url = self.get.metadata['url'] @@ -154,7 +155,9 @@ def get( response = self._client.send(request, header_parameters, stream=False, **operation_config) if response.status_code not in [200]: - raise models.ErrorResponseException(self._deserialize, response) + exp = CloudError(response) + exp.request_id = response.headers.get('x-ms-request-id') + raise exp deserialized = None @@ -167,3 +170,60 @@ def get( return deserialized get.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/pipelineruns/{runId}'} + + def cancel( + self, resource_group_name, factory_name, run_id, custom_headers=None, raw=False, **operation_config): + """Cancel a pipeline run by its run ID. + + :param resource_group_name: The resource group name. + :type resource_group_name: str + :param factory_name: The factory name. + :type factory_name: str + :param run_id: The pipeline run identifier. + :type run_id: str + :param dict custom_headers: headers that will be added to the request + :param bool raw: returns the direct response alongside the + deserialized response + :param operation_config: :ref:`Operation configuration + overrides`. + :return: None or ClientRawResponse if raw=true + :rtype: None or ~msrest.pipeline.ClientRawResponse + :raises: :class:`CloudError` + """ + # Construct URL + url = self.cancel.metadata['url'] + path_format_arguments = { + 'subscriptionId': self._serialize.url("self.config.subscription_id", self.config.subscription_id, 'str'), + 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str', max_length=90, min_length=1, pattern=r'^[-\w\._\(\)]+$'), + 'factoryName': self._serialize.url("factory_name", factory_name, 'str', max_length=63, min_length=3, pattern=r'^[A-Za-z0-9]+(?:-[A-Za-z0-9]+)*$'), + 'runId': self._serialize.url("run_id", run_id, 'str') + } + url = self._client.format_url(url, **path_format_arguments) + + # Construct parameters + query_parameters = {} + query_parameters['api-version'] = self._serialize.query("self.api_version", self.api_version, 'str') + + # Construct headers + header_parameters = {} + header_parameters['Content-Type'] = 'application/json; charset=utf-8' + if self.config.generate_client_request_id: + header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) + if custom_headers: + header_parameters.update(custom_headers) + if self.config.accept_language is not None: + header_parameters['accept-language'] = self._serialize.header("self.config.accept_language", self.config.accept_language, 'str') + + # Construct and send request + request = self._client.post(url, query_parameters) + response = self._client.send(request, header_parameters, stream=False, **operation_config) + + if response.status_code not in [200]: + exp = CloudError(response) + exp.request_id = response.headers.get('x-ms-request-id') + raise exp + + if raw: + client_raw_response = ClientRawResponse(None, response) + return client_raw_response + cancel.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/pipelineruns/{runId}/cancel'} diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/pipelines_operations.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/pipelines_operations.py index 5dbca8f0a008..61cf8f7bcda8 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/pipelines_operations.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/pipelines_operations.py @@ -11,6 +11,7 @@ import uuid from msrest.pipeline import ClientRawResponse +from msrestazure.azure_exceptions import CloudError from .. import models @@ -22,7 +23,7 @@ class PipelinesOperations(object): :param config: Configuration of service client. :param serializer: An object model serializer. :param deserializer: An object model deserializer. - :ivar api_version: The API version. Constant value: "2017-09-01-preview". + :ivar api_version: The API version. Constant value: "2018-06-01". """ models = models @@ -32,7 +33,7 @@ def __init__(self, client, config, serializer, deserializer): self._client = client self._serialize = serializer self._deserialize = deserializer - self.api_version = "2017-09-01-preview" + self.api_version = "2018-06-01" self.config = config @@ -52,8 +53,7 @@ def list_by_factory( :return: An iterator like instance of PipelineResource :rtype: ~azure.mgmt.datafactory.models.PipelineResourcePaged[~azure.mgmt.datafactory.models.PipelineResource] - :raises: - :class:`ErrorResponseException` + :raises: :class:`CloudError` """ def internal_paging(next_link=None, raw=False): @@ -91,7 +91,9 @@ def internal_paging(next_link=None, raw=False): request, header_parameters, stream=False, **operation_config) if response.status_code not in [200]: - raise models.ErrorResponseException(self._deserialize, response) + exp = CloudError(response) + exp.request_id = response.headers.get('x-ms-request-id') + raise exp return response @@ -130,8 +132,7 @@ def create_or_update( :return: PipelineResource or ClientRawResponse if raw=true :rtype: ~azure.mgmt.datafactory.models.PipelineResource or ~msrest.pipeline.ClientRawResponse - :raises: - :class:`ErrorResponseException` + :raises: :class:`CloudError` """ # Construct URL url = self.create_or_update.metadata['url'] @@ -168,7 +169,9 @@ def create_or_update( request, header_parameters, body_content, stream=False, **operation_config) if response.status_code not in [200]: - raise models.ErrorResponseException(self._deserialize, response) + exp = CloudError(response) + exp.request_id = response.headers.get('x-ms-request-id') + raise exp deserialized = None @@ -183,7 +186,7 @@ def create_or_update( create_or_update.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/pipelines/{pipelineName}'} def get( - self, resource_group_name, factory_name, pipeline_name, custom_headers=None, raw=False, **operation_config): + self, resource_group_name, factory_name, pipeline_name, if_none_match=None, custom_headers=None, raw=False, **operation_config): """Gets a pipeline. :param resource_group_name: The resource group name. @@ -192,6 +195,10 @@ def get( :type factory_name: str :param pipeline_name: The pipeline name. :type pipeline_name: str + :param if_none_match: ETag of the pipeline entity. Should only be + specified for get. If the ETag matches the existing entity tag, or if + * was provided, then no content will be returned. + :type if_none_match: str :param dict custom_headers: headers that will be added to the request :param bool raw: returns the direct response alongside the deserialized response @@ -200,8 +207,7 @@ def get( :return: PipelineResource or ClientRawResponse if raw=true :rtype: ~azure.mgmt.datafactory.models.PipelineResource or ~msrest.pipeline.ClientRawResponse - :raises: - :class:`ErrorResponseException` + :raises: :class:`CloudError` """ # Construct URL url = self.get.metadata['url'] @@ -224,6 +230,8 @@ def get( header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) if custom_headers: header_parameters.update(custom_headers) + if if_none_match is not None: + header_parameters['If-None-Match'] = self._serialize.header("if_none_match", if_none_match, 'str') if self.config.accept_language is not None: header_parameters['accept-language'] = self._serialize.header("self.config.accept_language", self.config.accept_language, 'str') @@ -231,8 +239,10 @@ def get( request = self._client.get(url, query_parameters) response = self._client.send(request, header_parameters, stream=False, **operation_config) - if response.status_code not in [200]: - raise models.ErrorResponseException(self._deserialize, response) + if response.status_code not in [200, 304]: + exp = CloudError(response) + exp.request_id = response.headers.get('x-ms-request-id') + raise exp deserialized = None @@ -263,8 +273,7 @@ def delete( overrides`. :return: None or ClientRawResponse if raw=true :rtype: None or ~msrest.pipeline.ClientRawResponse - :raises: - :class:`ErrorResponseException` + :raises: :class:`CloudError` """ # Construct URL url = self.delete.metadata['url'] @@ -295,7 +304,9 @@ def delete( response = self._client.send(request, header_parameters, stream=False, **operation_config) if response.status_code not in [200, 204]: - raise models.ErrorResponseException(self._deserialize, response) + exp = CloudError(response) + exp.request_id = response.headers.get('x-ms-request-id') + raise exp if raw: client_raw_response = ClientRawResponse(None, response) @@ -303,7 +314,7 @@ def delete( delete.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/pipelines/{pipelineName}'} def create_run( - self, resource_group_name, factory_name, pipeline_name, parameters=None, custom_headers=None, raw=False, **operation_config): + self, resource_group_name, factory_name, pipeline_name, reference_pipeline_run_id=None, parameters=None, custom_headers=None, raw=False, **operation_config): """Creates a run of a pipeline. :param resource_group_name: The resource group name. @@ -312,7 +323,12 @@ def create_run( :type factory_name: str :param pipeline_name: The pipeline name. :type pipeline_name: str - :param parameters: Parameters of the pipeline run. + :param reference_pipeline_run_id: The pipeline run identifier. If run + ID is specified the parameters of the the specified run will be used + to create a new run. + :type reference_pipeline_run_id: str + :param parameters: Parameters of the pipeline run. These parameters + will be used only if the runId is not specified. :type parameters: dict[str, object] :param dict custom_headers: headers that will be added to the request :param bool raw: returns the direct response alongside the @@ -322,8 +338,7 @@ def create_run( :return: CreateRunResponse or ClientRawResponse if raw=true :rtype: ~azure.mgmt.datafactory.models.CreateRunResponse or ~msrest.pipeline.ClientRawResponse - :raises: - :class:`ErrorResponseException` + :raises: :class:`CloudError` """ # Construct URL url = self.create_run.metadata['url'] @@ -338,6 +353,8 @@ def create_run( # Construct parameters query_parameters = {} query_parameters['api-version'] = self._serialize.query("self.api_version", self.api_version, 'str') + if reference_pipeline_run_id is not None: + query_parameters['referencePipelineRunId'] = self._serialize.query("reference_pipeline_run_id", reference_pipeline_run_id, 'str') # Construct headers header_parameters = {} @@ -360,12 +377,14 @@ def create_run( response = self._client.send( request, header_parameters, body_content, stream=False, **operation_config) - if response.status_code not in [202]: - raise models.ErrorResponseException(self._deserialize, response) + if response.status_code not in [200]: + exp = CloudError(response) + exp.request_id = response.headers.get('x-ms-request-id') + raise exp deserialized = None - if response.status_code == 202: + if response.status_code == 200: deserialized = self._deserialize('CreateRunResponse', response) if raw: diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/trigger_runs_operations.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/trigger_runs_operations.py new file mode 100644 index 000000000000..f56548720d87 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/trigger_runs_operations.py @@ -0,0 +1,107 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +import uuid +from msrest.pipeline import ClientRawResponse +from msrestazure.azure_exceptions import CloudError + +from .. import models + + +class TriggerRunsOperations(object): + """TriggerRunsOperations operations. + + :param client: Client for service requests. + :param config: Configuration of service client. + :param serializer: An object model serializer. + :param deserializer: An object model deserializer. + :ivar api_version: The API version. Constant value: "2018-06-01". + """ + + models = models + + def __init__(self, client, config, serializer, deserializer): + + self._client = client + self._serialize = serializer + self._deserialize = deserializer + self.api_version = "2018-06-01" + + self.config = config + + def query_by_factory( + self, resource_group_name, factory_name, filter_parameters, custom_headers=None, raw=False, **operation_config): + """Query trigger runs. + + :param resource_group_name: The resource group name. + :type resource_group_name: str + :param factory_name: The factory name. + :type factory_name: str + :param filter_parameters: Parameters to filter the pipeline run. + :type filter_parameters: + ~azure.mgmt.datafactory.models.RunFilterParameters + :param dict custom_headers: headers that will be added to the request + :param bool raw: returns the direct response alongside the + deserialized response + :param operation_config: :ref:`Operation configuration + overrides`. + :return: TriggerRunsQueryResponse or ClientRawResponse if raw=true + :rtype: ~azure.mgmt.datafactory.models.TriggerRunsQueryResponse or + ~msrest.pipeline.ClientRawResponse + :raises: :class:`CloudError` + """ + # Construct URL + url = self.query_by_factory.metadata['url'] + path_format_arguments = { + 'subscriptionId': self._serialize.url("self.config.subscription_id", self.config.subscription_id, 'str'), + 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str', max_length=90, min_length=1, pattern=r'^[-\w\._\(\)]+$'), + 'factoryName': self._serialize.url("factory_name", factory_name, 'str', max_length=63, min_length=3, pattern=r'^[A-Za-z0-9]+(?:-[A-Za-z0-9]+)*$') + } + url = self._client.format_url(url, **path_format_arguments) + + # Construct parameters + query_parameters = {} + query_parameters['api-version'] = self._serialize.query("self.api_version", self.api_version, 'str') + + # Construct headers + header_parameters = {} + header_parameters['Content-Type'] = 'application/json; charset=utf-8' + if self.config.generate_client_request_id: + header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) + if custom_headers: + header_parameters.update(custom_headers) + if self.config.accept_language is not None: + header_parameters['accept-language'] = self._serialize.header("self.config.accept_language", self.config.accept_language, 'str') + + # Construct body + body_content = self._serialize.body(filter_parameters, 'RunFilterParameters') + + # Construct and send request + request = self._client.post(url, query_parameters) + response = self._client.send( + request, header_parameters, body_content, stream=False, **operation_config) + + if response.status_code not in [200]: + exp = CloudError(response) + exp.request_id = response.headers.get('x-ms-request-id') + raise exp + + deserialized = None + + if response.status_code == 200: + deserialized = self._deserialize('TriggerRunsQueryResponse', response) + + if raw: + client_raw_response = ClientRawResponse(deserialized, response) + return client_raw_response + + return deserialized + query_by_factory.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/queryTriggerRuns'} diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/triggers_operations.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/triggers_operations.py index e54de00202ea..9c8359038228 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/triggers_operations.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/triggers_operations.py @@ -11,6 +11,7 @@ import uuid from msrest.pipeline import ClientRawResponse +from msrestazure.azure_exceptions import CloudError from msrest.polling import LROPoller, NoPolling from msrestazure.polling.arm_polling import ARMPolling @@ -24,7 +25,7 @@ class TriggersOperations(object): :param config: Configuration of service client. :param serializer: An object model serializer. :param deserializer: An object model deserializer. - :ivar api_version: The API version. Constant value: "2017-09-01-preview". + :ivar api_version: The API version. Constant value: "2018-06-01". """ models = models @@ -34,7 +35,7 @@ def __init__(self, client, config, serializer, deserializer): self._client = client self._serialize = serializer self._deserialize = deserializer - self.api_version = "2017-09-01-preview" + self.api_version = "2018-06-01" self.config = config @@ -54,8 +55,7 @@ def list_by_factory( :return: An iterator like instance of TriggerResource :rtype: ~azure.mgmt.datafactory.models.TriggerResourcePaged[~azure.mgmt.datafactory.models.TriggerResource] - :raises: - :class:`ErrorResponseException` + :raises: :class:`CloudError` """ def internal_paging(next_link=None, raw=False): @@ -93,7 +93,9 @@ def internal_paging(next_link=None, raw=False): request, header_parameters, stream=False, **operation_config) if response.status_code not in [200]: - raise models.ErrorResponseException(self._deserialize, response) + exp = CloudError(response) + exp.request_id = response.headers.get('x-ms-request-id') + raise exp return response @@ -132,8 +134,7 @@ def create_or_update( :return: TriggerResource or ClientRawResponse if raw=true :rtype: ~azure.mgmt.datafactory.models.TriggerResource or ~msrest.pipeline.ClientRawResponse - :raises: - :class:`ErrorResponseException` + :raises: :class:`CloudError` """ trigger = models.TriggerResource(properties=properties) @@ -172,7 +173,9 @@ def create_or_update( request, header_parameters, body_content, stream=False, **operation_config) if response.status_code not in [200]: - raise models.ErrorResponseException(self._deserialize, response) + exp = CloudError(response) + exp.request_id = response.headers.get('x-ms-request-id') + raise exp deserialized = None @@ -187,7 +190,7 @@ def create_or_update( create_or_update.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/triggers/{triggerName}'} def get( - self, resource_group_name, factory_name, trigger_name, custom_headers=None, raw=False, **operation_config): + self, resource_group_name, factory_name, trigger_name, if_none_match=None, custom_headers=None, raw=False, **operation_config): """Gets a trigger. :param resource_group_name: The resource group name. @@ -196,6 +199,10 @@ def get( :type factory_name: str :param trigger_name: The trigger name. :type trigger_name: str + :param if_none_match: ETag of the trigger entity. Should only be + specified for get. If the ETag matches the existing entity tag, or if + * was provided, then no content will be returned. + :type if_none_match: str :param dict custom_headers: headers that will be added to the request :param bool raw: returns the direct response alongside the deserialized response @@ -204,8 +211,7 @@ def get( :return: TriggerResource or ClientRawResponse if raw=true :rtype: ~azure.mgmt.datafactory.models.TriggerResource or ~msrest.pipeline.ClientRawResponse - :raises: - :class:`ErrorResponseException` + :raises: :class:`CloudError` """ # Construct URL url = self.get.metadata['url'] @@ -228,6 +234,8 @@ def get( header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) if custom_headers: header_parameters.update(custom_headers) + if if_none_match is not None: + header_parameters['If-None-Match'] = self._serialize.header("if_none_match", if_none_match, 'str') if self.config.accept_language is not None: header_parameters['accept-language'] = self._serialize.header("self.config.accept_language", self.config.accept_language, 'str') @@ -235,8 +243,10 @@ def get( request = self._client.get(url, query_parameters) response = self._client.send(request, header_parameters, stream=False, **operation_config) - if response.status_code not in [200]: - raise models.ErrorResponseException(self._deserialize, response) + if response.status_code not in [200, 304]: + exp = CloudError(response) + exp.request_id = response.headers.get('x-ms-request-id') + raise exp deserialized = None @@ -267,8 +277,7 @@ def delete( overrides`. :return: None or ClientRawResponse if raw=true :rtype: None or ~msrest.pipeline.ClientRawResponse - :raises: - :class:`ErrorResponseException` + :raises: :class:`CloudError` """ # Construct URL url = self.delete.metadata['url'] @@ -299,7 +308,9 @@ def delete( response = self._client.send(request, header_parameters, stream=False, **operation_config) if response.status_code not in [200, 204]: - raise models.ErrorResponseException(self._deserialize, response) + exp = CloudError(response) + exp.request_id = response.headers.get('x-ms-request-id') + raise exp if raw: client_raw_response = ClientRawResponse(None, response) @@ -338,7 +349,9 @@ def _start_initial( response = self._client.send(request, header_parameters, stream=False, **operation_config) if response.status_code not in [200]: - raise models.ErrorResponseException(self._deserialize, response) + exp = CloudError(response) + exp.request_id = response.headers.get('x-ms-request-id') + raise exp if raw: client_raw_response = ClientRawResponse(None, response) @@ -363,8 +376,7 @@ def start( ClientRawResponse if raw==True :rtype: ~msrestazure.azure_operation.AzureOperationPoller[None] or ~msrestazure.azure_operation.AzureOperationPoller[~msrest.pipeline.ClientRawResponse[None]] - :raises: - :class:`ErrorResponseException` + :raises: :class:`CloudError` """ raw_result = self._start_initial( resource_group_name=resource_group_name, @@ -421,7 +433,9 @@ def _stop_initial( response = self._client.send(request, header_parameters, stream=False, **operation_config) if response.status_code not in [200]: - raise models.ErrorResponseException(self._deserialize, response) + exp = CloudError(response) + exp.request_id = response.headers.get('x-ms-request-id') + raise exp if raw: client_raw_response = ClientRawResponse(None, response) @@ -446,8 +460,7 @@ def stop( ClientRawResponse if raw==True :rtype: ~msrestazure.azure_operation.AzureOperationPoller[None] or ~msrestazure.azure_operation.AzureOperationPoller[~msrest.pipeline.ClientRawResponse[None]] - :raises: - :class:`ErrorResponseException` + :raises: :class:`CloudError` """ raw_result = self._stop_initial( resource_group_name=resource_group_name, @@ -471,82 +484,3 @@ def get_long_running_output(response): else: polling_method = polling return LROPoller(self._client, raw_result, get_long_running_output, polling_method) stop.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/triggers/{triggerName}/stop'} - - def list_runs( - self, resource_group_name, factory_name, trigger_name, start_time, end_time, custom_headers=None, raw=False, **operation_config): - """List trigger runs. - - :param resource_group_name: The resource group name. - :type resource_group_name: str - :param factory_name: The factory name. - :type factory_name: str - :param trigger_name: The trigger name. - :type trigger_name: str - :param start_time: Start time for trigger runs. - :type start_time: datetime - :param end_time: End time for trigger runs. - :type end_time: datetime - :param dict custom_headers: headers that will be added to the request - :param bool raw: returns the direct response alongside the - deserialized response - :param operation_config: :ref:`Operation configuration - overrides`. - :return: An iterator like instance of TriggerRun - :rtype: - ~azure.mgmt.datafactory.models.TriggerRunPaged[~azure.mgmt.datafactory.models.TriggerRun] - :raises: - :class:`ErrorResponseException` - """ - def internal_paging(next_link=None, raw=False): - - if not next_link: - # Construct URL - url = self.list_runs.metadata['url'] - path_format_arguments = { - 'subscriptionId': self._serialize.url("self.config.subscription_id", self.config.subscription_id, 'str'), - 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str', max_length=90, min_length=1, pattern=r'^[-\w\._\(\)]+$'), - 'factoryName': self._serialize.url("factory_name", factory_name, 'str', max_length=63, min_length=3, pattern=r'^[A-Za-z0-9]+(?:-[A-Za-z0-9]+)*$'), - 'triggerName': self._serialize.url("trigger_name", trigger_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$') - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} - query_parameters['api-version'] = self._serialize.query("self.api_version", self.api_version, 'str') - query_parameters['startTime'] = self._serialize.query("start_time", start_time, 'iso-8601') - query_parameters['endTime'] = self._serialize.query("end_time", end_time, 'iso-8601') - - else: - url = next_link - query_parameters = {} - - # Construct headers - header_parameters = {} - header_parameters['Content-Type'] = 'application/json; charset=utf-8' - if self.config.generate_client_request_id: - header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) - if custom_headers: - header_parameters.update(custom_headers) - if self.config.accept_language is not None: - header_parameters['accept-language'] = self._serialize.header("self.config.accept_language", self.config.accept_language, 'str') - - # Construct and send request - request = self._client.get(url, query_parameters) - response = self._client.send( - request, header_parameters, stream=False, **operation_config) - - if response.status_code not in [200]: - raise models.ErrorResponseException(self._deserialize, response) - - return response - - # Deserialize response - deserialized = models.TriggerRunPaged(internal_paging, self._deserialize.dependencies) - - if raw: - header_dict = {} - client_raw_response = models.TriggerRunPaged(internal_paging, self._deserialize.dependencies, header_dict) - return client_raw_response - - return deserialized - list_runs.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/triggers/{triggerName}/triggerruns'} diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/version.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/version.py index 5a7feab42d26..a39916c162ce 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/version.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/version.py @@ -9,5 +9,5 @@ # regenerated. # -------------------------------------------------------------------------- -VERSION = "0.6.0" +VERSION = "1.0.0" From e97bbe5f56f17cefcc97dd7bc86a9a3ce4b9286e Mon Sep 17 00:00:00 2001 From: Azure SDK for Python bot Date: Tue, 10 Jul 2018 06:50:39 -0700 Subject: [PATCH 09/54] [AutoPR datafactory/resource-manager] [ADF v2]Add integration runtime sharing feature. (#2865) * Generated from 8252d05908b4f2a6dc1dff659c6eb61d6d0220eb Add integration runtime sharing feature. * Autorest 3.0.56 * Generated from 10967f6b83c1ff5532b2383c8fbac24f59c9e502 Add x-ms-client-name to LinkedIntegrationRuntimeRequest.factoryName * Generated from fce7b6fcf59f03d2866abfd45b122a02bcd12afd Change class name according review comments --- .../azure/mgmt/datafactory/models/__init__.py | 15 +++++ .../data_factory_management_client_enums.py | 1 + .../models/integration_runtime_status.py | 2 +- .../models/integration_runtime_status_py3.py | 2 +- .../models/linked_integration_runtime.py | 58 ++++++++++++++++ ...d_integration_runtime_key_authorization.py | 39 +++++++++++ ...tegration_runtime_key_authorization_py3.py | 39 +++++++++++ .../models/linked_integration_runtime_py3.py | 58 ++++++++++++++++ ..._integration_runtime_rbac_authorization.py | 41 ++++++++++++ ...egration_runtime_rbac_authorization_py3.py | 41 ++++++++++++ .../linked_integration_runtime_request.py | 35 ++++++++++ .../linked_integration_runtime_request_py3.py | 35 ++++++++++ .../models/linked_integration_runtime_type.py | 42 ++++++++++++ .../linked_integration_runtime_type_py3.py | 42 ++++++++++++ .../models/managed_integration_runtime.py | 2 +- .../models/managed_integration_runtime_py3.py | 2 +- .../managed_integration_runtime_status.py | 2 +- .../managed_integration_runtime_status_py3.py | 2 +- .../models/self_hosted_integration_runtime.py | 5 ++ .../self_hosted_integration_runtime_py3.py | 7 +- .../self_hosted_integration_runtime_status.py | 13 +++- ...f_hosted_integration_runtime_status_py3.py | 15 ++++- .../integration_runtimes_operations.py | 67 +++++++++++++++++++ 23 files changed, 555 insertions(+), 10 deletions(-) create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_key_authorization.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_key_authorization_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_rbac_authorization.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_rbac_authorization_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_request.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_request_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_type.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_type_py3.py diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py index 99f5c25e46e5..e18475326303 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py @@ -28,6 +28,7 @@ from .integration_runtime_status_list_response_py3 import IntegrationRuntimeStatusListResponse from .update_integration_runtime_request_py3 import UpdateIntegrationRuntimeRequest from .update_integration_runtime_node_request_py3 import UpdateIntegrationRuntimeNodeRequest + from .linked_integration_runtime_request_py3 import LinkedIntegrationRuntimeRequest from .parameter_specification_py3 import ParameterSpecification from .linked_service_py3 import LinkedService from .linked_service_resource_py3 import LinkedServiceResource @@ -315,12 +316,16 @@ from .if_condition_activity_py3 import IfConditionActivity from .execute_pipeline_activity_py3 import ExecutePipelineActivity from .control_activity_py3 import ControlActivity + from .linked_integration_runtime_py3 import LinkedIntegrationRuntime from .self_hosted_integration_runtime_node_py3 import SelfHostedIntegrationRuntimeNode from .self_hosted_integration_runtime_status_py3 import SelfHostedIntegrationRuntimeStatus from .managed_integration_runtime_operation_result_py3 import ManagedIntegrationRuntimeOperationResult from .managed_integration_runtime_error_py3 import ManagedIntegrationRuntimeError from .managed_integration_runtime_node_py3 import ManagedIntegrationRuntimeNode from .managed_integration_runtime_status_py3 import ManagedIntegrationRuntimeStatus + from .linked_integration_runtime_rbac_authorization_py3 import LinkedIntegrationRuntimeRbacAuthorization + from .linked_integration_runtime_key_authorization_py3 import LinkedIntegrationRuntimeKeyAuthorization + from .linked_integration_runtime_type_py3 import LinkedIntegrationRuntimeType from .self_hosted_integration_runtime_py3 import SelfHostedIntegrationRuntime from .integration_runtime_custom_setup_script_properties_py3 import IntegrationRuntimeCustomSetupScriptProperties from .integration_runtime_ssis_catalog_info_py3 import IntegrationRuntimeSsisCatalogInfo @@ -353,6 +358,7 @@ from .integration_runtime_status_list_response import IntegrationRuntimeStatusListResponse from .update_integration_runtime_request import UpdateIntegrationRuntimeRequest from .update_integration_runtime_node_request import UpdateIntegrationRuntimeNodeRequest + from .linked_integration_runtime_request import LinkedIntegrationRuntimeRequest from .parameter_specification import ParameterSpecification from .linked_service import LinkedService from .linked_service_resource import LinkedServiceResource @@ -640,12 +646,16 @@ from .if_condition_activity import IfConditionActivity from .execute_pipeline_activity import ExecutePipelineActivity from .control_activity import ControlActivity + from .linked_integration_runtime import LinkedIntegrationRuntime from .self_hosted_integration_runtime_node import SelfHostedIntegrationRuntimeNode from .self_hosted_integration_runtime_status import SelfHostedIntegrationRuntimeStatus from .managed_integration_runtime_operation_result import ManagedIntegrationRuntimeOperationResult from .managed_integration_runtime_error import ManagedIntegrationRuntimeError from .managed_integration_runtime_node import ManagedIntegrationRuntimeNode from .managed_integration_runtime_status import ManagedIntegrationRuntimeStatus + from .linked_integration_runtime_rbac_authorization import LinkedIntegrationRuntimeRbacAuthorization + from .linked_integration_runtime_key_authorization import LinkedIntegrationRuntimeKeyAuthorization + from .linked_integration_runtime_type import LinkedIntegrationRuntimeType from .self_hosted_integration_runtime import SelfHostedIntegrationRuntime from .integration_runtime_custom_setup_script_properties import IntegrationRuntimeCustomSetupScriptProperties from .integration_runtime_ssis_catalog_info import IntegrationRuntimeSsisCatalogInfo @@ -746,6 +756,7 @@ 'IntegrationRuntimeStatusListResponse', 'UpdateIntegrationRuntimeRequest', 'UpdateIntegrationRuntimeNodeRequest', + 'LinkedIntegrationRuntimeRequest', 'ParameterSpecification', 'LinkedService', 'LinkedServiceResource', @@ -1033,12 +1044,16 @@ 'IfConditionActivity', 'ExecutePipelineActivity', 'ControlActivity', + 'LinkedIntegrationRuntime', 'SelfHostedIntegrationRuntimeNode', 'SelfHostedIntegrationRuntimeStatus', 'ManagedIntegrationRuntimeOperationResult', 'ManagedIntegrationRuntimeError', 'ManagedIntegrationRuntimeNode', 'ManagedIntegrationRuntimeStatus', + 'LinkedIntegrationRuntimeRbacAuthorization', + 'LinkedIntegrationRuntimeKeyAuthorization', + 'LinkedIntegrationRuntimeType', 'SelfHostedIntegrationRuntime', 'IntegrationRuntimeCustomSetupScriptProperties', 'IntegrationRuntimeSsisCatalogInfo', diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_factory_management_client_enums.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_factory_management_client_enums.py index a7eb9a135489..363e6cfb9840 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_factory_management_client_enums.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_factory_management_client_enums.py @@ -23,6 +23,7 @@ class IntegrationRuntimeState(str, Enum): online = "Online" limited = "Limited" offline = "Offline" + access_denied = "AccessDenied" class IntegrationRuntimeAutoUpdate(str, Enum): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_status.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_status.py index 987fd8c7e0c4..64da6347f9ed 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_status.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_status.py @@ -32,7 +32,7 @@ class IntegrationRuntimeStatus(Model): :vartype data_factory_name: str :ivar state: The state of integration runtime. Possible values include: 'Initial', 'Stopped', 'Started', 'Starting', 'Stopping', - 'NeedRegistration', 'Online', 'Limited', 'Offline' + 'NeedRegistration', 'Online', 'Limited', 'Offline', 'AccessDenied' :vartype state: str or ~azure.mgmt.datafactory.models.IntegrationRuntimeState :param type: Required. Constant filled by server. diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_status_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_status_py3.py index 0f41e6f73f72..8541e04dc679 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_status_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/integration_runtime_status_py3.py @@ -32,7 +32,7 @@ class IntegrationRuntimeStatus(Model): :vartype data_factory_name: str :ivar state: The state of integration runtime. Possible values include: 'Initial', 'Stopped', 'Started', 'Starting', 'Stopping', - 'NeedRegistration', 'Online', 'Limited', 'Offline' + 'NeedRegistration', 'Online', 'Limited', 'Offline', 'AccessDenied' :vartype state: str or ~azure.mgmt.datafactory.models.IntegrationRuntimeState :param type: Required. Constant filled by server. diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime.py new file mode 100644 index 000000000000..f4a4e7eb8bf0 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime.py @@ -0,0 +1,58 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class LinkedIntegrationRuntime(Model): + """The linked integration runtime information. + + Variables are only populated by the server, and will be ignored when + sending a request. + + :ivar name: The name of the linked integration runtime. + :vartype name: str + :ivar subscription_id: The subscription ID for which the linked + integration runtime belong to. + :vartype subscription_id: str + :ivar data_factory_name: The name of the data factory for which the linked + integration runtime belong to. + :vartype data_factory_name: str + :ivar data_factory_location: The location of the data factory for which + the linked integration runtime belong to. + :vartype data_factory_location: str + :ivar create_time: The creating time of the linked integration runtime. + :vartype create_time: datetime + """ + + _validation = { + 'name': {'readonly': True}, + 'subscription_id': {'readonly': True}, + 'data_factory_name': {'readonly': True}, + 'data_factory_location': {'readonly': True}, + 'create_time': {'readonly': True}, + } + + _attribute_map = { + 'name': {'key': 'name', 'type': 'str'}, + 'subscription_id': {'key': 'subscriptionId', 'type': 'str'}, + 'data_factory_name': {'key': 'dataFactoryName', 'type': 'str'}, + 'data_factory_location': {'key': 'dataFactoryLocation', 'type': 'str'}, + 'create_time': {'key': 'createTime', 'type': 'iso-8601'}, + } + + def __init__(self, **kwargs): + super(LinkedIntegrationRuntime, self).__init__(**kwargs) + self.name = None + self.subscription_id = None + self.data_factory_name = None + self.data_factory_location = None + self.create_time = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_key_authorization.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_key_authorization.py new file mode 100644 index 000000000000..b7be47e8f096 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_key_authorization.py @@ -0,0 +1,39 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_integration_runtime_type import LinkedIntegrationRuntimeType + + +class LinkedIntegrationRuntimeKeyAuthorization(LinkedIntegrationRuntimeType): + """The key authorization type integration runtime. + + All required parameters must be populated in order to send to Azure. + + :param authorization_type: Required. Constant filled by server. + :type authorization_type: str + :param key: Required. The key used for authorization. + :type key: ~azure.mgmt.datafactory.models.SecureString + """ + + _validation = { + 'authorization_type': {'required': True}, + 'key': {'required': True}, + } + + _attribute_map = { + 'authorization_type': {'key': 'authorizationType', 'type': 'str'}, + 'key': {'key': 'key', 'type': 'SecureString'}, + } + + def __init__(self, **kwargs): + super(LinkedIntegrationRuntimeKeyAuthorization, self).__init__(**kwargs) + self.key = kwargs.get('key', None) + self.authorization_type = 'Key' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_key_authorization_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_key_authorization_py3.py new file mode 100644 index 000000000000..4a2ebd8d1003 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_key_authorization_py3.py @@ -0,0 +1,39 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_integration_runtime_type_py3 import LinkedIntegrationRuntimeType + + +class LinkedIntegrationRuntimeKeyAuthorization(LinkedIntegrationRuntimeType): + """The key authorization type integration runtime. + + All required parameters must be populated in order to send to Azure. + + :param authorization_type: Required. Constant filled by server. + :type authorization_type: str + :param key: Required. The key used for authorization. + :type key: ~azure.mgmt.datafactory.models.SecureString + """ + + _validation = { + 'authorization_type': {'required': True}, + 'key': {'required': True}, + } + + _attribute_map = { + 'authorization_type': {'key': 'authorizationType', 'type': 'str'}, + 'key': {'key': 'key', 'type': 'SecureString'}, + } + + def __init__(self, *, key, **kwargs) -> None: + super(LinkedIntegrationRuntimeKeyAuthorization, self).__init__(**kwargs) + self.key = key + self.authorization_type = 'Key' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_py3.py new file mode 100644 index 000000000000..6c831ab5f511 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_py3.py @@ -0,0 +1,58 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class LinkedIntegrationRuntime(Model): + """The linked integration runtime information. + + Variables are only populated by the server, and will be ignored when + sending a request. + + :ivar name: The name of the linked integration runtime. + :vartype name: str + :ivar subscription_id: The subscription ID for which the linked + integration runtime belong to. + :vartype subscription_id: str + :ivar data_factory_name: The name of the data factory for which the linked + integration runtime belong to. + :vartype data_factory_name: str + :ivar data_factory_location: The location of the data factory for which + the linked integration runtime belong to. + :vartype data_factory_location: str + :ivar create_time: The creating time of the linked integration runtime. + :vartype create_time: datetime + """ + + _validation = { + 'name': {'readonly': True}, + 'subscription_id': {'readonly': True}, + 'data_factory_name': {'readonly': True}, + 'data_factory_location': {'readonly': True}, + 'create_time': {'readonly': True}, + } + + _attribute_map = { + 'name': {'key': 'name', 'type': 'str'}, + 'subscription_id': {'key': 'subscriptionId', 'type': 'str'}, + 'data_factory_name': {'key': 'dataFactoryName', 'type': 'str'}, + 'data_factory_location': {'key': 'dataFactoryLocation', 'type': 'str'}, + 'create_time': {'key': 'createTime', 'type': 'iso-8601'}, + } + + def __init__(self, **kwargs) -> None: + super(LinkedIntegrationRuntime, self).__init__(**kwargs) + self.name = None + self.subscription_id = None + self.data_factory_name = None + self.data_factory_location = None + self.create_time = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_rbac_authorization.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_rbac_authorization.py new file mode 100644 index 000000000000..3fbc8dd9cac2 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_rbac_authorization.py @@ -0,0 +1,41 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_integration_runtime_type import LinkedIntegrationRuntimeType + + +class LinkedIntegrationRuntimeRbacAuthorization(LinkedIntegrationRuntimeType): + """The role based access control (RBAC) authorization type integration + runtime. + + All required parameters must be populated in order to send to Azure. + + :param authorization_type: Required. Constant filled by server. + :type authorization_type: str + :param resource_id: Required. The resource identifier of the integration + runtime to be shared. + :type resource_id: str + """ + + _validation = { + 'authorization_type': {'required': True}, + 'resource_id': {'required': True}, + } + + _attribute_map = { + 'authorization_type': {'key': 'authorizationType', 'type': 'str'}, + 'resource_id': {'key': 'resourceId', 'type': 'str'}, + } + + def __init__(self, **kwargs): + super(LinkedIntegrationRuntimeRbacAuthorization, self).__init__(**kwargs) + self.resource_id = kwargs.get('resource_id', None) + self.authorization_type = 'RBAC' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_rbac_authorization_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_rbac_authorization_py3.py new file mode 100644 index 000000000000..055b64809e18 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_rbac_authorization_py3.py @@ -0,0 +1,41 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_integration_runtime_type_py3 import LinkedIntegrationRuntimeType + + +class LinkedIntegrationRuntimeRbacAuthorization(LinkedIntegrationRuntimeType): + """The role based access control (RBAC) authorization type integration + runtime. + + All required parameters must be populated in order to send to Azure. + + :param authorization_type: Required. Constant filled by server. + :type authorization_type: str + :param resource_id: Required. The resource identifier of the integration + runtime to be shared. + :type resource_id: str + """ + + _validation = { + 'authorization_type': {'required': True}, + 'resource_id': {'required': True}, + } + + _attribute_map = { + 'authorization_type': {'key': 'authorizationType', 'type': 'str'}, + 'resource_id': {'key': 'resourceId', 'type': 'str'}, + } + + def __init__(self, *, resource_id: str, **kwargs) -> None: + super(LinkedIntegrationRuntimeRbacAuthorization, self).__init__(**kwargs) + self.resource_id = resource_id + self.authorization_type = 'RBAC' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_request.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_request.py new file mode 100644 index 000000000000..807757332b3e --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_request.py @@ -0,0 +1,35 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class LinkedIntegrationRuntimeRequest(Model): + """Data factory name for linked integration runtime request. + + All required parameters must be populated in order to send to Azure. + + :param linked_factory_name: Required. The data factory name for linked + integration runtime. + :type linked_factory_name: str + """ + + _validation = { + 'linked_factory_name': {'required': True}, + } + + _attribute_map = { + 'linked_factory_name': {'key': 'factoryName', 'type': 'str'}, + } + + def __init__(self, **kwargs): + super(LinkedIntegrationRuntimeRequest, self).__init__(**kwargs) + self.linked_factory_name = kwargs.get('linked_factory_name', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_request_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_request_py3.py new file mode 100644 index 000000000000..45362ab63ba3 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_request_py3.py @@ -0,0 +1,35 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class LinkedIntegrationRuntimeRequest(Model): + """Data factory name for linked integration runtime request. + + All required parameters must be populated in order to send to Azure. + + :param linked_factory_name: Required. The data factory name for linked + integration runtime. + :type linked_factory_name: str + """ + + _validation = { + 'linked_factory_name': {'required': True}, + } + + _attribute_map = { + 'linked_factory_name': {'key': 'factoryName', 'type': 'str'}, + } + + def __init__(self, *, linked_factory_name: str, **kwargs) -> None: + super(LinkedIntegrationRuntimeRequest, self).__init__(**kwargs) + self.linked_factory_name = linked_factory_name diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_type.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_type.py new file mode 100644 index 000000000000..446395bb9cbf --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_type.py @@ -0,0 +1,42 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class LinkedIntegrationRuntimeType(Model): + """The base definition of a linked integration runtime. + + You probably want to use the sub-classes and not this class directly. Known + sub-classes are: LinkedIntegrationRuntimeRbacAuthorization, + LinkedIntegrationRuntimeKeyAuthorization + + All required parameters must be populated in order to send to Azure. + + :param authorization_type: Required. Constant filled by server. + :type authorization_type: str + """ + + _validation = { + 'authorization_type': {'required': True}, + } + + _attribute_map = { + 'authorization_type': {'key': 'authorizationType', 'type': 'str'}, + } + + _subtype_map = { + 'authorization_type': {'RBAC': 'LinkedIntegrationRuntimeRbacAuthorization', 'Key': 'LinkedIntegrationRuntimeKeyAuthorization'} + } + + def __init__(self, **kwargs): + super(LinkedIntegrationRuntimeType, self).__init__(**kwargs) + self.authorization_type = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_type_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_type_py3.py new file mode 100644 index 000000000000..79468dc450d2 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_integration_runtime_type_py3.py @@ -0,0 +1,42 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class LinkedIntegrationRuntimeType(Model): + """The base definition of a linked integration runtime. + + You probably want to use the sub-classes and not this class directly. Known + sub-classes are: LinkedIntegrationRuntimeRbacAuthorization, + LinkedIntegrationRuntimeKeyAuthorization + + All required parameters must be populated in order to send to Azure. + + :param authorization_type: Required. Constant filled by server. + :type authorization_type: str + """ + + _validation = { + 'authorization_type': {'required': True}, + } + + _attribute_map = { + 'authorization_type': {'key': 'authorizationType', 'type': 'str'}, + } + + _subtype_map = { + 'authorization_type': {'RBAC': 'LinkedIntegrationRuntimeRbacAuthorization', 'Key': 'LinkedIntegrationRuntimeKeyAuthorization'} + } + + def __init__(self, **kwargs) -> None: + super(LinkedIntegrationRuntimeType, self).__init__(**kwargs) + self.authorization_type = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/managed_integration_runtime.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/managed_integration_runtime.py index 6d2afaffd90a..9cbc9e94e7c3 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/managed_integration_runtime.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/managed_integration_runtime.py @@ -31,7 +31,7 @@ class ManagedIntegrationRuntime(IntegrationRuntime): :ivar state: Integration runtime state, only valid for managed dedicated integration runtime. Possible values include: 'Initial', 'Stopped', 'Started', 'Starting', 'Stopping', 'NeedRegistration', 'Online', - 'Limited', 'Offline' + 'Limited', 'Offline', 'AccessDenied' :vartype state: str or ~azure.mgmt.datafactory.models.IntegrationRuntimeState :param compute_properties: The compute resource for managed integration diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/managed_integration_runtime_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/managed_integration_runtime_py3.py index 741564e6e531..0e71d8b09f4e 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/managed_integration_runtime_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/managed_integration_runtime_py3.py @@ -31,7 +31,7 @@ class ManagedIntegrationRuntime(IntegrationRuntime): :ivar state: Integration runtime state, only valid for managed dedicated integration runtime. Possible values include: 'Initial', 'Stopped', 'Started', 'Starting', 'Stopping', 'NeedRegistration', 'Online', - 'Limited', 'Offline' + 'Limited', 'Offline', 'AccessDenied' :vartype state: str or ~azure.mgmt.datafactory.models.IntegrationRuntimeState :param compute_properties: The compute resource for managed integration diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/managed_integration_runtime_status.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/managed_integration_runtime_status.py index 9ebe40deae74..17d21775f09f 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/managed_integration_runtime_status.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/managed_integration_runtime_status.py @@ -28,7 +28,7 @@ class ManagedIntegrationRuntimeStatus(IntegrationRuntimeStatus): :vartype data_factory_name: str :ivar state: The state of integration runtime. Possible values include: 'Initial', 'Stopped', 'Started', 'Starting', 'Stopping', - 'NeedRegistration', 'Online', 'Limited', 'Offline' + 'NeedRegistration', 'Online', 'Limited', 'Offline', 'AccessDenied' :vartype state: str or ~azure.mgmt.datafactory.models.IntegrationRuntimeState :param type: Required. Constant filled by server. diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/managed_integration_runtime_status_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/managed_integration_runtime_status_py3.py index 873c885e7c5d..03d9451045bd 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/managed_integration_runtime_status_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/managed_integration_runtime_status_py3.py @@ -28,7 +28,7 @@ class ManagedIntegrationRuntimeStatus(IntegrationRuntimeStatus): :vartype data_factory_name: str :ivar state: The state of integration runtime. Possible values include: 'Initial', 'Stopped', 'Started', 'Starting', 'Stopping', - 'NeedRegistration', 'Online', 'Limited', 'Offline' + 'NeedRegistration', 'Online', 'Limited', 'Offline', 'AccessDenied' :vartype state: str or ~azure.mgmt.datafactory.models.IntegrationRuntimeState :param type: Required. Constant filled by server. diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime.py index b9c6c2870fff..20744f02306d 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime.py @@ -24,6 +24,9 @@ class SelfHostedIntegrationRuntime(IntegrationRuntime): :type description: str :param type: Required. Constant filled by server. :type type: str + :param linked_info: + :type linked_info: + ~azure.mgmt.datafactory.models.LinkedIntegrationRuntimeType """ _validation = { @@ -34,8 +37,10 @@ class SelfHostedIntegrationRuntime(IntegrationRuntime): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'type': {'key': 'type', 'type': 'str'}, + 'linked_info': {'key': 'typeProperties.linkedInfo', 'type': 'LinkedIntegrationRuntimeType'}, } def __init__(self, **kwargs): super(SelfHostedIntegrationRuntime, self).__init__(**kwargs) + self.linked_info = kwargs.get('linked_info', None) self.type = 'SelfHosted' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime_py3.py index 744d92d28a5a..a25d04373849 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime_py3.py @@ -24,6 +24,9 @@ class SelfHostedIntegrationRuntime(IntegrationRuntime): :type description: str :param type: Required. Constant filled by server. :type type: str + :param linked_info: + :type linked_info: + ~azure.mgmt.datafactory.models.LinkedIntegrationRuntimeType """ _validation = { @@ -34,8 +37,10 @@ class SelfHostedIntegrationRuntime(IntegrationRuntime): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'type': {'key': 'type', 'type': 'str'}, + 'linked_info': {'key': 'typeProperties.linkedInfo', 'type': 'LinkedIntegrationRuntimeType'}, } - def __init__(self, *, additional_properties=None, description: str=None, **kwargs) -> None: + def __init__(self, *, additional_properties=None, description: str=None, linked_info=None, **kwargs) -> None: super(SelfHostedIntegrationRuntime, self).__init__(additional_properties=additional_properties, description=description, **kwargs) + self.linked_info = linked_info self.type = 'SelfHosted' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime_status.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime_status.py index 3a07410bb21e..3ce6b3ef11c5 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime_status.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime_status.py @@ -28,7 +28,7 @@ class SelfHostedIntegrationRuntimeStatus(IntegrationRuntimeStatus): :vartype data_factory_name: str :ivar state: The state of integration runtime. Possible values include: 'Initial', 'Stopped', 'Started', 'Starting', 'Stopping', - 'NeedRegistration', 'Online', 'Limited', 'Offline' + 'NeedRegistration', 'Online', 'Limited', 'Offline', 'AccessDenied' :vartype state: str or ~azure.mgmt.datafactory.models.IntegrationRuntimeState :param type: Required. Constant filled by server. @@ -69,6 +69,12 @@ class SelfHostedIntegrationRuntimeStatus(IntegrationRuntimeStatus): ~azure.mgmt.datafactory.models.IntegrationRuntimeAutoUpdate :ivar version_status: Status of the integration runtime version. :vartype version_status: str + :param links: The list of linked integration runtimes that are created to + share with this integration runtime. + :type links: list[~azure.mgmt.datafactory.models.LinkedIntegrationRuntime] + :ivar shared_with_factories: The MSI-s of the data factories to which the + integration runtime is shared. + :vartype shared_with_factories: list[str] :ivar pushed_version: The version that the integration runtime is going to update to. :vartype pushed_version: str @@ -91,6 +97,7 @@ class SelfHostedIntegrationRuntimeStatus(IntegrationRuntimeStatus): 'service_urls': {'readonly': True}, 'auto_update': {'readonly': True}, 'version_status': {'readonly': True}, + 'shared_with_factories': {'readonly': True}, 'pushed_version': {'readonly': True}, 'latest_version': {'readonly': True}, } @@ -112,6 +119,8 @@ class SelfHostedIntegrationRuntimeStatus(IntegrationRuntimeStatus): 'service_urls': {'key': 'typeProperties.serviceUrls', 'type': '[str]'}, 'auto_update': {'key': 'typeProperties.autoUpdate', 'type': 'str'}, 'version_status': {'key': 'typeProperties.versionStatus', 'type': 'str'}, + 'links': {'key': 'typeProperties.links', 'type': '[LinkedIntegrationRuntime]'}, + 'shared_with_factories': {'key': 'typeProperties.sharedWithFactories', 'type': '[str]'}, 'pushed_version': {'key': 'typeProperties.pushedVersion', 'type': 'str'}, 'latest_version': {'key': 'typeProperties.latestVersion', 'type': 'str'}, } @@ -130,6 +139,8 @@ def __init__(self, **kwargs): self.service_urls = None self.auto_update = None self.version_status = None + self.links = kwargs.get('links', None) + self.shared_with_factories = None self.pushed_version = None self.latest_version = None self.type = 'SelfHosted' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime_status_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime_status_py3.py index 0714c35ca388..ae921e17d161 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime_status_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime_status_py3.py @@ -28,7 +28,7 @@ class SelfHostedIntegrationRuntimeStatus(IntegrationRuntimeStatus): :vartype data_factory_name: str :ivar state: The state of integration runtime. Possible values include: 'Initial', 'Stopped', 'Started', 'Starting', 'Stopping', - 'NeedRegistration', 'Online', 'Limited', 'Offline' + 'NeedRegistration', 'Online', 'Limited', 'Offline', 'AccessDenied' :vartype state: str or ~azure.mgmt.datafactory.models.IntegrationRuntimeState :param type: Required. Constant filled by server. @@ -69,6 +69,12 @@ class SelfHostedIntegrationRuntimeStatus(IntegrationRuntimeStatus): ~azure.mgmt.datafactory.models.IntegrationRuntimeAutoUpdate :ivar version_status: Status of the integration runtime version. :vartype version_status: str + :param links: The list of linked integration runtimes that are created to + share with this integration runtime. + :type links: list[~azure.mgmt.datafactory.models.LinkedIntegrationRuntime] + :ivar shared_with_factories: The MSI-s of the data factories to which the + integration runtime is shared. + :vartype shared_with_factories: list[str] :ivar pushed_version: The version that the integration runtime is going to update to. :vartype pushed_version: str @@ -91,6 +97,7 @@ class SelfHostedIntegrationRuntimeStatus(IntegrationRuntimeStatus): 'service_urls': {'readonly': True}, 'auto_update': {'readonly': True}, 'version_status': {'readonly': True}, + 'shared_with_factories': {'readonly': True}, 'pushed_version': {'readonly': True}, 'latest_version': {'readonly': True}, } @@ -112,11 +119,13 @@ class SelfHostedIntegrationRuntimeStatus(IntegrationRuntimeStatus): 'service_urls': {'key': 'typeProperties.serviceUrls', 'type': '[str]'}, 'auto_update': {'key': 'typeProperties.autoUpdate', 'type': 'str'}, 'version_status': {'key': 'typeProperties.versionStatus', 'type': 'str'}, + 'links': {'key': 'typeProperties.links', 'type': '[LinkedIntegrationRuntime]'}, + 'shared_with_factories': {'key': 'typeProperties.sharedWithFactories', 'type': '[str]'}, 'pushed_version': {'key': 'typeProperties.pushedVersion', 'type': 'str'}, 'latest_version': {'key': 'typeProperties.latestVersion', 'type': 'str'}, } - def __init__(self, *, additional_properties=None, nodes=None, **kwargs) -> None: + def __init__(self, *, additional_properties=None, nodes=None, links=None, **kwargs) -> None: super(SelfHostedIntegrationRuntimeStatus, self).__init__(additional_properties=additional_properties, **kwargs) self.create_time = None self.task_queue_id = None @@ -130,6 +139,8 @@ def __init__(self, *, additional_properties=None, nodes=None, **kwargs) -> None: self.service_urls = None self.auto_update = None self.version_status = None + self.links = links + self.shared_with_factories = None self.pushed_version = None self.latest_version = None self.type = 'SelfHosted' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/integration_runtimes_operations.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/integration_runtimes_operations.py index 6fe48b40a33c..1593ce82058b 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/integration_runtimes_operations.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/integration_runtimes_operations.py @@ -1041,3 +1041,70 @@ def upgrade( client_raw_response = ClientRawResponse(None, response) return client_raw_response upgrade.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/integrationRuntimes/{integrationRuntimeName}/upgrade'} + + def remove_links( + self, resource_group_name, factory_name, integration_runtime_name, linked_factory_name, custom_headers=None, raw=False, **operation_config): + """Remove all linked integration runtimes under specific data factory in a + self-hosted integration runtime. + + :param resource_group_name: The resource group name. + :type resource_group_name: str + :param factory_name: The factory name. + :type factory_name: str + :param integration_runtime_name: The integration runtime name. + :type integration_runtime_name: str + :param linked_factory_name: The data factory name for linked + integration runtime. + :type linked_factory_name: str + :param dict custom_headers: headers that will be added to the request + :param bool raw: returns the direct response alongside the + deserialized response + :param operation_config: :ref:`Operation configuration + overrides`. + :return: None or ClientRawResponse if raw=true + :rtype: None or ~msrest.pipeline.ClientRawResponse + :raises: :class:`CloudError` + """ + linked_integration_runtime_request = models.LinkedIntegrationRuntimeRequest(linked_factory_name=linked_factory_name) + + # Construct URL + url = self.remove_links.metadata['url'] + path_format_arguments = { + 'subscriptionId': self._serialize.url("self.config.subscription_id", self.config.subscription_id, 'str'), + 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str', max_length=90, min_length=1, pattern=r'^[-\w\._\(\)]+$'), + 'factoryName': self._serialize.url("factory_name", factory_name, 'str', max_length=63, min_length=3, pattern=r'^[A-Za-z0-9]+(?:-[A-Za-z0-9]+)*$'), + 'integrationRuntimeName': self._serialize.url("integration_runtime_name", integration_runtime_name, 'str', max_length=63, min_length=3, pattern=r'^[A-Za-z0-9]+(?:-[A-Za-z0-9]+)*$') + } + url = self._client.format_url(url, **path_format_arguments) + + # Construct parameters + query_parameters = {} + query_parameters['api-version'] = self._serialize.query("self.api_version", self.api_version, 'str') + + # Construct headers + header_parameters = {} + header_parameters['Content-Type'] = 'application/json; charset=utf-8' + if self.config.generate_client_request_id: + header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) + if custom_headers: + header_parameters.update(custom_headers) + if self.config.accept_language is not None: + header_parameters['accept-language'] = self._serialize.header("self.config.accept_language", self.config.accept_language, 'str') + + # Construct body + body_content = self._serialize.body(linked_integration_runtime_request, 'LinkedIntegrationRuntimeRequest') + + # Construct and send request + request = self._client.post(url, query_parameters) + response = self._client.send( + request, header_parameters, body_content, stream=False, **operation_config) + + if response.status_code not in [200]: + exp = CloudError(response) + exp.request_id = response.headers.get('x-ms-request-id') + raise exp + + if raw: + client_raw_response = ClientRawResponse(None, response) + return client_raw_response + remove_links.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/integrationRuntimes/{integrationRuntimeName}/removeLinks'} From fd03cc537526c202ca1fe081b998efcc69fff8e4 Mon Sep 17 00:00:00 2001 From: Azure SDK for Python bot Date: Wed, 11 Jul 2018 11:28:44 -0700 Subject: [PATCH 10/54] Generated from 150ad35384422f89c66b0814294f439b35e2d2ed (#2906) Remove sharedWithFactories property from SelfHostedIntegrationRuntimeStatusTypeProperties --- .../models/self_hosted_integration_runtime_status.py | 6 ------ .../models/self_hosted_integration_runtime_status_py3.py | 6 ------ 2 files changed, 12 deletions(-) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime_status.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime_status.py index 3ce6b3ef11c5..c1ad1df7deda 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime_status.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime_status.py @@ -72,9 +72,6 @@ class SelfHostedIntegrationRuntimeStatus(IntegrationRuntimeStatus): :param links: The list of linked integration runtimes that are created to share with this integration runtime. :type links: list[~azure.mgmt.datafactory.models.LinkedIntegrationRuntime] - :ivar shared_with_factories: The MSI-s of the data factories to which the - integration runtime is shared. - :vartype shared_with_factories: list[str] :ivar pushed_version: The version that the integration runtime is going to update to. :vartype pushed_version: str @@ -97,7 +94,6 @@ class SelfHostedIntegrationRuntimeStatus(IntegrationRuntimeStatus): 'service_urls': {'readonly': True}, 'auto_update': {'readonly': True}, 'version_status': {'readonly': True}, - 'shared_with_factories': {'readonly': True}, 'pushed_version': {'readonly': True}, 'latest_version': {'readonly': True}, } @@ -120,7 +116,6 @@ class SelfHostedIntegrationRuntimeStatus(IntegrationRuntimeStatus): 'auto_update': {'key': 'typeProperties.autoUpdate', 'type': 'str'}, 'version_status': {'key': 'typeProperties.versionStatus', 'type': 'str'}, 'links': {'key': 'typeProperties.links', 'type': '[LinkedIntegrationRuntime]'}, - 'shared_with_factories': {'key': 'typeProperties.sharedWithFactories', 'type': '[str]'}, 'pushed_version': {'key': 'typeProperties.pushedVersion', 'type': 'str'}, 'latest_version': {'key': 'typeProperties.latestVersion', 'type': 'str'}, } @@ -140,7 +135,6 @@ def __init__(self, **kwargs): self.auto_update = None self.version_status = None self.links = kwargs.get('links', None) - self.shared_with_factories = None self.pushed_version = None self.latest_version = None self.type = 'SelfHosted' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime_status_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime_status_py3.py index ae921e17d161..100cc77982ff 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime_status_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime_status_py3.py @@ -72,9 +72,6 @@ class SelfHostedIntegrationRuntimeStatus(IntegrationRuntimeStatus): :param links: The list of linked integration runtimes that are created to share with this integration runtime. :type links: list[~azure.mgmt.datafactory.models.LinkedIntegrationRuntime] - :ivar shared_with_factories: The MSI-s of the data factories to which the - integration runtime is shared. - :vartype shared_with_factories: list[str] :ivar pushed_version: The version that the integration runtime is going to update to. :vartype pushed_version: str @@ -97,7 +94,6 @@ class SelfHostedIntegrationRuntimeStatus(IntegrationRuntimeStatus): 'service_urls': {'readonly': True}, 'auto_update': {'readonly': True}, 'version_status': {'readonly': True}, - 'shared_with_factories': {'readonly': True}, 'pushed_version': {'readonly': True}, 'latest_version': {'readonly': True}, } @@ -120,7 +116,6 @@ class SelfHostedIntegrationRuntimeStatus(IntegrationRuntimeStatus): 'auto_update': {'key': 'typeProperties.autoUpdate', 'type': 'str'}, 'version_status': {'key': 'typeProperties.versionStatus', 'type': 'str'}, 'links': {'key': 'typeProperties.links', 'type': '[LinkedIntegrationRuntime]'}, - 'shared_with_factories': {'key': 'typeProperties.sharedWithFactories', 'type': '[str]'}, 'pushed_version': {'key': 'typeProperties.pushedVersion', 'type': 'str'}, 'latest_version': {'key': 'typeProperties.latestVersion', 'type': 'str'}, } @@ -140,7 +135,6 @@ def __init__(self, *, additional_properties=None, nodes=None, links=None, **kwar self.auto_update = None self.version_status = None self.links = links - self.shared_with_factories = None self.pushed_version = None self.latest_version = None self.type = 'SelfHosted' From 7dc28c8dc7111070eaed71b56a568e0fa7a1272c Mon Sep 17 00:00:00 2001 From: Azure SDK for Python bot Date: Wed, 11 Jul 2018 15:24:15 -0700 Subject: [PATCH 11/54] =?UTF-8?q?[AutoPR=20datafactory/resource-manager]?= =?UTF-8?q?=20[Datafactory]=20DatabricksSparkJarActivity,=20DatabricksSpar?= =?UTF-8?q?kPythonActivity=20and=20other=20D=E2=80=A6=20(#2907)?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit * Generated from b8adcbf95165499712d30b78fa15d908ebf5e356 DatabricksSparkJarActivity, DatabricksSparkPythonActivity and other Databricks enhancements * Generated from b8adcbf95165499712d30b78fa15d908ebf5e356 DatabricksSparkJarActivity, DatabricksSparkPythonActivity and other Databricks enhancements --- .../azure/mgmt/datafactory/models/__init__.py | 6 ++ .../models/azure_databricks_linked_service.py | 4 + .../azure_databricks_linked_service_py3.py | 6 +- .../models/databricks_notebook_activity.py | 5 ++ .../databricks_notebook_activity_py3.py | 7 +- .../models/databricks_spark_jar_activity.py | 75 +++++++++++++++++++ .../databricks_spark_jar_activity_py3.py | 75 +++++++++++++++++++ .../databricks_spark_python_activity.py | 75 +++++++++++++++++++ .../databricks_spark_python_activity_py3.py | 75 +++++++++++++++++++ .../datafactory/models/execution_activity.py | 5 +- .../models/execution_activity_py3.py | 5 +- 11 files changed, 332 insertions(+), 6 deletions(-) create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_spark_jar_activity.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_spark_jar_activity_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_spark_python_activity.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_spark_python_activity_py3.py diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py index e18475326303..52649cab31ec 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py @@ -212,6 +212,8 @@ from .schedule_trigger_py3 import ScheduleTrigger from .multiple_pipeline_trigger_py3 import MultiplePipelineTrigger from .activity_policy_py3 import ActivityPolicy + from .databricks_spark_python_activity_py3 import DatabricksSparkPythonActivity + from .databricks_spark_jar_activity_py3 import DatabricksSparkJarActivity from .databricks_notebook_activity_py3 import DatabricksNotebookActivity from .data_lake_analytics_usql_activity_py3 import DataLakeAnalyticsUSQLActivity from .azure_ml_update_resource_activity_py3 import AzureMLUpdateResourceActivity @@ -542,6 +544,8 @@ from .schedule_trigger import ScheduleTrigger from .multiple_pipeline_trigger import MultiplePipelineTrigger from .activity_policy import ActivityPolicy + from .databricks_spark_python_activity import DatabricksSparkPythonActivity + from .databricks_spark_jar_activity import DatabricksSparkJarActivity from .databricks_notebook_activity import DatabricksNotebookActivity from .data_lake_analytics_usql_activity import DataLakeAnalyticsUSQLActivity from .azure_ml_update_resource_activity import AzureMLUpdateResourceActivity @@ -940,6 +944,8 @@ 'ScheduleTrigger', 'MultiplePipelineTrigger', 'ActivityPolicy', + 'DatabricksSparkPythonActivity', + 'DatabricksSparkJarActivity', 'DatabricksNotebookActivity', 'DataLakeAnalyticsUSQLActivity', 'AzureMLUpdateResourceActivity', diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_databricks_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_databricks_linked_service.py index 6f683d9b6062..270c9ff302a2 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_databricks_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_databricks_linked_service.py @@ -59,6 +59,8 @@ class AzureDatabricksLinkedService(LinkedService): :param new_cluster_spark_conf: a set of optional, user-specified Spark configuration key-value pairs. :type new_cluster_spark_conf: dict[str, object] + :param new_cluster_custom_tags: Additional tags for cluster resources. + :type new_cluster_custom_tags: dict[str, object] :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -85,6 +87,7 @@ class AzureDatabricksLinkedService(LinkedService): 'new_cluster_num_of_worker': {'key': 'typeProperties.newClusterNumOfWorker', 'type': 'object'}, 'new_cluster_node_type': {'key': 'typeProperties.newClusterNodeType', 'type': 'object'}, 'new_cluster_spark_conf': {'key': 'typeProperties.newClusterSparkConf', 'type': '{object}'}, + 'new_cluster_custom_tags': {'key': 'typeProperties.newClusterCustomTags', 'type': '{object}'}, 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } @@ -97,5 +100,6 @@ def __init__(self, **kwargs): self.new_cluster_num_of_worker = kwargs.get('new_cluster_num_of_worker', None) self.new_cluster_node_type = kwargs.get('new_cluster_node_type', None) self.new_cluster_spark_conf = kwargs.get('new_cluster_spark_conf', None) + self.new_cluster_custom_tags = kwargs.get('new_cluster_custom_tags', None) self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'AzureDatabricks' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_databricks_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_databricks_linked_service_py3.py index f6113d0209b9..aa74ccc05987 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_databricks_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_databricks_linked_service_py3.py @@ -59,6 +59,8 @@ class AzureDatabricksLinkedService(LinkedService): :param new_cluster_spark_conf: a set of optional, user-specified Spark configuration key-value pairs. :type new_cluster_spark_conf: dict[str, object] + :param new_cluster_custom_tags: Additional tags for cluster resources. + :type new_cluster_custom_tags: dict[str, object] :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -85,10 +87,11 @@ class AzureDatabricksLinkedService(LinkedService): 'new_cluster_num_of_worker': {'key': 'typeProperties.newClusterNumOfWorker', 'type': 'object'}, 'new_cluster_node_type': {'key': 'typeProperties.newClusterNodeType', 'type': 'object'}, 'new_cluster_spark_conf': {'key': 'typeProperties.newClusterSparkConf', 'type': '{object}'}, + 'new_cluster_custom_tags': {'key': 'typeProperties.newClusterCustomTags', 'type': '{object}'}, 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, *, domain, access_token, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, existing_cluster_id=None, new_cluster_version=None, new_cluster_num_of_worker=None, new_cluster_node_type=None, new_cluster_spark_conf=None, encrypted_credential=None, **kwargs) -> None: + def __init__(self, *, domain, access_token, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, existing_cluster_id=None, new_cluster_version=None, new_cluster_num_of_worker=None, new_cluster_node_type=None, new_cluster_spark_conf=None, new_cluster_custom_tags=None, encrypted_credential=None, **kwargs) -> None: super(AzureDatabricksLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) self.domain = domain self.access_token = access_token @@ -97,5 +100,6 @@ def __init__(self, *, domain, access_token, additional_properties=None, connect_ self.new_cluster_num_of_worker = new_cluster_num_of_worker self.new_cluster_node_type = new_cluster_node_type self.new_cluster_spark_conf = new_cluster_spark_conf + self.new_cluster_custom_tags = new_cluster_custom_tags self.encrypted_credential = encrypted_credential self.type = 'AzureDatabricks' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_notebook_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_notebook_activity.py index fa0dfc114141..f9a337e26e33 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_notebook_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_notebook_activity.py @@ -43,6 +43,9 @@ class DatabricksNotebookActivity(ExecutionActivity): job.If the notebook takes a parameter that is not specified, the default value from the notebook will be used. :type base_parameters: dict[str, object] + :param libraries: A list of libraries to be installed on the cluster that + will execute the job. + :type libraries: list[dict[str, object]] """ _validation = { @@ -62,10 +65,12 @@ class DatabricksNotebookActivity(ExecutionActivity): 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, 'notebook_path': {'key': 'typeProperties.notebookPath', 'type': 'object'}, 'base_parameters': {'key': 'typeProperties.baseParameters', 'type': '{object}'}, + 'libraries': {'key': 'typeProperties.libraries', 'type': '[{object}]'}, } def __init__(self, **kwargs): super(DatabricksNotebookActivity, self).__init__(**kwargs) self.notebook_path = kwargs.get('notebook_path', None) self.base_parameters = kwargs.get('base_parameters', None) + self.libraries = kwargs.get('libraries', None) self.type = 'DatabricksNotebook' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_notebook_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_notebook_activity_py3.py index 2b0ed28abd87..aeb97b6e28b0 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_notebook_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_notebook_activity_py3.py @@ -43,6 +43,9 @@ class DatabricksNotebookActivity(ExecutionActivity): job.If the notebook takes a parameter that is not specified, the default value from the notebook will be used. :type base_parameters: dict[str, object] + :param libraries: A list of libraries to be installed on the cluster that + will execute the job. + :type libraries: list[dict[str, object]] """ _validation = { @@ -62,10 +65,12 @@ class DatabricksNotebookActivity(ExecutionActivity): 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, 'notebook_path': {'key': 'typeProperties.notebookPath', 'type': 'object'}, 'base_parameters': {'key': 'typeProperties.baseParameters', 'type': '{object}'}, + 'libraries': {'key': 'typeProperties.libraries', 'type': '[{object}]'}, } - def __init__(self, *, name: str, notebook_path, additional_properties=None, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, base_parameters=None, **kwargs) -> None: + def __init__(self, *, name: str, notebook_path, additional_properties=None, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, base_parameters=None, libraries=None, **kwargs) -> None: super(DatabricksNotebookActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) self.notebook_path = notebook_path self.base_parameters = base_parameters + self.libraries = libraries self.type = 'DatabricksNotebook' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_spark_jar_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_spark_jar_activity.py new file mode 100644 index 000000000000..432caf54bdc3 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_spark_jar_activity.py @@ -0,0 +1,75 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .execution_activity import ExecutionActivity + + +class DatabricksSparkJarActivity(ExecutionActivity): + """DatabricksSparkJar activity. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param name: Required. Activity name. + :type name: str + :param description: Activity description. + :type description: str + :param depends_on: Activity depends on condition. + :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param user_properties: Activity user properties. + :type user_properties: dict[str, str] + :param type: Required. Constant filled by server. + :type type: str + :param linked_service_name: Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param policy: Activity policy. + :type policy: ~azure.mgmt.datafactory.models.ActivityPolicy + :param main_class_name: Required. The full name of the class containing + the main method to be executed. This class must be contained in a JAR + provided as a library. Type: string (or Expression with resultType + string). + :type main_class_name: object + :param parameters: Parameters that will be passed to the main method. + :type parameters: list[object] + :param libraries: A list of libraries to be installed on the cluster that + will execute the job. + :type libraries: list[dict[str, object]] + """ + + _validation = { + 'name': {'required': True}, + 'type': {'required': True}, + 'main_class_name': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'name': {'key': 'name', 'type': 'str'}, + 'description': {'key': 'description', 'type': 'str'}, + 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'user_properties': {'key': 'userProperties', 'type': '{str}'}, + 'type': {'key': 'type', 'type': 'str'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, + 'main_class_name': {'key': 'typeProperties.mainClassName', 'type': 'object'}, + 'parameters': {'key': 'typeProperties.parameters', 'type': '[object]'}, + 'libraries': {'key': 'typeProperties.libraries', 'type': '[{object}]'}, + } + + def __init__(self, **kwargs): + super(DatabricksSparkJarActivity, self).__init__(**kwargs) + self.main_class_name = kwargs.get('main_class_name', None) + self.parameters = kwargs.get('parameters', None) + self.libraries = kwargs.get('libraries', None) + self.type = 'DatabricksSparkJar' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_spark_jar_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_spark_jar_activity_py3.py new file mode 100644 index 000000000000..99224eade24f --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_spark_jar_activity_py3.py @@ -0,0 +1,75 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .execution_activity_py3 import ExecutionActivity + + +class DatabricksSparkJarActivity(ExecutionActivity): + """DatabricksSparkJar activity. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param name: Required. Activity name. + :type name: str + :param description: Activity description. + :type description: str + :param depends_on: Activity depends on condition. + :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param user_properties: Activity user properties. + :type user_properties: dict[str, str] + :param type: Required. Constant filled by server. + :type type: str + :param linked_service_name: Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param policy: Activity policy. + :type policy: ~azure.mgmt.datafactory.models.ActivityPolicy + :param main_class_name: Required. The full name of the class containing + the main method to be executed. This class must be contained in a JAR + provided as a library. Type: string (or Expression with resultType + string). + :type main_class_name: object + :param parameters: Parameters that will be passed to the main method. + :type parameters: list[object] + :param libraries: A list of libraries to be installed on the cluster that + will execute the job. + :type libraries: list[dict[str, object]] + """ + + _validation = { + 'name': {'required': True}, + 'type': {'required': True}, + 'main_class_name': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'name': {'key': 'name', 'type': 'str'}, + 'description': {'key': 'description', 'type': 'str'}, + 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'user_properties': {'key': 'userProperties', 'type': '{str}'}, + 'type': {'key': 'type', 'type': 'str'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, + 'main_class_name': {'key': 'typeProperties.mainClassName', 'type': 'object'}, + 'parameters': {'key': 'typeProperties.parameters', 'type': '[object]'}, + 'libraries': {'key': 'typeProperties.libraries', 'type': '[{object}]'}, + } + + def __init__(self, *, name: str, main_class_name, additional_properties=None, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, parameters=None, libraries=None, **kwargs) -> None: + super(DatabricksSparkJarActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) + self.main_class_name = main_class_name + self.parameters = parameters + self.libraries = libraries + self.type = 'DatabricksSparkJar' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_spark_python_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_spark_python_activity.py new file mode 100644 index 000000000000..76a949177514 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_spark_python_activity.py @@ -0,0 +1,75 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .execution_activity import ExecutionActivity + + +class DatabricksSparkPythonActivity(ExecutionActivity): + """DatabricksSparkPython activity. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param name: Required. Activity name. + :type name: str + :param description: Activity description. + :type description: str + :param depends_on: Activity depends on condition. + :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param user_properties: Activity user properties. + :type user_properties: dict[str, str] + :param type: Required. Constant filled by server. + :type type: str + :param linked_service_name: Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param policy: Activity policy. + :type policy: ~azure.mgmt.datafactory.models.ActivityPolicy + :param python_file: Required. The URI of the Python file to be executed. + DBFS paths are supported. Type: string (or Expression with resultType + string). + :type python_file: object + :param parameters: Command line parameters that will be passed to the + Python file. + :type parameters: list[object] + :param libraries: A list of libraries to be installed on the cluster that + will execute the job. + :type libraries: list[dict[str, object]] + """ + + _validation = { + 'name': {'required': True}, + 'type': {'required': True}, + 'python_file': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'name': {'key': 'name', 'type': 'str'}, + 'description': {'key': 'description', 'type': 'str'}, + 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'user_properties': {'key': 'userProperties', 'type': '{str}'}, + 'type': {'key': 'type', 'type': 'str'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, + 'python_file': {'key': 'typeProperties.pythonFile', 'type': 'object'}, + 'parameters': {'key': 'typeProperties.parameters', 'type': '[object]'}, + 'libraries': {'key': 'typeProperties.libraries', 'type': '[{object}]'}, + } + + def __init__(self, **kwargs): + super(DatabricksSparkPythonActivity, self).__init__(**kwargs) + self.python_file = kwargs.get('python_file', None) + self.parameters = kwargs.get('parameters', None) + self.libraries = kwargs.get('libraries', None) + self.type = 'DatabricksSparkPython' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_spark_python_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_spark_python_activity_py3.py new file mode 100644 index 000000000000..26039a6958c4 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_spark_python_activity_py3.py @@ -0,0 +1,75 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .execution_activity_py3 import ExecutionActivity + + +class DatabricksSparkPythonActivity(ExecutionActivity): + """DatabricksSparkPython activity. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param name: Required. Activity name. + :type name: str + :param description: Activity description. + :type description: str + :param depends_on: Activity depends on condition. + :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param user_properties: Activity user properties. + :type user_properties: dict[str, str] + :param type: Required. Constant filled by server. + :type type: str + :param linked_service_name: Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param policy: Activity policy. + :type policy: ~azure.mgmt.datafactory.models.ActivityPolicy + :param python_file: Required. The URI of the Python file to be executed. + DBFS paths are supported. Type: string (or Expression with resultType + string). + :type python_file: object + :param parameters: Command line parameters that will be passed to the + Python file. + :type parameters: list[object] + :param libraries: A list of libraries to be installed on the cluster that + will execute the job. + :type libraries: list[dict[str, object]] + """ + + _validation = { + 'name': {'required': True}, + 'type': {'required': True}, + 'python_file': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'name': {'key': 'name', 'type': 'str'}, + 'description': {'key': 'description', 'type': 'str'}, + 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'user_properties': {'key': 'userProperties', 'type': '{str}'}, + 'type': {'key': 'type', 'type': 'str'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, + 'python_file': {'key': 'typeProperties.pythonFile', 'type': 'object'}, + 'parameters': {'key': 'typeProperties.parameters', 'type': '[object]'}, + 'libraries': {'key': 'typeProperties.libraries', 'type': '[{object}]'}, + } + + def __init__(self, *, name: str, python_file, additional_properties=None, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, parameters=None, libraries=None, **kwargs) -> None: + super(DatabricksSparkPythonActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) + self.python_file = python_file + self.parameters = parameters + self.libraries = libraries + self.type = 'DatabricksSparkPython' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execution_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execution_activity.py index 8e88fec0e8bd..f8c64232576b 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execution_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execution_activity.py @@ -16,7 +16,8 @@ class ExecutionActivity(Activity): """Base class for all execution activities. You probably want to use the sub-classes and not this class directly. Known - sub-classes are: DatabricksNotebookActivity, DataLakeAnalyticsUSQLActivity, + sub-classes are: DatabricksSparkPythonActivity, DatabricksSparkJarActivity, + DatabricksNotebookActivity, DataLakeAnalyticsUSQLActivity, AzureMLUpdateResourceActivity, AzureMLBatchExecutionActivity, GetMetadataActivity, WebActivity, LookupActivity, SqlServerStoredProcedureActivity, CustomActivity, @@ -63,7 +64,7 @@ class ExecutionActivity(Activity): } _subtype_map = { - 'type': {'DatabricksNotebook': 'DatabricksNotebookActivity', 'DataLakeAnalyticsU-SQL': 'DataLakeAnalyticsUSQLActivity', 'AzureMLUpdateResource': 'AzureMLUpdateResourceActivity', 'AzureMLBatchExecution': 'AzureMLBatchExecutionActivity', 'GetMetadata': 'GetMetadataActivity', 'WebActivity': 'WebActivity', 'Lookup': 'LookupActivity', 'SqlServerStoredProcedure': 'SqlServerStoredProcedureActivity', 'Custom': 'CustomActivity', 'ExecuteSSISPackage': 'ExecuteSSISPackageActivity', 'HDInsightSpark': 'HDInsightSparkActivity', 'HDInsightStreaming': 'HDInsightStreamingActivity', 'HDInsightMapReduce': 'HDInsightMapReduceActivity', 'HDInsightPig': 'HDInsightPigActivity', 'HDInsightHive': 'HDInsightHiveActivity', 'Copy': 'CopyActivity'} + 'type': {'DatabricksSparkPython': 'DatabricksSparkPythonActivity', 'DatabricksSparkJar': 'DatabricksSparkJarActivity', 'DatabricksNotebook': 'DatabricksNotebookActivity', 'DataLakeAnalyticsU-SQL': 'DataLakeAnalyticsUSQLActivity', 'AzureMLUpdateResource': 'AzureMLUpdateResourceActivity', 'AzureMLBatchExecution': 'AzureMLBatchExecutionActivity', 'GetMetadata': 'GetMetadataActivity', 'WebActivity': 'WebActivity', 'Lookup': 'LookupActivity', 'SqlServerStoredProcedure': 'SqlServerStoredProcedureActivity', 'Custom': 'CustomActivity', 'ExecuteSSISPackage': 'ExecuteSSISPackageActivity', 'HDInsightSpark': 'HDInsightSparkActivity', 'HDInsightStreaming': 'HDInsightStreamingActivity', 'HDInsightMapReduce': 'HDInsightMapReduceActivity', 'HDInsightPig': 'HDInsightPigActivity', 'HDInsightHive': 'HDInsightHiveActivity', 'Copy': 'CopyActivity'} } def __init__(self, **kwargs): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execution_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execution_activity_py3.py index d7290f78706a..e54a7b1fea57 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execution_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execution_activity_py3.py @@ -16,7 +16,8 @@ class ExecutionActivity(Activity): """Base class for all execution activities. You probably want to use the sub-classes and not this class directly. Known - sub-classes are: DatabricksNotebookActivity, DataLakeAnalyticsUSQLActivity, + sub-classes are: DatabricksSparkPythonActivity, DatabricksSparkJarActivity, + DatabricksNotebookActivity, DataLakeAnalyticsUSQLActivity, AzureMLUpdateResourceActivity, AzureMLBatchExecutionActivity, GetMetadataActivity, WebActivity, LookupActivity, SqlServerStoredProcedureActivity, CustomActivity, @@ -63,7 +64,7 @@ class ExecutionActivity(Activity): } _subtype_map = { - 'type': {'DatabricksNotebook': 'DatabricksNotebookActivity', 'DataLakeAnalyticsU-SQL': 'DataLakeAnalyticsUSQLActivity', 'AzureMLUpdateResource': 'AzureMLUpdateResourceActivity', 'AzureMLBatchExecution': 'AzureMLBatchExecutionActivity', 'GetMetadata': 'GetMetadataActivity', 'WebActivity': 'WebActivity', 'Lookup': 'LookupActivity', 'SqlServerStoredProcedure': 'SqlServerStoredProcedureActivity', 'Custom': 'CustomActivity', 'ExecuteSSISPackage': 'ExecuteSSISPackageActivity', 'HDInsightSpark': 'HDInsightSparkActivity', 'HDInsightStreaming': 'HDInsightStreamingActivity', 'HDInsightMapReduce': 'HDInsightMapReduceActivity', 'HDInsightPig': 'HDInsightPigActivity', 'HDInsightHive': 'HDInsightHiveActivity', 'Copy': 'CopyActivity'} + 'type': {'DatabricksSparkPython': 'DatabricksSparkPythonActivity', 'DatabricksSparkJar': 'DatabricksSparkJarActivity', 'DatabricksNotebook': 'DatabricksNotebookActivity', 'DataLakeAnalyticsU-SQL': 'DataLakeAnalyticsUSQLActivity', 'AzureMLUpdateResource': 'AzureMLUpdateResourceActivity', 'AzureMLBatchExecution': 'AzureMLBatchExecutionActivity', 'GetMetadata': 'GetMetadataActivity', 'WebActivity': 'WebActivity', 'Lookup': 'LookupActivity', 'SqlServerStoredProcedure': 'SqlServerStoredProcedureActivity', 'Custom': 'CustomActivity', 'ExecuteSSISPackage': 'ExecuteSSISPackageActivity', 'HDInsightSpark': 'HDInsightSparkActivity', 'HDInsightStreaming': 'HDInsightStreamingActivity', 'HDInsightMapReduce': 'HDInsightMapReduceActivity', 'HDInsightPig': 'HDInsightPigActivity', 'HDInsightHive': 'HDInsightHiveActivity', 'Copy': 'CopyActivity'} } def __init__(self, *, name: str, additional_properties=None, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, **kwargs) -> None: From 5f8603db67c7b2635956b08af1da65840aebbab8 Mon Sep 17 00:00:00 2001 From: Azure SDK for Python bot Date: Tue, 24 Jul 2018 11:23:48 -0700 Subject: [PATCH 12/54] Generated from 3bcb5068dd38686779fe7f7aee5b11b6da83f680 (#2988) Updating User Properties --- .../azure/mgmt/datafactory/models/__init__.py | 3 ++ .../azure/mgmt/datafactory/models/activity.py | 9 +---- .../mgmt/datafactory/models/activity_py3.py | 11 ++--- .../azure_ml_batch_execution_activity.py | 8 +--- .../azure_ml_batch_execution_activity_py3.py | 12 ++---- .../azure_ml_update_resource_activity.py | 8 +--- .../azure_ml_update_resource_activity_py3.py | 12 ++---- .../datafactory/models/control_activity.py | 8 +--- .../models/control_activity_py3.py | 12 ++---- .../mgmt/datafactory/models/copy_activity.py | 8 +--- .../datafactory/models/copy_activity_py3.py | 12 ++---- .../datafactory/models/custom_activity.py | 8 +--- .../datafactory/models/custom_activity_py3.py | 12 ++---- .../data_lake_analytics_usql_activity.py | 8 +--- .../data_lake_analytics_usql_activity_py3.py | 12 ++---- .../models/databricks_notebook_activity.py | 8 +--- .../databricks_notebook_activity_py3.py | 12 ++---- .../models/databricks_spark_jar_activity.py | 8 +--- .../databricks_spark_jar_activity_py3.py | 12 ++---- .../databricks_spark_python_activity.py | 8 +--- .../databricks_spark_python_activity_py3.py | 12 ++---- .../models/execute_pipeline_activity.py | 8 +--- .../models/execute_pipeline_activity_py3.py | 12 ++---- .../models/execute_ssis_package_activity.py | 8 +--- .../execute_ssis_package_activity_py3.py | 12 ++---- .../datafactory/models/execution_activity.py | 8 +--- .../models/execution_activity_py3.py | 12 ++---- .../datafactory/models/filter_activity.py | 8 +--- .../datafactory/models/filter_activity_py3.py | 12 ++---- .../datafactory/models/for_each_activity.py | 8 +--- .../models/for_each_activity_py3.py | 12 ++---- .../models/get_metadata_activity.py | 8 +--- .../models/get_metadata_activity_py3.py | 12 ++---- .../models/hd_insight_hive_activity.py | 8 +--- .../models/hd_insight_hive_activity_py3.py | 12 ++---- .../models/hd_insight_map_reduce_activity.py | 8 +--- .../hd_insight_map_reduce_activity_py3.py | 12 ++---- .../models/hd_insight_pig_activity.py | 8 +--- .../models/hd_insight_pig_activity_py3.py | 12 ++---- .../models/hd_insight_spark_activity.py | 8 +--- .../models/hd_insight_spark_activity_py3.py | 12 ++---- .../models/hd_insight_streaming_activity.py | 8 +--- .../hd_insight_streaming_activity_py3.py | 12 ++---- .../models/if_condition_activity.py | 8 +--- .../models/if_condition_activity_py3.py | 12 ++---- .../datafactory/models/lookup_activity.py | 8 +--- .../datafactory/models/lookup_activity_py3.py | 12 ++---- .../sql_server_stored_procedure_activity.py | 8 +--- ...ql_server_stored_procedure_activity_py3.py | 12 ++---- .../mgmt/datafactory/models/until_activity.py | 8 +--- .../datafactory/models/until_activity_py3.py | 12 ++---- .../mgmt/datafactory/models/user_property.py | 40 +++++++++++++++++++ .../datafactory/models/user_property_py3.py | 40 +++++++++++++++++++ .../mgmt/datafactory/models/wait_activity.py | 8 +--- .../datafactory/models/wait_activity_py3.py | 12 ++---- .../mgmt/datafactory/models/web_activity.py | 8 +--- .../datafactory/models/web_activity_py3.py | 12 ++---- 57 files changed, 244 insertions(+), 379 deletions(-) create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/user_property.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/user_property_py3.py diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py index 52649cab31ec..2397ab63e4c5 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py @@ -35,6 +35,7 @@ from .dataset_py3 import Dataset from .dataset_resource_py3 import DatasetResource from .activity_dependency_py3 import ActivityDependency + from .user_property_py3 import UserProperty from .activity_py3 import Activity from .pipeline_resource_py3 import PipelineResource from .trigger_py3 import Trigger @@ -367,6 +368,7 @@ from .dataset import Dataset from .dataset_resource import DatasetResource from .activity_dependency import ActivityDependency + from .user_property import UserProperty from .activity import Activity from .pipeline_resource import PipelineResource from .trigger import Trigger @@ -767,6 +769,7 @@ 'Dataset', 'DatasetResource', 'ActivityDependency', + 'UserProperty', 'Activity', 'PipelineResource', 'Trigger', diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/activity.py index 438e4b23f7eb..06dca4b71127 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/activity.py @@ -20,9 +20,6 @@ class Activity(Model): All required parameters must be populated in order to send to Azure. - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -30,7 +27,7 @@ class Activity(Model): :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: dict[str, str] + :type user_properties: list[~azure.mgmt.datafactory.models.UserProperty] :param type: Required. Constant filled by server. :type type: str """ @@ -41,11 +38,10 @@ class Activity(Model): } _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '{str}'}, + 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, 'type': {'key': 'type', 'type': 'str'}, } @@ -55,7 +51,6 @@ class Activity(Model): def __init__(self, **kwargs): super(Activity, self).__init__(**kwargs) - self.additional_properties = kwargs.get('additional_properties', None) self.name = kwargs.get('name', None) self.description = kwargs.get('description', None) self.depends_on = kwargs.get('depends_on', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/activity_py3.py index 48d13e8e631d..a36cee0fc595 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/activity_py3.py @@ -20,9 +20,6 @@ class Activity(Model): All required parameters must be populated in order to send to Azure. - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -30,7 +27,7 @@ class Activity(Model): :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: dict[str, str] + :type user_properties: list[~azure.mgmt.datafactory.models.UserProperty] :param type: Required. Constant filled by server. :type type: str """ @@ -41,11 +38,10 @@ class Activity(Model): } _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '{str}'}, + 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, 'type': {'key': 'type', 'type': 'str'}, } @@ -53,9 +49,8 @@ class Activity(Model): 'type': {'Execution': 'ExecutionActivity', 'Container': 'ControlActivity'} } - def __init__(self, *, name: str, additional_properties=None, description: str=None, depends_on=None, user_properties=None, **kwargs) -> None: + def __init__(self, *, name: str, description: str=None, depends_on=None, user_properties=None, **kwargs) -> None: super(Activity, self).__init__(**kwargs) - self.additional_properties = additional_properties self.name = name self.description = description self.depends_on = depends_on diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_batch_execution_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_batch_execution_activity.py index e9c935e65205..47a7a6dbd41e 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_batch_execution_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_batch_execution_activity.py @@ -17,9 +17,6 @@ class AzureMLBatchExecutionActivity(ExecutionActivity): All required parameters must be populated in order to send to Azure. - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -27,7 +24,7 @@ class AzureMLBatchExecutionActivity(ExecutionActivity): :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: dict[str, str] + :type user_properties: list[~azure.mgmt.datafactory.models.UserProperty] :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. @@ -61,11 +58,10 @@ class AzureMLBatchExecutionActivity(ExecutionActivity): } _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '{str}'}, + 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, 'type': {'key': 'type', 'type': 'str'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_batch_execution_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_batch_execution_activity_py3.py index e455622c0a89..eaac413d277e 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_batch_execution_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_batch_execution_activity_py3.py @@ -17,9 +17,6 @@ class AzureMLBatchExecutionActivity(ExecutionActivity): All required parameters must be populated in order to send to Azure. - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -27,7 +24,7 @@ class AzureMLBatchExecutionActivity(ExecutionActivity): :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: dict[str, str] + :type user_properties: list[~azure.mgmt.datafactory.models.UserProperty] :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. @@ -61,11 +58,10 @@ class AzureMLBatchExecutionActivity(ExecutionActivity): } _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '{str}'}, + 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, 'type': {'key': 'type', 'type': 'str'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, @@ -74,8 +70,8 @@ class AzureMLBatchExecutionActivity(ExecutionActivity): 'web_service_inputs': {'key': 'typeProperties.webServiceInputs', 'type': '{AzureMLWebServiceFile}'}, } - def __init__(self, *, name: str, additional_properties=None, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, global_parameters=None, web_service_outputs=None, web_service_inputs=None, **kwargs) -> None: - super(AzureMLBatchExecutionActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) + def __init__(self, *, name: str, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, global_parameters=None, web_service_outputs=None, web_service_inputs=None, **kwargs) -> None: + super(AzureMLBatchExecutionActivity, self).__init__(name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) self.global_parameters = global_parameters self.web_service_outputs = web_service_outputs self.web_service_inputs = web_service_inputs diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_update_resource_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_update_resource_activity.py index 6c809dd27d4d..ad5911008bf2 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_update_resource_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_update_resource_activity.py @@ -17,9 +17,6 @@ class AzureMLUpdateResourceActivity(ExecutionActivity): All required parameters must be populated in order to send to Azure. - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -27,7 +24,7 @@ class AzureMLUpdateResourceActivity(ExecutionActivity): :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: dict[str, str] + :type user_properties: list[~azure.mgmt.datafactory.models.UserProperty] :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. @@ -60,11 +57,10 @@ class AzureMLUpdateResourceActivity(ExecutionActivity): } _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '{str}'}, + 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, 'type': {'key': 'type', 'type': 'str'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_update_resource_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_update_resource_activity_py3.py index 13cbc43d5de5..68b721fc94f1 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_update_resource_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_update_resource_activity_py3.py @@ -17,9 +17,6 @@ class AzureMLUpdateResourceActivity(ExecutionActivity): All required parameters must be populated in order to send to Azure. - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -27,7 +24,7 @@ class AzureMLUpdateResourceActivity(ExecutionActivity): :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: dict[str, str] + :type user_properties: list[~azure.mgmt.datafactory.models.UserProperty] :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. @@ -60,11 +57,10 @@ class AzureMLUpdateResourceActivity(ExecutionActivity): } _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '{str}'}, + 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, 'type': {'key': 'type', 'type': 'str'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, @@ -73,8 +69,8 @@ class AzureMLUpdateResourceActivity(ExecutionActivity): 'trained_model_file_path': {'key': 'typeProperties.trainedModelFilePath', 'type': 'object'}, } - def __init__(self, *, name: str, trained_model_name, trained_model_linked_service_name, trained_model_file_path, additional_properties=None, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, **kwargs) -> None: - super(AzureMLUpdateResourceActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) + def __init__(self, *, name: str, trained_model_name, trained_model_linked_service_name, trained_model_file_path, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, **kwargs) -> None: + super(AzureMLUpdateResourceActivity, self).__init__(name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) self.trained_model_name = trained_model_name self.trained_model_linked_service_name = trained_model_linked_service_name self.trained_model_file_path = trained_model_file_path diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/control_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/control_activity.py index b1915fc2e454..08cbef7c9f94 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/control_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/control_activity.py @@ -21,9 +21,6 @@ class ControlActivity(Activity): All required parameters must be populated in order to send to Azure. - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -31,7 +28,7 @@ class ControlActivity(Activity): :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: dict[str, str] + :type user_properties: list[~azure.mgmt.datafactory.models.UserProperty] :param type: Required. Constant filled by server. :type type: str """ @@ -42,11 +39,10 @@ class ControlActivity(Activity): } _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '{str}'}, + 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, 'type': {'key': 'type', 'type': 'str'}, } diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/control_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/control_activity_py3.py index 929788720bc4..244eda50b460 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/control_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/control_activity_py3.py @@ -21,9 +21,6 @@ class ControlActivity(Activity): All required parameters must be populated in order to send to Azure. - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -31,7 +28,7 @@ class ControlActivity(Activity): :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: dict[str, str] + :type user_properties: list[~azure.mgmt.datafactory.models.UserProperty] :param type: Required. Constant filled by server. :type type: str """ @@ -42,11 +39,10 @@ class ControlActivity(Activity): } _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '{str}'}, + 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, 'type': {'key': 'type', 'type': 'str'}, } @@ -54,6 +50,6 @@ class ControlActivity(Activity): 'type': {'Filter': 'FilterActivity', 'Until': 'UntilActivity', 'Wait': 'WaitActivity', 'ForEach': 'ForEachActivity', 'IfCondition': 'IfConditionActivity', 'ExecutePipeline': 'ExecutePipelineActivity'} } - def __init__(self, *, name: str, additional_properties=None, description: str=None, depends_on=None, user_properties=None, **kwargs) -> None: - super(ControlActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, **kwargs) + def __init__(self, *, name: str, description: str=None, depends_on=None, user_properties=None, **kwargs) -> None: + super(ControlActivity, self).__init__(name=name, description=description, depends_on=depends_on, user_properties=user_properties, **kwargs) self.type = 'Container' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_activity.py index e26170d07c14..6010102627ea 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_activity.py @@ -17,9 +17,6 @@ class CopyActivity(ExecutionActivity): All required parameters must be populated in order to send to Azure. - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -27,7 +24,7 @@ class CopyActivity(ExecutionActivity): :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: dict[str, str] + :type user_properties: list[~azure.mgmt.datafactory.models.UserProperty] :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. @@ -79,11 +76,10 @@ class CopyActivity(ExecutionActivity): } _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '{str}'}, + 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, 'type': {'key': 'type', 'type': 'str'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_activity_py3.py index bdbd626bb317..248c0ed24f80 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_activity_py3.py @@ -17,9 +17,6 @@ class CopyActivity(ExecutionActivity): All required parameters must be populated in order to send to Azure. - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -27,7 +24,7 @@ class CopyActivity(ExecutionActivity): :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: dict[str, str] + :type user_properties: list[~azure.mgmt.datafactory.models.UserProperty] :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. @@ -79,11 +76,10 @@ class CopyActivity(ExecutionActivity): } _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '{str}'}, + 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, 'type': {'key': 'type', 'type': 'str'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, @@ -100,8 +96,8 @@ class CopyActivity(ExecutionActivity): 'outputs': {'key': 'outputs', 'type': '[DatasetReference]'}, } - def __init__(self, *, name: str, source, sink, additional_properties=None, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, translator=None, enable_staging=None, staging_settings=None, parallel_copies=None, data_integration_units=None, enable_skip_incompatible_row=None, redirect_incompatible_row_settings=None, inputs=None, outputs=None, **kwargs) -> None: - super(CopyActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) + def __init__(self, *, name: str, source, sink, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, translator=None, enable_staging=None, staging_settings=None, parallel_copies=None, data_integration_units=None, enable_skip_incompatible_row=None, redirect_incompatible_row_settings=None, inputs=None, outputs=None, **kwargs) -> None: + super(CopyActivity, self).__init__(name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) self.source = source self.sink = sink self.translator = translator diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_activity.py index 9c0d65a18a16..b8211e0b33ad 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_activity.py @@ -17,9 +17,6 @@ class CustomActivity(ExecutionActivity): All required parameters must be populated in order to send to Azure. - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -27,7 +24,7 @@ class CustomActivity(ExecutionActivity): :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: dict[str, str] + :type user_properties: list[~azure.mgmt.datafactory.models.UserProperty] :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. @@ -61,11 +58,10 @@ class CustomActivity(ExecutionActivity): } _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '{str}'}, + 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, 'type': {'key': 'type', 'type': 'str'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_activity_py3.py index abaee4d53210..d5bbf82e7a93 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_activity_py3.py @@ -17,9 +17,6 @@ class CustomActivity(ExecutionActivity): All required parameters must be populated in order to send to Azure. - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -27,7 +24,7 @@ class CustomActivity(ExecutionActivity): :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: dict[str, str] + :type user_properties: list[~azure.mgmt.datafactory.models.UserProperty] :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. @@ -61,11 +58,10 @@ class CustomActivity(ExecutionActivity): } _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '{str}'}, + 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, 'type': {'key': 'type', 'type': 'str'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, @@ -76,8 +72,8 @@ class CustomActivity(ExecutionActivity): 'extended_properties': {'key': 'typeProperties.extendedProperties', 'type': '{object}'}, } - def __init__(self, *, name: str, command, additional_properties=None, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, resource_linked_service=None, folder_path=None, reference_objects=None, extended_properties=None, **kwargs) -> None: - super(CustomActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) + def __init__(self, *, name: str, command, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, resource_linked_service=None, folder_path=None, reference_objects=None, extended_properties=None, **kwargs) -> None: + super(CustomActivity, self).__init__(name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) self.command = command self.resource_linked_service = resource_linked_service self.folder_path = folder_path diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_lake_analytics_usql_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_lake_analytics_usql_activity.py index e35da3e25273..8cd09b705857 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_lake_analytics_usql_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_lake_analytics_usql_activity.py @@ -17,9 +17,6 @@ class DataLakeAnalyticsUSQLActivity(ExecutionActivity): All required parameters must be populated in order to send to Azure. - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -27,7 +24,7 @@ class DataLakeAnalyticsUSQLActivity(ExecutionActivity): :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: dict[str, str] + :type user_properties: list[~azure.mgmt.datafactory.models.UserProperty] :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. @@ -69,11 +66,10 @@ class DataLakeAnalyticsUSQLActivity(ExecutionActivity): } _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '{str}'}, + 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, 'type': {'key': 'type', 'type': 'str'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_lake_analytics_usql_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_lake_analytics_usql_activity_py3.py index 5e4e4968b535..25242ae40705 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_lake_analytics_usql_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_lake_analytics_usql_activity_py3.py @@ -17,9 +17,6 @@ class DataLakeAnalyticsUSQLActivity(ExecutionActivity): All required parameters must be populated in order to send to Azure. - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -27,7 +24,7 @@ class DataLakeAnalyticsUSQLActivity(ExecutionActivity): :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: dict[str, str] + :type user_properties: list[~azure.mgmt.datafactory.models.UserProperty] :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. @@ -69,11 +66,10 @@ class DataLakeAnalyticsUSQLActivity(ExecutionActivity): } _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '{str}'}, + 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, 'type': {'key': 'type', 'type': 'str'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, @@ -86,8 +82,8 @@ class DataLakeAnalyticsUSQLActivity(ExecutionActivity): 'compilation_mode': {'key': 'typeProperties.compilationMode', 'type': 'object'}, } - def __init__(self, *, name: str, script_path, script_linked_service, additional_properties=None, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, degree_of_parallelism=None, priority=None, parameters=None, runtime_version=None, compilation_mode=None, **kwargs) -> None: - super(DataLakeAnalyticsUSQLActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) + def __init__(self, *, name: str, script_path, script_linked_service, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, degree_of_parallelism=None, priority=None, parameters=None, runtime_version=None, compilation_mode=None, **kwargs) -> None: + super(DataLakeAnalyticsUSQLActivity, self).__init__(name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) self.script_path = script_path self.script_linked_service = script_linked_service self.degree_of_parallelism = degree_of_parallelism diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_notebook_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_notebook_activity.py index f9a337e26e33..9d61043c2a3a 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_notebook_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_notebook_activity.py @@ -17,9 +17,6 @@ class DatabricksNotebookActivity(ExecutionActivity): All required parameters must be populated in order to send to Azure. - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -27,7 +24,7 @@ class DatabricksNotebookActivity(ExecutionActivity): :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: dict[str, str] + :type user_properties: list[~azure.mgmt.datafactory.models.UserProperty] :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. @@ -55,11 +52,10 @@ class DatabricksNotebookActivity(ExecutionActivity): } _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '{str}'}, + 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, 'type': {'key': 'type', 'type': 'str'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_notebook_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_notebook_activity_py3.py index aeb97b6e28b0..0cd5e6bdb25d 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_notebook_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_notebook_activity_py3.py @@ -17,9 +17,6 @@ class DatabricksNotebookActivity(ExecutionActivity): All required parameters must be populated in order to send to Azure. - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -27,7 +24,7 @@ class DatabricksNotebookActivity(ExecutionActivity): :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: dict[str, str] + :type user_properties: list[~azure.mgmt.datafactory.models.UserProperty] :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. @@ -55,11 +52,10 @@ class DatabricksNotebookActivity(ExecutionActivity): } _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '{str}'}, + 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, 'type': {'key': 'type', 'type': 'str'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, @@ -68,8 +64,8 @@ class DatabricksNotebookActivity(ExecutionActivity): 'libraries': {'key': 'typeProperties.libraries', 'type': '[{object}]'}, } - def __init__(self, *, name: str, notebook_path, additional_properties=None, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, base_parameters=None, libraries=None, **kwargs) -> None: - super(DatabricksNotebookActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) + def __init__(self, *, name: str, notebook_path, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, base_parameters=None, libraries=None, **kwargs) -> None: + super(DatabricksNotebookActivity, self).__init__(name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) self.notebook_path = notebook_path self.base_parameters = base_parameters self.libraries = libraries diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_spark_jar_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_spark_jar_activity.py index 432caf54bdc3..42d12eba7117 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_spark_jar_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_spark_jar_activity.py @@ -17,9 +17,6 @@ class DatabricksSparkJarActivity(ExecutionActivity): All required parameters must be populated in order to send to Azure. - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -27,7 +24,7 @@ class DatabricksSparkJarActivity(ExecutionActivity): :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: dict[str, str] + :type user_properties: list[~azure.mgmt.datafactory.models.UserProperty] :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. @@ -54,11 +51,10 @@ class DatabricksSparkJarActivity(ExecutionActivity): } _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '{str}'}, + 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, 'type': {'key': 'type', 'type': 'str'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_spark_jar_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_spark_jar_activity_py3.py index 99224eade24f..2c88c324bbf0 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_spark_jar_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_spark_jar_activity_py3.py @@ -17,9 +17,6 @@ class DatabricksSparkJarActivity(ExecutionActivity): All required parameters must be populated in order to send to Azure. - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -27,7 +24,7 @@ class DatabricksSparkJarActivity(ExecutionActivity): :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: dict[str, str] + :type user_properties: list[~azure.mgmt.datafactory.models.UserProperty] :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. @@ -54,11 +51,10 @@ class DatabricksSparkJarActivity(ExecutionActivity): } _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '{str}'}, + 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, 'type': {'key': 'type', 'type': 'str'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, @@ -67,8 +63,8 @@ class DatabricksSparkJarActivity(ExecutionActivity): 'libraries': {'key': 'typeProperties.libraries', 'type': '[{object}]'}, } - def __init__(self, *, name: str, main_class_name, additional_properties=None, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, parameters=None, libraries=None, **kwargs) -> None: - super(DatabricksSparkJarActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) + def __init__(self, *, name: str, main_class_name, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, parameters=None, libraries=None, **kwargs) -> None: + super(DatabricksSparkJarActivity, self).__init__(name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) self.main_class_name = main_class_name self.parameters = parameters self.libraries = libraries diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_spark_python_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_spark_python_activity.py index 76a949177514..9cbb17a7ddf7 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_spark_python_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_spark_python_activity.py @@ -17,9 +17,6 @@ class DatabricksSparkPythonActivity(ExecutionActivity): All required parameters must be populated in order to send to Azure. - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -27,7 +24,7 @@ class DatabricksSparkPythonActivity(ExecutionActivity): :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: dict[str, str] + :type user_properties: list[~azure.mgmt.datafactory.models.UserProperty] :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. @@ -54,11 +51,10 @@ class DatabricksSparkPythonActivity(ExecutionActivity): } _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '{str}'}, + 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, 'type': {'key': 'type', 'type': 'str'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_spark_python_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_spark_python_activity_py3.py index 26039a6958c4..7090ee70681c 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_spark_python_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_spark_python_activity_py3.py @@ -17,9 +17,6 @@ class DatabricksSparkPythonActivity(ExecutionActivity): All required parameters must be populated in order to send to Azure. - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -27,7 +24,7 @@ class DatabricksSparkPythonActivity(ExecutionActivity): :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: dict[str, str] + :type user_properties: list[~azure.mgmt.datafactory.models.UserProperty] :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. @@ -54,11 +51,10 @@ class DatabricksSparkPythonActivity(ExecutionActivity): } _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '{str}'}, + 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, 'type': {'key': 'type', 'type': 'str'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, @@ -67,8 +63,8 @@ class DatabricksSparkPythonActivity(ExecutionActivity): 'libraries': {'key': 'typeProperties.libraries', 'type': '[{object}]'}, } - def __init__(self, *, name: str, python_file, additional_properties=None, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, parameters=None, libraries=None, **kwargs) -> None: - super(DatabricksSparkPythonActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) + def __init__(self, *, name: str, python_file, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, parameters=None, libraries=None, **kwargs) -> None: + super(DatabricksSparkPythonActivity, self).__init__(name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) self.python_file = python_file self.parameters = parameters self.libraries = libraries diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_pipeline_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_pipeline_activity.py index beab510d3367..24ffefbfbb1e 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_pipeline_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_pipeline_activity.py @@ -17,9 +17,6 @@ class ExecutePipelineActivity(ControlActivity): All required parameters must be populated in order to send to Azure. - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -27,7 +24,7 @@ class ExecutePipelineActivity(ControlActivity): :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: dict[str, str] + :type user_properties: list[~azure.mgmt.datafactory.models.UserProperty] :param type: Required. Constant filled by server. :type type: str :param pipeline: Required. Pipeline reference. @@ -46,11 +43,10 @@ class ExecutePipelineActivity(ControlActivity): } _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '{str}'}, + 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, 'type': {'key': 'type', 'type': 'str'}, 'pipeline': {'key': 'typeProperties.pipeline', 'type': 'PipelineReference'}, 'parameters': {'key': 'typeProperties.parameters', 'type': '{object}'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_pipeline_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_pipeline_activity_py3.py index ef878fec54fc..3bf9865f9f71 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_pipeline_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_pipeline_activity_py3.py @@ -17,9 +17,6 @@ class ExecutePipelineActivity(ControlActivity): All required parameters must be populated in order to send to Azure. - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -27,7 +24,7 @@ class ExecutePipelineActivity(ControlActivity): :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: dict[str, str] + :type user_properties: list[~azure.mgmt.datafactory.models.UserProperty] :param type: Required. Constant filled by server. :type type: str :param pipeline: Required. Pipeline reference. @@ -46,19 +43,18 @@ class ExecutePipelineActivity(ControlActivity): } _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '{str}'}, + 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, 'type': {'key': 'type', 'type': 'str'}, 'pipeline': {'key': 'typeProperties.pipeline', 'type': 'PipelineReference'}, 'parameters': {'key': 'typeProperties.parameters', 'type': '{object}'}, 'wait_on_completion': {'key': 'typeProperties.waitOnCompletion', 'type': 'bool'}, } - def __init__(self, *, name: str, pipeline, additional_properties=None, description: str=None, depends_on=None, user_properties=None, parameters=None, wait_on_completion: bool=None, **kwargs) -> None: - super(ExecutePipelineActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, **kwargs) + def __init__(self, *, name: str, pipeline, description: str=None, depends_on=None, user_properties=None, parameters=None, wait_on_completion: bool=None, **kwargs) -> None: + super(ExecutePipelineActivity, self).__init__(name=name, description=description, depends_on=depends_on, user_properties=user_properties, **kwargs) self.pipeline = pipeline self.parameters = parameters self.wait_on_completion = wait_on_completion diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_ssis_package_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_ssis_package_activity.py index 66282d877881..6df95be62e2b 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_ssis_package_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_ssis_package_activity.py @@ -17,9 +17,6 @@ class ExecuteSSISPackageActivity(ExecutionActivity): All required parameters must be populated in order to send to Azure. - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -27,7 +24,7 @@ class ExecuteSSISPackageActivity(ExecutionActivity): :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: dict[str, str] + :type user_properties: list[~azure.mgmt.datafactory.models.UserProperty] :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. @@ -77,11 +74,10 @@ class ExecuteSSISPackageActivity(ExecutionActivity): } _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '{str}'}, + 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, 'type': {'key': 'type', 'type': 'str'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_ssis_package_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_ssis_package_activity_py3.py index a1437141b785..d1cbdf2a66d2 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_ssis_package_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_ssis_package_activity_py3.py @@ -17,9 +17,6 @@ class ExecuteSSISPackageActivity(ExecutionActivity): All required parameters must be populated in order to send to Azure. - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -27,7 +24,7 @@ class ExecuteSSISPackageActivity(ExecutionActivity): :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: dict[str, str] + :type user_properties: list[~azure.mgmt.datafactory.models.UserProperty] :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. @@ -77,11 +74,10 @@ class ExecuteSSISPackageActivity(ExecutionActivity): } _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '{str}'}, + 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, 'type': {'key': 'type', 'type': 'str'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, @@ -97,8 +93,8 @@ class ExecuteSSISPackageActivity(ExecutionActivity): 'property_overrides': {'key': 'typeProperties.propertyOverrides', 'type': '{SSISPropertyOverride}'}, } - def __init__(self, *, name: str, package_location, connect_via, additional_properties=None, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, runtime=None, logging_level: str=None, environment_path: str=None, project_parameters=None, package_parameters=None, project_connection_managers=None, package_connection_managers=None, property_overrides=None, **kwargs) -> None: - super(ExecuteSSISPackageActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) + def __init__(self, *, name: str, package_location, connect_via, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, runtime=None, logging_level: str=None, environment_path: str=None, project_parameters=None, package_parameters=None, project_connection_managers=None, package_connection_managers=None, property_overrides=None, **kwargs) -> None: + super(ExecuteSSISPackageActivity, self).__init__(name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) self.package_location = package_location self.runtime = runtime self.logging_level = logging_level diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execution_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execution_activity.py index f8c64232576b..dcd134d880fd 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execution_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execution_activity.py @@ -27,9 +27,6 @@ class ExecutionActivity(Activity): All required parameters must be populated in order to send to Azure. - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -37,7 +34,7 @@ class ExecutionActivity(Activity): :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: dict[str, str] + :type user_properties: list[~azure.mgmt.datafactory.models.UserProperty] :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. @@ -53,11 +50,10 @@ class ExecutionActivity(Activity): } _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '{str}'}, + 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, 'type': {'key': 'type', 'type': 'str'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execution_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execution_activity_py3.py index e54a7b1fea57..299e327a7d11 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execution_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execution_activity_py3.py @@ -27,9 +27,6 @@ class ExecutionActivity(Activity): All required parameters must be populated in order to send to Azure. - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -37,7 +34,7 @@ class ExecutionActivity(Activity): :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: dict[str, str] + :type user_properties: list[~azure.mgmt.datafactory.models.UserProperty] :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. @@ -53,11 +50,10 @@ class ExecutionActivity(Activity): } _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '{str}'}, + 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, 'type': {'key': 'type', 'type': 'str'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, @@ -67,8 +63,8 @@ class ExecutionActivity(Activity): 'type': {'DatabricksSparkPython': 'DatabricksSparkPythonActivity', 'DatabricksSparkJar': 'DatabricksSparkJarActivity', 'DatabricksNotebook': 'DatabricksNotebookActivity', 'DataLakeAnalyticsU-SQL': 'DataLakeAnalyticsUSQLActivity', 'AzureMLUpdateResource': 'AzureMLUpdateResourceActivity', 'AzureMLBatchExecution': 'AzureMLBatchExecutionActivity', 'GetMetadata': 'GetMetadataActivity', 'WebActivity': 'WebActivity', 'Lookup': 'LookupActivity', 'SqlServerStoredProcedure': 'SqlServerStoredProcedureActivity', 'Custom': 'CustomActivity', 'ExecuteSSISPackage': 'ExecuteSSISPackageActivity', 'HDInsightSpark': 'HDInsightSparkActivity', 'HDInsightStreaming': 'HDInsightStreamingActivity', 'HDInsightMapReduce': 'HDInsightMapReduceActivity', 'HDInsightPig': 'HDInsightPigActivity', 'HDInsightHive': 'HDInsightHiveActivity', 'Copy': 'CopyActivity'} } - def __init__(self, *, name: str, additional_properties=None, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, **kwargs) -> None: - super(ExecutionActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, **kwargs) + def __init__(self, *, name: str, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, **kwargs) -> None: + super(ExecutionActivity, self).__init__(name=name, description=description, depends_on=depends_on, user_properties=user_properties, **kwargs) self.linked_service_name = linked_service_name self.policy = policy self.type = 'Execution' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/filter_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/filter_activity.py index 29a386fdf0eb..d8fdaf2d47b3 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/filter_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/filter_activity.py @@ -17,9 +17,6 @@ class FilterActivity(ControlActivity): All required parameters must be populated in order to send to Azure. - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -27,7 +24,7 @@ class FilterActivity(ControlActivity): :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: dict[str, str] + :type user_properties: list[~azure.mgmt.datafactory.models.UserProperty] :param type: Required. Constant filled by server. :type type: str :param items: Required. Input array on which filter should be applied. @@ -44,11 +41,10 @@ class FilterActivity(ControlActivity): } _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '{str}'}, + 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, 'type': {'key': 'type', 'type': 'str'}, 'items': {'key': 'typeProperties.items', 'type': 'Expression'}, 'condition': {'key': 'typeProperties.condition', 'type': 'Expression'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/filter_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/filter_activity_py3.py index 970767da8b32..5756334b35d5 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/filter_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/filter_activity_py3.py @@ -17,9 +17,6 @@ class FilterActivity(ControlActivity): All required parameters must be populated in order to send to Azure. - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -27,7 +24,7 @@ class FilterActivity(ControlActivity): :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: dict[str, str] + :type user_properties: list[~azure.mgmt.datafactory.models.UserProperty] :param type: Required. Constant filled by server. :type type: str :param items: Required. Input array on which filter should be applied. @@ -44,18 +41,17 @@ class FilterActivity(ControlActivity): } _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '{str}'}, + 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, 'type': {'key': 'type', 'type': 'str'}, 'items': {'key': 'typeProperties.items', 'type': 'Expression'}, 'condition': {'key': 'typeProperties.condition', 'type': 'Expression'}, } - def __init__(self, *, name: str, items, condition, additional_properties=None, description: str=None, depends_on=None, user_properties=None, **kwargs) -> None: - super(FilterActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, **kwargs) + def __init__(self, *, name: str, items, condition, description: str=None, depends_on=None, user_properties=None, **kwargs) -> None: + super(FilterActivity, self).__init__(name=name, description=description, depends_on=depends_on, user_properties=user_properties, **kwargs) self.items = items self.condition = condition self.type = 'Filter' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/for_each_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/for_each_activity.py index e5c37d53a17e..25f30e46dad7 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/for_each_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/for_each_activity.py @@ -18,9 +18,6 @@ class ForEachActivity(ControlActivity): All required parameters must be populated in order to send to Azure. - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -28,7 +25,7 @@ class ForEachActivity(ControlActivity): :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: dict[str, str] + :type user_properties: list[~azure.mgmt.datafactory.models.UserProperty] :param type: Required. Constant filled by server. :type type: str :param is_sequential: Should the loop be executed in sequence or in @@ -52,11 +49,10 @@ class ForEachActivity(ControlActivity): } _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '{str}'}, + 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, 'type': {'key': 'type', 'type': 'str'}, 'is_sequential': {'key': 'typeProperties.isSequential', 'type': 'bool'}, 'batch_count': {'key': 'typeProperties.batchCount', 'type': 'int'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/for_each_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/for_each_activity_py3.py index f33af7b8e254..87ae2a2e73fd 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/for_each_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/for_each_activity_py3.py @@ -18,9 +18,6 @@ class ForEachActivity(ControlActivity): All required parameters must be populated in order to send to Azure. - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -28,7 +25,7 @@ class ForEachActivity(ControlActivity): :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: dict[str, str] + :type user_properties: list[~azure.mgmt.datafactory.models.UserProperty] :param type: Required. Constant filled by server. :type type: str :param is_sequential: Should the loop be executed in sequence or in @@ -52,11 +49,10 @@ class ForEachActivity(ControlActivity): } _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '{str}'}, + 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, 'type': {'key': 'type', 'type': 'str'}, 'is_sequential': {'key': 'typeProperties.isSequential', 'type': 'bool'}, 'batch_count': {'key': 'typeProperties.batchCount', 'type': 'int'}, @@ -64,8 +60,8 @@ class ForEachActivity(ControlActivity): 'activities': {'key': 'typeProperties.activities', 'type': '[Activity]'}, } - def __init__(self, *, name: str, items, activities, additional_properties=None, description: str=None, depends_on=None, user_properties=None, is_sequential: bool=None, batch_count: int=None, **kwargs) -> None: - super(ForEachActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, **kwargs) + def __init__(self, *, name: str, items, activities, description: str=None, depends_on=None, user_properties=None, is_sequential: bool=None, batch_count: int=None, **kwargs) -> None: + super(ForEachActivity, self).__init__(name=name, description=description, depends_on=depends_on, user_properties=user_properties, **kwargs) self.is_sequential = is_sequential self.batch_count = batch_count self.items = items diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/get_metadata_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/get_metadata_activity.py index 2d5d2c99447e..187cad77ecff 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/get_metadata_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/get_metadata_activity.py @@ -17,9 +17,6 @@ class GetMetadataActivity(ExecutionActivity): All required parameters must be populated in order to send to Azure. - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -27,7 +24,7 @@ class GetMetadataActivity(ExecutionActivity): :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: dict[str, str] + :type user_properties: list[~azure.mgmt.datafactory.models.UserProperty] :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. @@ -48,11 +45,10 @@ class GetMetadataActivity(ExecutionActivity): } _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '{str}'}, + 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, 'type': {'key': 'type', 'type': 'str'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/get_metadata_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/get_metadata_activity_py3.py index b992f462b6f3..215a26397f7d 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/get_metadata_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/get_metadata_activity_py3.py @@ -17,9 +17,6 @@ class GetMetadataActivity(ExecutionActivity): All required parameters must be populated in order to send to Azure. - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -27,7 +24,7 @@ class GetMetadataActivity(ExecutionActivity): :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: dict[str, str] + :type user_properties: list[~azure.mgmt.datafactory.models.UserProperty] :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. @@ -48,11 +45,10 @@ class GetMetadataActivity(ExecutionActivity): } _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '{str}'}, + 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, 'type': {'key': 'type', 'type': 'str'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, @@ -60,8 +56,8 @@ class GetMetadataActivity(ExecutionActivity): 'field_list': {'key': 'typeProperties.fieldList', 'type': '[object]'}, } - def __init__(self, *, name: str, dataset, additional_properties=None, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, field_list=None, **kwargs) -> None: - super(GetMetadataActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) + def __init__(self, *, name: str, dataset, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, field_list=None, **kwargs) -> None: + super(GetMetadataActivity, self).__init__(name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) self.dataset = dataset self.field_list = field_list self.type = 'GetMetadata' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_hive_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_hive_activity.py index a856c100b8b0..a3a4117608b4 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_hive_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_hive_activity.py @@ -17,9 +17,6 @@ class HDInsightHiveActivity(ExecutionActivity): All required parameters must be populated in order to send to Azure. - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -27,7 +24,7 @@ class HDInsightHiveActivity(ExecutionActivity): :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: dict[str, str] + :type user_properties: list[~azure.mgmt.datafactory.models.UserProperty] :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. @@ -60,11 +57,10 @@ class HDInsightHiveActivity(ExecutionActivity): } _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '{str}'}, + 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, 'type': {'key': 'type', 'type': 'str'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_hive_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_hive_activity_py3.py index 3ced50d708ac..d78931ee19b5 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_hive_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_hive_activity_py3.py @@ -17,9 +17,6 @@ class HDInsightHiveActivity(ExecutionActivity): All required parameters must be populated in order to send to Azure. - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -27,7 +24,7 @@ class HDInsightHiveActivity(ExecutionActivity): :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: dict[str, str] + :type user_properties: list[~azure.mgmt.datafactory.models.UserProperty] :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. @@ -60,11 +57,10 @@ class HDInsightHiveActivity(ExecutionActivity): } _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '{str}'}, + 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, 'type': {'key': 'type', 'type': 'str'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, @@ -76,8 +72,8 @@ class HDInsightHiveActivity(ExecutionActivity): 'defines': {'key': 'typeProperties.defines', 'type': '{object}'}, } - def __init__(self, *, name: str, additional_properties=None, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, storage_linked_services=None, arguments=None, get_debug_info=None, script_path=None, script_linked_service=None, defines=None, **kwargs) -> None: - super(HDInsightHiveActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) + def __init__(self, *, name: str, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, storage_linked_services=None, arguments=None, get_debug_info=None, script_path=None, script_linked_service=None, defines=None, **kwargs) -> None: + super(HDInsightHiveActivity, self).__init__(name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) self.storage_linked_services = storage_linked_services self.arguments = arguments self.get_debug_info = get_debug_info diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_map_reduce_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_map_reduce_activity.py index 293013d73afc..11f23563c2ca 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_map_reduce_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_map_reduce_activity.py @@ -17,9 +17,6 @@ class HDInsightMapReduceActivity(ExecutionActivity): All required parameters must be populated in order to send to Azure. - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -27,7 +24,7 @@ class HDInsightMapReduceActivity(ExecutionActivity): :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: dict[str, str] + :type user_properties: list[~azure.mgmt.datafactory.models.UserProperty] :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. @@ -68,11 +65,10 @@ class HDInsightMapReduceActivity(ExecutionActivity): } _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '{str}'}, + 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, 'type': {'key': 'type', 'type': 'str'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_map_reduce_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_map_reduce_activity_py3.py index 830df1d395c2..9d5626490df3 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_map_reduce_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_map_reduce_activity_py3.py @@ -17,9 +17,6 @@ class HDInsightMapReduceActivity(ExecutionActivity): All required parameters must be populated in order to send to Azure. - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -27,7 +24,7 @@ class HDInsightMapReduceActivity(ExecutionActivity): :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: dict[str, str] + :type user_properties: list[~azure.mgmt.datafactory.models.UserProperty] :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. @@ -68,11 +65,10 @@ class HDInsightMapReduceActivity(ExecutionActivity): } _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '{str}'}, + 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, 'type': {'key': 'type', 'type': 'str'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, @@ -86,8 +82,8 @@ class HDInsightMapReduceActivity(ExecutionActivity): 'defines': {'key': 'typeProperties.defines', 'type': '{object}'}, } - def __init__(self, *, name: str, class_name, jar_file_path, additional_properties=None, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, storage_linked_services=None, arguments=None, get_debug_info=None, jar_linked_service=None, jar_libs=None, defines=None, **kwargs) -> None: - super(HDInsightMapReduceActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) + def __init__(self, *, name: str, class_name, jar_file_path, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, storage_linked_services=None, arguments=None, get_debug_info=None, jar_linked_service=None, jar_libs=None, defines=None, **kwargs) -> None: + super(HDInsightMapReduceActivity, self).__init__(name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) self.storage_linked_services = storage_linked_services self.arguments = arguments self.get_debug_info = get_debug_info diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_pig_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_pig_activity.py index 9400060e9d12..2a68916a36fe 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_pig_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_pig_activity.py @@ -17,9 +17,6 @@ class HDInsightPigActivity(ExecutionActivity): All required parameters must be populated in order to send to Azure. - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -27,7 +24,7 @@ class HDInsightPigActivity(ExecutionActivity): :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: dict[str, str] + :type user_properties: list[~azure.mgmt.datafactory.models.UserProperty] :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. @@ -60,11 +57,10 @@ class HDInsightPigActivity(ExecutionActivity): } _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '{str}'}, + 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, 'type': {'key': 'type', 'type': 'str'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_pig_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_pig_activity_py3.py index 839d193c5059..137767256ed5 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_pig_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_pig_activity_py3.py @@ -17,9 +17,6 @@ class HDInsightPigActivity(ExecutionActivity): All required parameters must be populated in order to send to Azure. - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -27,7 +24,7 @@ class HDInsightPigActivity(ExecutionActivity): :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: dict[str, str] + :type user_properties: list[~azure.mgmt.datafactory.models.UserProperty] :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. @@ -60,11 +57,10 @@ class HDInsightPigActivity(ExecutionActivity): } _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '{str}'}, + 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, 'type': {'key': 'type', 'type': 'str'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, @@ -76,8 +72,8 @@ class HDInsightPigActivity(ExecutionActivity): 'defines': {'key': 'typeProperties.defines', 'type': '{object}'}, } - def __init__(self, *, name: str, additional_properties=None, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, storage_linked_services=None, arguments=None, get_debug_info=None, script_path=None, script_linked_service=None, defines=None, **kwargs) -> None: - super(HDInsightPigActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) + def __init__(self, *, name: str, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, storage_linked_services=None, arguments=None, get_debug_info=None, script_path=None, script_linked_service=None, defines=None, **kwargs) -> None: + super(HDInsightPigActivity, self).__init__(name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) self.storage_linked_services = storage_linked_services self.arguments = arguments self.get_debug_info = get_debug_info diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_spark_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_spark_activity.py index 422177713af4..76c9a359ae1b 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_spark_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_spark_activity.py @@ -17,9 +17,6 @@ class HDInsightSparkActivity(ExecutionActivity): All required parameters must be populated in order to send to Azure. - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -27,7 +24,7 @@ class HDInsightSparkActivity(ExecutionActivity): :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: dict[str, str] + :type user_properties: list[~azure.mgmt.datafactory.models.UserProperty] :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. @@ -69,11 +66,10 @@ class HDInsightSparkActivity(ExecutionActivity): } _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '{str}'}, + 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, 'type': {'key': 'type', 'type': 'str'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_spark_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_spark_activity_py3.py index b3a44e425368..528bb71450d3 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_spark_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_spark_activity_py3.py @@ -17,9 +17,6 @@ class HDInsightSparkActivity(ExecutionActivity): All required parameters must be populated in order to send to Azure. - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -27,7 +24,7 @@ class HDInsightSparkActivity(ExecutionActivity): :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: dict[str, str] + :type user_properties: list[~azure.mgmt.datafactory.models.UserProperty] :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. @@ -69,11 +66,10 @@ class HDInsightSparkActivity(ExecutionActivity): } _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '{str}'}, + 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, 'type': {'key': 'type', 'type': 'str'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, @@ -87,8 +83,8 @@ class HDInsightSparkActivity(ExecutionActivity): 'spark_config': {'key': 'typeProperties.sparkConfig', 'type': '{object}'}, } - def __init__(self, *, name: str, root_path, entry_file_path, additional_properties=None, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, arguments=None, get_debug_info=None, spark_job_linked_service=None, class_name: str=None, proxy_user=None, spark_config=None, **kwargs) -> None: - super(HDInsightSparkActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) + def __init__(self, *, name: str, root_path, entry_file_path, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, arguments=None, get_debug_info=None, spark_job_linked_service=None, class_name: str=None, proxy_user=None, spark_config=None, **kwargs) -> None: + super(HDInsightSparkActivity, self).__init__(name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) self.root_path = root_path self.entry_file_path = entry_file_path self.arguments = arguments diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_streaming_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_streaming_activity.py index fc2bcc3ac1b8..87e7fe5cbe3d 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_streaming_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_streaming_activity.py @@ -17,9 +17,6 @@ class HDInsightStreamingActivity(ExecutionActivity): All required parameters must be populated in order to send to Azure. - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -27,7 +24,7 @@ class HDInsightStreamingActivity(ExecutionActivity): :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: dict[str, str] + :type user_properties: list[~azure.mgmt.datafactory.models.UserProperty] :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. @@ -83,11 +80,10 @@ class HDInsightStreamingActivity(ExecutionActivity): } _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '{str}'}, + 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, 'type': {'key': 'type', 'type': 'str'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_streaming_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_streaming_activity_py3.py index 40ae73bf18ec..195903974063 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_streaming_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_streaming_activity_py3.py @@ -17,9 +17,6 @@ class HDInsightStreamingActivity(ExecutionActivity): All required parameters must be populated in order to send to Azure. - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -27,7 +24,7 @@ class HDInsightStreamingActivity(ExecutionActivity): :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: dict[str, str] + :type user_properties: list[~azure.mgmt.datafactory.models.UserProperty] :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. @@ -83,11 +80,10 @@ class HDInsightStreamingActivity(ExecutionActivity): } _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '{str}'}, + 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, 'type': {'key': 'type', 'type': 'str'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, @@ -105,8 +101,8 @@ class HDInsightStreamingActivity(ExecutionActivity): 'defines': {'key': 'typeProperties.defines', 'type': '{object}'}, } - def __init__(self, *, name: str, mapper, reducer, input, output, file_paths, additional_properties=None, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, storage_linked_services=None, arguments=None, get_debug_info=None, file_linked_service=None, combiner=None, command_environment=None, defines=None, **kwargs) -> None: - super(HDInsightStreamingActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) + def __init__(self, *, name: str, mapper, reducer, input, output, file_paths, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, storage_linked_services=None, arguments=None, get_debug_info=None, file_linked_service=None, combiner=None, command_environment=None, defines=None, **kwargs) -> None: + super(HDInsightStreamingActivity, self).__init__(name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) self.storage_linked_services = storage_linked_services self.arguments = arguments self.get_debug_info = get_debug_info diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/if_condition_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/if_condition_activity.py index 1e66cc0a875b..9f210ac81682 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/if_condition_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/if_condition_activity.py @@ -19,9 +19,6 @@ class IfConditionActivity(ControlActivity): All required parameters must be populated in order to send to Azure. - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -29,7 +26,7 @@ class IfConditionActivity(ControlActivity): :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: dict[str, str] + :type user_properties: list[~azure.mgmt.datafactory.models.UserProperty] :param type: Required. Constant filled by server. :type type: str :param expression: Required. An expression that would evaluate to Boolean. @@ -53,11 +50,10 @@ class IfConditionActivity(ControlActivity): } _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '{str}'}, + 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, 'type': {'key': 'type', 'type': 'str'}, 'expression': {'key': 'typeProperties.expression', 'type': 'Expression'}, 'if_true_activities': {'key': 'typeProperties.ifTrueActivities', 'type': '[Activity]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/if_condition_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/if_condition_activity_py3.py index 00798ac67bff..404aaa8bef4d 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/if_condition_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/if_condition_activity_py3.py @@ -19,9 +19,6 @@ class IfConditionActivity(ControlActivity): All required parameters must be populated in order to send to Azure. - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -29,7 +26,7 @@ class IfConditionActivity(ControlActivity): :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: dict[str, str] + :type user_properties: list[~azure.mgmt.datafactory.models.UserProperty] :param type: Required. Constant filled by server. :type type: str :param expression: Required. An expression that would evaluate to Boolean. @@ -53,19 +50,18 @@ class IfConditionActivity(ControlActivity): } _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '{str}'}, + 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, 'type': {'key': 'type', 'type': 'str'}, 'expression': {'key': 'typeProperties.expression', 'type': 'Expression'}, 'if_true_activities': {'key': 'typeProperties.ifTrueActivities', 'type': '[Activity]'}, 'if_false_activities': {'key': 'typeProperties.ifFalseActivities', 'type': '[Activity]'}, } - def __init__(self, *, name: str, expression, additional_properties=None, description: str=None, depends_on=None, user_properties=None, if_true_activities=None, if_false_activities=None, **kwargs) -> None: - super(IfConditionActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, **kwargs) + def __init__(self, *, name: str, expression, description: str=None, depends_on=None, user_properties=None, if_true_activities=None, if_false_activities=None, **kwargs) -> None: + super(IfConditionActivity, self).__init__(name=name, description=description, depends_on=depends_on, user_properties=user_properties, **kwargs) self.expression = expression self.if_true_activities = if_true_activities self.if_false_activities = if_false_activities diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/lookup_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/lookup_activity.py index 0cf215272adf..4573d9369246 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/lookup_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/lookup_activity.py @@ -17,9 +17,6 @@ class LookupActivity(ExecutionActivity): All required parameters must be populated in order to send to Azure. - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -27,7 +24,7 @@ class LookupActivity(ExecutionActivity): :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: dict[str, str] + :type user_properties: list[~azure.mgmt.datafactory.models.UserProperty] :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. @@ -53,11 +50,10 @@ class LookupActivity(ExecutionActivity): } _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '{str}'}, + 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, 'type': {'key': 'type', 'type': 'str'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/lookup_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/lookup_activity_py3.py index ef81802111c0..22ba9d129585 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/lookup_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/lookup_activity_py3.py @@ -17,9 +17,6 @@ class LookupActivity(ExecutionActivity): All required parameters must be populated in order to send to Azure. - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -27,7 +24,7 @@ class LookupActivity(ExecutionActivity): :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: dict[str, str] + :type user_properties: list[~azure.mgmt.datafactory.models.UserProperty] :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. @@ -53,11 +50,10 @@ class LookupActivity(ExecutionActivity): } _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '{str}'}, + 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, 'type': {'key': 'type', 'type': 'str'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, @@ -66,8 +62,8 @@ class LookupActivity(ExecutionActivity): 'first_row_only': {'key': 'typeProperties.firstRowOnly', 'type': 'object'}, } - def __init__(self, *, name: str, source, dataset, additional_properties=None, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, first_row_only=None, **kwargs) -> None: - super(LookupActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) + def __init__(self, *, name: str, source, dataset, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, first_row_only=None, **kwargs) -> None: + super(LookupActivity, self).__init__(name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) self.source = source self.dataset = dataset self.first_row_only = first_row_only diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_stored_procedure_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_stored_procedure_activity.py index 5a80087427fa..163339b67ec9 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_stored_procedure_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_stored_procedure_activity.py @@ -17,9 +17,6 @@ class SqlServerStoredProcedureActivity(ExecutionActivity): All required parameters must be populated in order to send to Azure. - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -27,7 +24,7 @@ class SqlServerStoredProcedureActivity(ExecutionActivity): :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: dict[str, str] + :type user_properties: list[~azure.mgmt.datafactory.models.UserProperty] :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. @@ -51,11 +48,10 @@ class SqlServerStoredProcedureActivity(ExecutionActivity): } _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '{str}'}, + 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, 'type': {'key': 'type', 'type': 'str'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_stored_procedure_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_stored_procedure_activity_py3.py index f7200f8f5776..db487bd47b9f 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_stored_procedure_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_stored_procedure_activity_py3.py @@ -17,9 +17,6 @@ class SqlServerStoredProcedureActivity(ExecutionActivity): All required parameters must be populated in order to send to Azure. - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -27,7 +24,7 @@ class SqlServerStoredProcedureActivity(ExecutionActivity): :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: dict[str, str] + :type user_properties: list[~azure.mgmt.datafactory.models.UserProperty] :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. @@ -51,11 +48,10 @@ class SqlServerStoredProcedureActivity(ExecutionActivity): } _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '{str}'}, + 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, 'type': {'key': 'type', 'type': 'str'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, @@ -63,8 +59,8 @@ class SqlServerStoredProcedureActivity(ExecutionActivity): 'stored_procedure_parameters': {'key': 'typeProperties.storedProcedureParameters', 'type': '{StoredProcedureParameter}'}, } - def __init__(self, *, name: str, stored_procedure_name, additional_properties=None, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, stored_procedure_parameters=None, **kwargs) -> None: - super(SqlServerStoredProcedureActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) + def __init__(self, *, name: str, stored_procedure_name, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, stored_procedure_parameters=None, **kwargs) -> None: + super(SqlServerStoredProcedureActivity, self).__init__(name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) self.stored_procedure_name = stored_procedure_name self.stored_procedure_parameters = stored_procedure_parameters self.type = 'SqlServerStoredProcedure' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/until_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/until_activity.py index 707c02e538c5..f9e7472c93f3 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/until_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/until_activity.py @@ -18,9 +18,6 @@ class UntilActivity(ControlActivity): All required parameters must be populated in order to send to Azure. - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -28,7 +25,7 @@ class UntilActivity(ControlActivity): :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: dict[str, str] + :type user_properties: list[~azure.mgmt.datafactory.models.UserProperty] :param type: Required. Constant filled by server. :type type: str :param expression: Required. An expression that would evaluate to Boolean. @@ -53,11 +50,10 @@ class UntilActivity(ControlActivity): } _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '{str}'}, + 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, 'type': {'key': 'type', 'type': 'str'}, 'expression': {'key': 'typeProperties.expression', 'type': 'Expression'}, 'timeout': {'key': 'typeProperties.timeout', 'type': 'object'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/until_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/until_activity_py3.py index ac23fd6d4e92..795f8ed5a814 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/until_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/until_activity_py3.py @@ -18,9 +18,6 @@ class UntilActivity(ControlActivity): All required parameters must be populated in order to send to Azure. - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -28,7 +25,7 @@ class UntilActivity(ControlActivity): :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: dict[str, str] + :type user_properties: list[~azure.mgmt.datafactory.models.UserProperty] :param type: Required. Constant filled by server. :type type: str :param expression: Required. An expression that would evaluate to Boolean. @@ -53,19 +50,18 @@ class UntilActivity(ControlActivity): } _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '{str}'}, + 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, 'type': {'key': 'type', 'type': 'str'}, 'expression': {'key': 'typeProperties.expression', 'type': 'Expression'}, 'timeout': {'key': 'typeProperties.timeout', 'type': 'object'}, 'activities': {'key': 'typeProperties.activities', 'type': '[Activity]'}, } - def __init__(self, *, name: str, expression, activities, additional_properties=None, description: str=None, depends_on=None, user_properties=None, timeout=None, **kwargs) -> None: - super(UntilActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, **kwargs) + def __init__(self, *, name: str, expression, activities, description: str=None, depends_on=None, user_properties=None, timeout=None, **kwargs) -> None: + super(UntilActivity, self).__init__(name=name, description=description, depends_on=depends_on, user_properties=user_properties, **kwargs) self.expression = expression self.timeout = timeout self.activities = activities diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/user_property.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/user_property.py new file mode 100644 index 000000000000..b29aa61a0e6e --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/user_property.py @@ -0,0 +1,40 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class UserProperty(Model): + """User property. + + All required parameters must be populated in order to send to Azure. + + :param name: Required. User proprety name. + :type name: str + :param value: Required. User proprety value. Type: string (or Expression + with resultType string). + :type value: object + """ + + _validation = { + 'name': {'required': True}, + 'value': {'required': True}, + } + + _attribute_map = { + 'name': {'key': 'name', 'type': 'str'}, + 'value': {'key': 'value', 'type': 'object'}, + } + + def __init__(self, **kwargs): + super(UserProperty, self).__init__(**kwargs) + self.name = kwargs.get('name', None) + self.value = kwargs.get('value', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/user_property_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/user_property_py3.py new file mode 100644 index 000000000000..52691ac1d4e4 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/user_property_py3.py @@ -0,0 +1,40 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class UserProperty(Model): + """User property. + + All required parameters must be populated in order to send to Azure. + + :param name: Required. User proprety name. + :type name: str + :param value: Required. User proprety value. Type: string (or Expression + with resultType string). + :type value: object + """ + + _validation = { + 'name': {'required': True}, + 'value': {'required': True}, + } + + _attribute_map = { + 'name': {'key': 'name', 'type': 'str'}, + 'value': {'key': 'value', 'type': 'object'}, + } + + def __init__(self, *, name: str, value, **kwargs) -> None: + super(UserProperty, self).__init__(**kwargs) + self.name = name + self.value = value diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/wait_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/wait_activity.py index e2000f3f588c..987abc3eea78 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/wait_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/wait_activity.py @@ -17,9 +17,6 @@ class WaitActivity(ControlActivity): All required parameters must be populated in order to send to Azure. - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -27,7 +24,7 @@ class WaitActivity(ControlActivity): :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: dict[str, str] + :type user_properties: list[~azure.mgmt.datafactory.models.UserProperty] :param type: Required. Constant filled by server. :type type: str :param wait_time_in_seconds: Required. Duration in seconds. @@ -41,11 +38,10 @@ class WaitActivity(ControlActivity): } _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '{str}'}, + 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, 'type': {'key': 'type', 'type': 'str'}, 'wait_time_in_seconds': {'key': 'typeProperties.waitTimeInSeconds', 'type': 'int'}, } diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/wait_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/wait_activity_py3.py index 4170198d8d3b..31cc91ca1f3b 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/wait_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/wait_activity_py3.py @@ -17,9 +17,6 @@ class WaitActivity(ControlActivity): All required parameters must be populated in order to send to Azure. - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -27,7 +24,7 @@ class WaitActivity(ControlActivity): :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: dict[str, str] + :type user_properties: list[~azure.mgmt.datafactory.models.UserProperty] :param type: Required. Constant filled by server. :type type: str :param wait_time_in_seconds: Required. Duration in seconds. @@ -41,16 +38,15 @@ class WaitActivity(ControlActivity): } _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '{str}'}, + 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, 'type': {'key': 'type', 'type': 'str'}, 'wait_time_in_seconds': {'key': 'typeProperties.waitTimeInSeconds', 'type': 'int'}, } - def __init__(self, *, name: str, wait_time_in_seconds: int, additional_properties=None, description: str=None, depends_on=None, user_properties=None, **kwargs) -> None: - super(WaitActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, **kwargs) + def __init__(self, *, name: str, wait_time_in_seconds: int, description: str=None, depends_on=None, user_properties=None, **kwargs) -> None: + super(WaitActivity, self).__init__(name=name, description=description, depends_on=depends_on, user_properties=user_properties, **kwargs) self.wait_time_in_seconds = wait_time_in_seconds self.type = 'Wait' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_activity.py index 017aea067c90..8d175329d949 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_activity.py @@ -17,9 +17,6 @@ class WebActivity(ExecutionActivity): All required parameters must be populated in order to send to Azure. - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -27,7 +24,7 @@ class WebActivity(ExecutionActivity): :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: dict[str, str] + :type user_properties: list[~azure.mgmt.datafactory.models.UserProperty] :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. @@ -69,11 +66,10 @@ class WebActivity(ExecutionActivity): } _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '{str}'}, + 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, 'type': {'key': 'type', 'type': 'str'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_activity_py3.py index 5bdf1c4a0ac8..7d06e9a72884 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_activity_py3.py @@ -17,9 +17,6 @@ class WebActivity(ExecutionActivity): All required parameters must be populated in order to send to Azure. - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -27,7 +24,7 @@ class WebActivity(ExecutionActivity): :param depends_on: Activity depends on condition. :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: dict[str, str] + :type user_properties: list[~azure.mgmt.datafactory.models.UserProperty] :param type: Required. Constant filled by server. :type type: str :param linked_service_name: Linked service reference. @@ -69,11 +66,10 @@ class WebActivity(ExecutionActivity): } _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '{str}'}, + 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, 'type': {'key': 'type', 'type': 'str'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, @@ -86,8 +82,8 @@ class WebActivity(ExecutionActivity): 'linked_services': {'key': 'typeProperties.linkedServices', 'type': '[LinkedServiceReference]'}, } - def __init__(self, *, name: str, method, url, additional_properties=None, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, headers=None, body=None, authentication=None, datasets=None, linked_services=None, **kwargs) -> None: - super(WebActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) + def __init__(self, *, name: str, method, url, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, headers=None, body=None, authentication=None, datasets=None, linked_services=None, **kwargs) -> None: + super(WebActivity, self).__init__(name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) self.method = method self.url = url self.headers = headers From a5ab9c777c470c752947fbe4c097012866797d0f Mon Sep 17 00:00:00 2001 From: Azure SDK for Python bot Date: Tue, 24 Jul 2018 13:18:37 -0700 Subject: [PATCH 13/54] Generated from 04d330debfe9f9f86aece8489cb5758f4025c53a (#2996) [ADF] Bringing back additionalProperties in Activity removed by mistake --- .../azure/mgmt/datafactory/models/activity.py | 5 +++++ .../azure/mgmt/datafactory/models/activity_py3.py | 7 ++++++- .../models/azure_ml_batch_execution_activity.py | 4 ++++ .../models/azure_ml_batch_execution_activity_py3.py | 8 ++++++-- .../models/azure_ml_update_resource_activity.py | 4 ++++ .../models/azure_ml_update_resource_activity_py3.py | 8 ++++++-- .../azure/mgmt/datafactory/models/control_activity.py | 4 ++++ .../azure/mgmt/datafactory/models/control_activity_py3.py | 8 ++++++-- .../azure/mgmt/datafactory/models/copy_activity.py | 4 ++++ .../azure/mgmt/datafactory/models/copy_activity_py3.py | 8 ++++++-- .../azure/mgmt/datafactory/models/custom_activity.py | 4 ++++ .../azure/mgmt/datafactory/models/custom_activity_py3.py | 8 ++++++-- .../models/data_lake_analytics_usql_activity.py | 4 ++++ .../models/data_lake_analytics_usql_activity_py3.py | 8 ++++++-- .../datafactory/models/databricks_notebook_activity.py | 4 ++++ .../models/databricks_notebook_activity_py3.py | 8 ++++++-- .../datafactory/models/databricks_spark_jar_activity.py | 4 ++++ .../models/databricks_spark_jar_activity_py3.py | 8 ++++++-- .../models/databricks_spark_python_activity.py | 4 ++++ .../models/databricks_spark_python_activity_py3.py | 8 ++++++-- .../mgmt/datafactory/models/execute_pipeline_activity.py | 4 ++++ .../datafactory/models/execute_pipeline_activity_py3.py | 8 ++++++-- .../datafactory/models/execute_ssis_package_activity.py | 4 ++++ .../models/execute_ssis_package_activity_py3.py | 8 ++++++-- .../azure/mgmt/datafactory/models/execution_activity.py | 4 ++++ .../mgmt/datafactory/models/execution_activity_py3.py | 8 ++++++-- .../azure/mgmt/datafactory/models/filter_activity.py | 4 ++++ .../azure/mgmt/datafactory/models/filter_activity_py3.py | 8 ++++++-- .../azure/mgmt/datafactory/models/for_each_activity.py | 4 ++++ .../mgmt/datafactory/models/for_each_activity_py3.py | 8 ++++++-- .../mgmt/datafactory/models/get_metadata_activity.py | 4 ++++ .../mgmt/datafactory/models/get_metadata_activity_py3.py | 8 ++++++-- .../mgmt/datafactory/models/hd_insight_hive_activity.py | 4 ++++ .../datafactory/models/hd_insight_hive_activity_py3.py | 8 ++++++-- .../datafactory/models/hd_insight_map_reduce_activity.py | 4 ++++ .../models/hd_insight_map_reduce_activity_py3.py | 8 ++++++-- .../mgmt/datafactory/models/hd_insight_pig_activity.py | 4 ++++ .../datafactory/models/hd_insight_pig_activity_py3.py | 8 ++++++-- .../mgmt/datafactory/models/hd_insight_spark_activity.py | 4 ++++ .../datafactory/models/hd_insight_spark_activity_py3.py | 8 ++++++-- .../datafactory/models/hd_insight_streaming_activity.py | 4 ++++ .../models/hd_insight_streaming_activity_py3.py | 8 ++++++-- .../mgmt/datafactory/models/if_condition_activity.py | 4 ++++ .../mgmt/datafactory/models/if_condition_activity_py3.py | 8 ++++++-- .../azure/mgmt/datafactory/models/lookup_activity.py | 4 ++++ .../azure/mgmt/datafactory/models/lookup_activity_py3.py | 8 ++++++-- .../models/sql_server_stored_procedure_activity.py | 4 ++++ .../models/sql_server_stored_procedure_activity_py3.py | 8 ++++++-- .../azure/mgmt/datafactory/models/until_activity.py | 4 ++++ .../azure/mgmt/datafactory/models/until_activity_py3.py | 8 ++++++-- .../azure/mgmt/datafactory/models/wait_activity.py | 4 ++++ .../azure/mgmt/datafactory/models/wait_activity_py3.py | 8 ++++++-- .../azure/mgmt/datafactory/models/web_activity.py | 4 ++++ .../azure/mgmt/datafactory/models/web_activity_py3.py | 8 ++++++-- 54 files changed, 271 insertions(+), 53 deletions(-) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/activity.py index 06dca4b71127..72d920f1d04c 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/activity.py @@ -20,6 +20,9 @@ class Activity(Model): All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -38,6 +41,7 @@ class Activity(Model): } _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, @@ -51,6 +55,7 @@ class Activity(Model): def __init__(self, **kwargs): super(Activity, self).__init__(**kwargs) + self.additional_properties = kwargs.get('additional_properties', None) self.name = kwargs.get('name', None) self.description = kwargs.get('description', None) self.depends_on = kwargs.get('depends_on', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/activity_py3.py index a36cee0fc595..b5997c9352e1 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/activity_py3.py @@ -20,6 +20,9 @@ class Activity(Model): All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -38,6 +41,7 @@ class Activity(Model): } _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, @@ -49,8 +53,9 @@ class Activity(Model): 'type': {'Execution': 'ExecutionActivity', 'Container': 'ControlActivity'} } - def __init__(self, *, name: str, description: str=None, depends_on=None, user_properties=None, **kwargs) -> None: + def __init__(self, *, name: str, additional_properties=None, description: str=None, depends_on=None, user_properties=None, **kwargs) -> None: super(Activity, self).__init__(**kwargs) + self.additional_properties = additional_properties self.name = name self.description = description self.depends_on = depends_on diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_batch_execution_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_batch_execution_activity.py index 47a7a6dbd41e..f6c7c75a1299 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_batch_execution_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_batch_execution_activity.py @@ -17,6 +17,9 @@ class AzureMLBatchExecutionActivity(ExecutionActivity): All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -58,6 +61,7 @@ class AzureMLBatchExecutionActivity(ExecutionActivity): } _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_batch_execution_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_batch_execution_activity_py3.py index eaac413d277e..e273c0b38128 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_batch_execution_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_batch_execution_activity_py3.py @@ -17,6 +17,9 @@ class AzureMLBatchExecutionActivity(ExecutionActivity): All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -58,6 +61,7 @@ class AzureMLBatchExecutionActivity(ExecutionActivity): } _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, @@ -70,8 +74,8 @@ class AzureMLBatchExecutionActivity(ExecutionActivity): 'web_service_inputs': {'key': 'typeProperties.webServiceInputs', 'type': '{AzureMLWebServiceFile}'}, } - def __init__(self, *, name: str, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, global_parameters=None, web_service_outputs=None, web_service_inputs=None, **kwargs) -> None: - super(AzureMLBatchExecutionActivity, self).__init__(name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) + def __init__(self, *, name: str, additional_properties=None, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, global_parameters=None, web_service_outputs=None, web_service_inputs=None, **kwargs) -> None: + super(AzureMLBatchExecutionActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) self.global_parameters = global_parameters self.web_service_outputs = web_service_outputs self.web_service_inputs = web_service_inputs diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_update_resource_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_update_resource_activity.py index ad5911008bf2..c47a2d81648e 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_update_resource_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_update_resource_activity.py @@ -17,6 +17,9 @@ class AzureMLUpdateResourceActivity(ExecutionActivity): All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -57,6 +60,7 @@ class AzureMLUpdateResourceActivity(ExecutionActivity): } _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_update_resource_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_update_resource_activity_py3.py index 68b721fc94f1..50a5932f0bf0 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_update_resource_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_ml_update_resource_activity_py3.py @@ -17,6 +17,9 @@ class AzureMLUpdateResourceActivity(ExecutionActivity): All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -57,6 +60,7 @@ class AzureMLUpdateResourceActivity(ExecutionActivity): } _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, @@ -69,8 +73,8 @@ class AzureMLUpdateResourceActivity(ExecutionActivity): 'trained_model_file_path': {'key': 'typeProperties.trainedModelFilePath', 'type': 'object'}, } - def __init__(self, *, name: str, trained_model_name, trained_model_linked_service_name, trained_model_file_path, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, **kwargs) -> None: - super(AzureMLUpdateResourceActivity, self).__init__(name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) + def __init__(self, *, name: str, trained_model_name, trained_model_linked_service_name, trained_model_file_path, additional_properties=None, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, **kwargs) -> None: + super(AzureMLUpdateResourceActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) self.trained_model_name = trained_model_name self.trained_model_linked_service_name = trained_model_linked_service_name self.trained_model_file_path = trained_model_file_path diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/control_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/control_activity.py index 08cbef7c9f94..87677dcfa6bf 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/control_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/control_activity.py @@ -21,6 +21,9 @@ class ControlActivity(Activity): All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -39,6 +42,7 @@ class ControlActivity(Activity): } _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/control_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/control_activity_py3.py index 244eda50b460..09004ce2cee2 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/control_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/control_activity_py3.py @@ -21,6 +21,9 @@ class ControlActivity(Activity): All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -39,6 +42,7 @@ class ControlActivity(Activity): } _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, @@ -50,6 +54,6 @@ class ControlActivity(Activity): 'type': {'Filter': 'FilterActivity', 'Until': 'UntilActivity', 'Wait': 'WaitActivity', 'ForEach': 'ForEachActivity', 'IfCondition': 'IfConditionActivity', 'ExecutePipeline': 'ExecutePipelineActivity'} } - def __init__(self, *, name: str, description: str=None, depends_on=None, user_properties=None, **kwargs) -> None: - super(ControlActivity, self).__init__(name=name, description=description, depends_on=depends_on, user_properties=user_properties, **kwargs) + def __init__(self, *, name: str, additional_properties=None, description: str=None, depends_on=None, user_properties=None, **kwargs) -> None: + super(ControlActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, **kwargs) self.type = 'Container' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_activity.py index 6010102627ea..997c3820cb00 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_activity.py @@ -17,6 +17,9 @@ class CopyActivity(ExecutionActivity): All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -76,6 +79,7 @@ class CopyActivity(ExecutionActivity): } _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_activity_py3.py index 248c0ed24f80..8a740db78983 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_activity_py3.py @@ -17,6 +17,9 @@ class CopyActivity(ExecutionActivity): All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -76,6 +79,7 @@ class CopyActivity(ExecutionActivity): } _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, @@ -96,8 +100,8 @@ class CopyActivity(ExecutionActivity): 'outputs': {'key': 'outputs', 'type': '[DatasetReference]'}, } - def __init__(self, *, name: str, source, sink, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, translator=None, enable_staging=None, staging_settings=None, parallel_copies=None, data_integration_units=None, enable_skip_incompatible_row=None, redirect_incompatible_row_settings=None, inputs=None, outputs=None, **kwargs) -> None: - super(CopyActivity, self).__init__(name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) + def __init__(self, *, name: str, source, sink, additional_properties=None, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, translator=None, enable_staging=None, staging_settings=None, parallel_copies=None, data_integration_units=None, enable_skip_incompatible_row=None, redirect_incompatible_row_settings=None, inputs=None, outputs=None, **kwargs) -> None: + super(CopyActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) self.source = source self.sink = sink self.translator = translator diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_activity.py index b8211e0b33ad..f7eceb72ff3b 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_activity.py @@ -17,6 +17,9 @@ class CustomActivity(ExecutionActivity): All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -58,6 +61,7 @@ class CustomActivity(ExecutionActivity): } _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_activity_py3.py index d5bbf82e7a93..b82ac57bca4d 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_activity_py3.py @@ -17,6 +17,9 @@ class CustomActivity(ExecutionActivity): All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -58,6 +61,7 @@ class CustomActivity(ExecutionActivity): } _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, @@ -72,8 +76,8 @@ class CustomActivity(ExecutionActivity): 'extended_properties': {'key': 'typeProperties.extendedProperties', 'type': '{object}'}, } - def __init__(self, *, name: str, command, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, resource_linked_service=None, folder_path=None, reference_objects=None, extended_properties=None, **kwargs) -> None: - super(CustomActivity, self).__init__(name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) + def __init__(self, *, name: str, command, additional_properties=None, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, resource_linked_service=None, folder_path=None, reference_objects=None, extended_properties=None, **kwargs) -> None: + super(CustomActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) self.command = command self.resource_linked_service = resource_linked_service self.folder_path = folder_path diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_lake_analytics_usql_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_lake_analytics_usql_activity.py index 8cd09b705857..364dfd79d71a 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_lake_analytics_usql_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_lake_analytics_usql_activity.py @@ -17,6 +17,9 @@ class DataLakeAnalyticsUSQLActivity(ExecutionActivity): All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -66,6 +69,7 @@ class DataLakeAnalyticsUSQLActivity(ExecutionActivity): } _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_lake_analytics_usql_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_lake_analytics_usql_activity_py3.py index 25242ae40705..22623aa3622c 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_lake_analytics_usql_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_lake_analytics_usql_activity_py3.py @@ -17,6 +17,9 @@ class DataLakeAnalyticsUSQLActivity(ExecutionActivity): All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -66,6 +69,7 @@ class DataLakeAnalyticsUSQLActivity(ExecutionActivity): } _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, @@ -82,8 +86,8 @@ class DataLakeAnalyticsUSQLActivity(ExecutionActivity): 'compilation_mode': {'key': 'typeProperties.compilationMode', 'type': 'object'}, } - def __init__(self, *, name: str, script_path, script_linked_service, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, degree_of_parallelism=None, priority=None, parameters=None, runtime_version=None, compilation_mode=None, **kwargs) -> None: - super(DataLakeAnalyticsUSQLActivity, self).__init__(name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) + def __init__(self, *, name: str, script_path, script_linked_service, additional_properties=None, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, degree_of_parallelism=None, priority=None, parameters=None, runtime_version=None, compilation_mode=None, **kwargs) -> None: + super(DataLakeAnalyticsUSQLActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) self.script_path = script_path self.script_linked_service = script_linked_service self.degree_of_parallelism = degree_of_parallelism diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_notebook_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_notebook_activity.py index 9d61043c2a3a..a49bd973e2b9 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_notebook_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_notebook_activity.py @@ -17,6 +17,9 @@ class DatabricksNotebookActivity(ExecutionActivity): All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -52,6 +55,7 @@ class DatabricksNotebookActivity(ExecutionActivity): } _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_notebook_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_notebook_activity_py3.py index 0cd5e6bdb25d..7d2d464b7a1a 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_notebook_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_notebook_activity_py3.py @@ -17,6 +17,9 @@ class DatabricksNotebookActivity(ExecutionActivity): All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -52,6 +55,7 @@ class DatabricksNotebookActivity(ExecutionActivity): } _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, @@ -64,8 +68,8 @@ class DatabricksNotebookActivity(ExecutionActivity): 'libraries': {'key': 'typeProperties.libraries', 'type': '[{object}]'}, } - def __init__(self, *, name: str, notebook_path, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, base_parameters=None, libraries=None, **kwargs) -> None: - super(DatabricksNotebookActivity, self).__init__(name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) + def __init__(self, *, name: str, notebook_path, additional_properties=None, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, base_parameters=None, libraries=None, **kwargs) -> None: + super(DatabricksNotebookActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) self.notebook_path = notebook_path self.base_parameters = base_parameters self.libraries = libraries diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_spark_jar_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_spark_jar_activity.py index 42d12eba7117..51e7245d12fe 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_spark_jar_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_spark_jar_activity.py @@ -17,6 +17,9 @@ class DatabricksSparkJarActivity(ExecutionActivity): All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -51,6 +54,7 @@ class DatabricksSparkJarActivity(ExecutionActivity): } _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_spark_jar_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_spark_jar_activity_py3.py index 2c88c324bbf0..6c33f3b51d1e 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_spark_jar_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_spark_jar_activity_py3.py @@ -17,6 +17,9 @@ class DatabricksSparkJarActivity(ExecutionActivity): All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -51,6 +54,7 @@ class DatabricksSparkJarActivity(ExecutionActivity): } _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, @@ -63,8 +67,8 @@ class DatabricksSparkJarActivity(ExecutionActivity): 'libraries': {'key': 'typeProperties.libraries', 'type': '[{object}]'}, } - def __init__(self, *, name: str, main_class_name, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, parameters=None, libraries=None, **kwargs) -> None: - super(DatabricksSparkJarActivity, self).__init__(name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) + def __init__(self, *, name: str, main_class_name, additional_properties=None, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, parameters=None, libraries=None, **kwargs) -> None: + super(DatabricksSparkJarActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) self.main_class_name = main_class_name self.parameters = parameters self.libraries = libraries diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_spark_python_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_spark_python_activity.py index 9cbb17a7ddf7..56178d3882c5 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_spark_python_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_spark_python_activity.py @@ -17,6 +17,9 @@ class DatabricksSparkPythonActivity(ExecutionActivity): All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -51,6 +54,7 @@ class DatabricksSparkPythonActivity(ExecutionActivity): } _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_spark_python_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_spark_python_activity_py3.py index 7090ee70681c..5b16d0d5e9ef 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_spark_python_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/databricks_spark_python_activity_py3.py @@ -17,6 +17,9 @@ class DatabricksSparkPythonActivity(ExecutionActivity): All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -51,6 +54,7 @@ class DatabricksSparkPythonActivity(ExecutionActivity): } _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, @@ -63,8 +67,8 @@ class DatabricksSparkPythonActivity(ExecutionActivity): 'libraries': {'key': 'typeProperties.libraries', 'type': '[{object}]'}, } - def __init__(self, *, name: str, python_file, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, parameters=None, libraries=None, **kwargs) -> None: - super(DatabricksSparkPythonActivity, self).__init__(name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) + def __init__(self, *, name: str, python_file, additional_properties=None, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, parameters=None, libraries=None, **kwargs) -> None: + super(DatabricksSparkPythonActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) self.python_file = python_file self.parameters = parameters self.libraries = libraries diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_pipeline_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_pipeline_activity.py index 24ffefbfbb1e..0008b5eee153 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_pipeline_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_pipeline_activity.py @@ -17,6 +17,9 @@ class ExecutePipelineActivity(ControlActivity): All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -43,6 +46,7 @@ class ExecutePipelineActivity(ControlActivity): } _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_pipeline_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_pipeline_activity_py3.py index 3bf9865f9f71..addaafabe7b0 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_pipeline_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_pipeline_activity_py3.py @@ -17,6 +17,9 @@ class ExecutePipelineActivity(ControlActivity): All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -43,6 +46,7 @@ class ExecutePipelineActivity(ControlActivity): } _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, @@ -53,8 +57,8 @@ class ExecutePipelineActivity(ControlActivity): 'wait_on_completion': {'key': 'typeProperties.waitOnCompletion', 'type': 'bool'}, } - def __init__(self, *, name: str, pipeline, description: str=None, depends_on=None, user_properties=None, parameters=None, wait_on_completion: bool=None, **kwargs) -> None: - super(ExecutePipelineActivity, self).__init__(name=name, description=description, depends_on=depends_on, user_properties=user_properties, **kwargs) + def __init__(self, *, name: str, pipeline, additional_properties=None, description: str=None, depends_on=None, user_properties=None, parameters=None, wait_on_completion: bool=None, **kwargs) -> None: + super(ExecutePipelineActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, **kwargs) self.pipeline = pipeline self.parameters = parameters self.wait_on_completion = wait_on_completion diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_ssis_package_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_ssis_package_activity.py index 6df95be62e2b..3491cbe93982 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_ssis_package_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_ssis_package_activity.py @@ -17,6 +17,9 @@ class ExecuteSSISPackageActivity(ExecutionActivity): All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -74,6 +77,7 @@ class ExecuteSSISPackageActivity(ExecutionActivity): } _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_ssis_package_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_ssis_package_activity_py3.py index d1cbdf2a66d2..b00419f52c55 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_ssis_package_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_ssis_package_activity_py3.py @@ -17,6 +17,9 @@ class ExecuteSSISPackageActivity(ExecutionActivity): All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -74,6 +77,7 @@ class ExecuteSSISPackageActivity(ExecutionActivity): } _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, @@ -93,8 +97,8 @@ class ExecuteSSISPackageActivity(ExecutionActivity): 'property_overrides': {'key': 'typeProperties.propertyOverrides', 'type': '{SSISPropertyOverride}'}, } - def __init__(self, *, name: str, package_location, connect_via, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, runtime=None, logging_level: str=None, environment_path: str=None, project_parameters=None, package_parameters=None, project_connection_managers=None, package_connection_managers=None, property_overrides=None, **kwargs) -> None: - super(ExecuteSSISPackageActivity, self).__init__(name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) + def __init__(self, *, name: str, package_location, connect_via, additional_properties=None, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, runtime=None, logging_level: str=None, environment_path: str=None, project_parameters=None, package_parameters=None, project_connection_managers=None, package_connection_managers=None, property_overrides=None, **kwargs) -> None: + super(ExecuteSSISPackageActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) self.package_location = package_location self.runtime = runtime self.logging_level = logging_level diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execution_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execution_activity.py index dcd134d880fd..a9362afc43a9 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execution_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execution_activity.py @@ -27,6 +27,9 @@ class ExecutionActivity(Activity): All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -50,6 +53,7 @@ class ExecutionActivity(Activity): } _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execution_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execution_activity_py3.py index 299e327a7d11..8762d742eb93 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execution_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execution_activity_py3.py @@ -27,6 +27,9 @@ class ExecutionActivity(Activity): All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -50,6 +53,7 @@ class ExecutionActivity(Activity): } _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, @@ -63,8 +67,8 @@ class ExecutionActivity(Activity): 'type': {'DatabricksSparkPython': 'DatabricksSparkPythonActivity', 'DatabricksSparkJar': 'DatabricksSparkJarActivity', 'DatabricksNotebook': 'DatabricksNotebookActivity', 'DataLakeAnalyticsU-SQL': 'DataLakeAnalyticsUSQLActivity', 'AzureMLUpdateResource': 'AzureMLUpdateResourceActivity', 'AzureMLBatchExecution': 'AzureMLBatchExecutionActivity', 'GetMetadata': 'GetMetadataActivity', 'WebActivity': 'WebActivity', 'Lookup': 'LookupActivity', 'SqlServerStoredProcedure': 'SqlServerStoredProcedureActivity', 'Custom': 'CustomActivity', 'ExecuteSSISPackage': 'ExecuteSSISPackageActivity', 'HDInsightSpark': 'HDInsightSparkActivity', 'HDInsightStreaming': 'HDInsightStreamingActivity', 'HDInsightMapReduce': 'HDInsightMapReduceActivity', 'HDInsightPig': 'HDInsightPigActivity', 'HDInsightHive': 'HDInsightHiveActivity', 'Copy': 'CopyActivity'} } - def __init__(self, *, name: str, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, **kwargs) -> None: - super(ExecutionActivity, self).__init__(name=name, description=description, depends_on=depends_on, user_properties=user_properties, **kwargs) + def __init__(self, *, name: str, additional_properties=None, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, **kwargs) -> None: + super(ExecutionActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, **kwargs) self.linked_service_name = linked_service_name self.policy = policy self.type = 'Execution' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/filter_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/filter_activity.py index d8fdaf2d47b3..1346bb234695 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/filter_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/filter_activity.py @@ -17,6 +17,9 @@ class FilterActivity(ControlActivity): All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -41,6 +44,7 @@ class FilterActivity(ControlActivity): } _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/filter_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/filter_activity_py3.py index 5756334b35d5..a07cf01d1dd5 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/filter_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/filter_activity_py3.py @@ -17,6 +17,9 @@ class FilterActivity(ControlActivity): All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -41,6 +44,7 @@ class FilterActivity(ControlActivity): } _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, @@ -50,8 +54,8 @@ class FilterActivity(ControlActivity): 'condition': {'key': 'typeProperties.condition', 'type': 'Expression'}, } - def __init__(self, *, name: str, items, condition, description: str=None, depends_on=None, user_properties=None, **kwargs) -> None: - super(FilterActivity, self).__init__(name=name, description=description, depends_on=depends_on, user_properties=user_properties, **kwargs) + def __init__(self, *, name: str, items, condition, additional_properties=None, description: str=None, depends_on=None, user_properties=None, **kwargs) -> None: + super(FilterActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, **kwargs) self.items = items self.condition = condition self.type = 'Filter' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/for_each_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/for_each_activity.py index 25f30e46dad7..5edfa2a8140e 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/for_each_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/for_each_activity.py @@ -18,6 +18,9 @@ class ForEachActivity(ControlActivity): All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -49,6 +52,7 @@ class ForEachActivity(ControlActivity): } _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/for_each_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/for_each_activity_py3.py index 87ae2a2e73fd..7c5c887bb1d9 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/for_each_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/for_each_activity_py3.py @@ -18,6 +18,9 @@ class ForEachActivity(ControlActivity): All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -49,6 +52,7 @@ class ForEachActivity(ControlActivity): } _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, @@ -60,8 +64,8 @@ class ForEachActivity(ControlActivity): 'activities': {'key': 'typeProperties.activities', 'type': '[Activity]'}, } - def __init__(self, *, name: str, items, activities, description: str=None, depends_on=None, user_properties=None, is_sequential: bool=None, batch_count: int=None, **kwargs) -> None: - super(ForEachActivity, self).__init__(name=name, description=description, depends_on=depends_on, user_properties=user_properties, **kwargs) + def __init__(self, *, name: str, items, activities, additional_properties=None, description: str=None, depends_on=None, user_properties=None, is_sequential: bool=None, batch_count: int=None, **kwargs) -> None: + super(ForEachActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, **kwargs) self.is_sequential = is_sequential self.batch_count = batch_count self.items = items diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/get_metadata_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/get_metadata_activity.py index 187cad77ecff..7941189f2dcd 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/get_metadata_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/get_metadata_activity.py @@ -17,6 +17,9 @@ class GetMetadataActivity(ExecutionActivity): All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -45,6 +48,7 @@ class GetMetadataActivity(ExecutionActivity): } _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/get_metadata_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/get_metadata_activity_py3.py index 215a26397f7d..b4d8eb17cab1 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/get_metadata_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/get_metadata_activity_py3.py @@ -17,6 +17,9 @@ class GetMetadataActivity(ExecutionActivity): All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -45,6 +48,7 @@ class GetMetadataActivity(ExecutionActivity): } _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, @@ -56,8 +60,8 @@ class GetMetadataActivity(ExecutionActivity): 'field_list': {'key': 'typeProperties.fieldList', 'type': '[object]'}, } - def __init__(self, *, name: str, dataset, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, field_list=None, **kwargs) -> None: - super(GetMetadataActivity, self).__init__(name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) + def __init__(self, *, name: str, dataset, additional_properties=None, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, field_list=None, **kwargs) -> None: + super(GetMetadataActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) self.dataset = dataset self.field_list = field_list self.type = 'GetMetadata' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_hive_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_hive_activity.py index a3a4117608b4..4944c4ceff75 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_hive_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_hive_activity.py @@ -17,6 +17,9 @@ class HDInsightHiveActivity(ExecutionActivity): All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -57,6 +60,7 @@ class HDInsightHiveActivity(ExecutionActivity): } _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_hive_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_hive_activity_py3.py index d78931ee19b5..748c1d5fc9dc 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_hive_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_hive_activity_py3.py @@ -17,6 +17,9 @@ class HDInsightHiveActivity(ExecutionActivity): All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -57,6 +60,7 @@ class HDInsightHiveActivity(ExecutionActivity): } _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, @@ -72,8 +76,8 @@ class HDInsightHiveActivity(ExecutionActivity): 'defines': {'key': 'typeProperties.defines', 'type': '{object}'}, } - def __init__(self, *, name: str, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, storage_linked_services=None, arguments=None, get_debug_info=None, script_path=None, script_linked_service=None, defines=None, **kwargs) -> None: - super(HDInsightHiveActivity, self).__init__(name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) + def __init__(self, *, name: str, additional_properties=None, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, storage_linked_services=None, arguments=None, get_debug_info=None, script_path=None, script_linked_service=None, defines=None, **kwargs) -> None: + super(HDInsightHiveActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) self.storage_linked_services = storage_linked_services self.arguments = arguments self.get_debug_info = get_debug_info diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_map_reduce_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_map_reduce_activity.py index 11f23563c2ca..20655843e1db 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_map_reduce_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_map_reduce_activity.py @@ -17,6 +17,9 @@ class HDInsightMapReduceActivity(ExecutionActivity): All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -65,6 +68,7 @@ class HDInsightMapReduceActivity(ExecutionActivity): } _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_map_reduce_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_map_reduce_activity_py3.py index 9d5626490df3..dffa9f119069 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_map_reduce_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_map_reduce_activity_py3.py @@ -17,6 +17,9 @@ class HDInsightMapReduceActivity(ExecutionActivity): All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -65,6 +68,7 @@ class HDInsightMapReduceActivity(ExecutionActivity): } _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, @@ -82,8 +86,8 @@ class HDInsightMapReduceActivity(ExecutionActivity): 'defines': {'key': 'typeProperties.defines', 'type': '{object}'}, } - def __init__(self, *, name: str, class_name, jar_file_path, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, storage_linked_services=None, arguments=None, get_debug_info=None, jar_linked_service=None, jar_libs=None, defines=None, **kwargs) -> None: - super(HDInsightMapReduceActivity, self).__init__(name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) + def __init__(self, *, name: str, class_name, jar_file_path, additional_properties=None, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, storage_linked_services=None, arguments=None, get_debug_info=None, jar_linked_service=None, jar_libs=None, defines=None, **kwargs) -> None: + super(HDInsightMapReduceActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) self.storage_linked_services = storage_linked_services self.arguments = arguments self.get_debug_info = get_debug_info diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_pig_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_pig_activity.py index 2a68916a36fe..61b939076db6 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_pig_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_pig_activity.py @@ -17,6 +17,9 @@ class HDInsightPigActivity(ExecutionActivity): All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -57,6 +60,7 @@ class HDInsightPigActivity(ExecutionActivity): } _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_pig_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_pig_activity_py3.py index 137767256ed5..fb149df91f39 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_pig_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_pig_activity_py3.py @@ -17,6 +17,9 @@ class HDInsightPigActivity(ExecutionActivity): All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -57,6 +60,7 @@ class HDInsightPigActivity(ExecutionActivity): } _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, @@ -72,8 +76,8 @@ class HDInsightPigActivity(ExecutionActivity): 'defines': {'key': 'typeProperties.defines', 'type': '{object}'}, } - def __init__(self, *, name: str, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, storage_linked_services=None, arguments=None, get_debug_info=None, script_path=None, script_linked_service=None, defines=None, **kwargs) -> None: - super(HDInsightPigActivity, self).__init__(name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) + def __init__(self, *, name: str, additional_properties=None, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, storage_linked_services=None, arguments=None, get_debug_info=None, script_path=None, script_linked_service=None, defines=None, **kwargs) -> None: + super(HDInsightPigActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) self.storage_linked_services = storage_linked_services self.arguments = arguments self.get_debug_info = get_debug_info diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_spark_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_spark_activity.py index 76c9a359ae1b..7822344f012f 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_spark_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_spark_activity.py @@ -17,6 +17,9 @@ class HDInsightSparkActivity(ExecutionActivity): All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -66,6 +69,7 @@ class HDInsightSparkActivity(ExecutionActivity): } _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_spark_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_spark_activity_py3.py index 528bb71450d3..3f305901abb7 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_spark_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_spark_activity_py3.py @@ -17,6 +17,9 @@ class HDInsightSparkActivity(ExecutionActivity): All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -66,6 +69,7 @@ class HDInsightSparkActivity(ExecutionActivity): } _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, @@ -83,8 +87,8 @@ class HDInsightSparkActivity(ExecutionActivity): 'spark_config': {'key': 'typeProperties.sparkConfig', 'type': '{object}'}, } - def __init__(self, *, name: str, root_path, entry_file_path, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, arguments=None, get_debug_info=None, spark_job_linked_service=None, class_name: str=None, proxy_user=None, spark_config=None, **kwargs) -> None: - super(HDInsightSparkActivity, self).__init__(name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) + def __init__(self, *, name: str, root_path, entry_file_path, additional_properties=None, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, arguments=None, get_debug_info=None, spark_job_linked_service=None, class_name: str=None, proxy_user=None, spark_config=None, **kwargs) -> None: + super(HDInsightSparkActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) self.root_path = root_path self.entry_file_path = entry_file_path self.arguments = arguments diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_streaming_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_streaming_activity.py index 87e7fe5cbe3d..42146a5d6cc6 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_streaming_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_streaming_activity.py @@ -17,6 +17,9 @@ class HDInsightStreamingActivity(ExecutionActivity): All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -80,6 +83,7 @@ class HDInsightStreamingActivity(ExecutionActivity): } _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_streaming_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_streaming_activity_py3.py index 195903974063..2f5a301ff880 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_streaming_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_streaming_activity_py3.py @@ -17,6 +17,9 @@ class HDInsightStreamingActivity(ExecutionActivity): All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -80,6 +83,7 @@ class HDInsightStreamingActivity(ExecutionActivity): } _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, @@ -101,8 +105,8 @@ class HDInsightStreamingActivity(ExecutionActivity): 'defines': {'key': 'typeProperties.defines', 'type': '{object}'}, } - def __init__(self, *, name: str, mapper, reducer, input, output, file_paths, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, storage_linked_services=None, arguments=None, get_debug_info=None, file_linked_service=None, combiner=None, command_environment=None, defines=None, **kwargs) -> None: - super(HDInsightStreamingActivity, self).__init__(name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) + def __init__(self, *, name: str, mapper, reducer, input, output, file_paths, additional_properties=None, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, storage_linked_services=None, arguments=None, get_debug_info=None, file_linked_service=None, combiner=None, command_environment=None, defines=None, **kwargs) -> None: + super(HDInsightStreamingActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) self.storage_linked_services = storage_linked_services self.arguments = arguments self.get_debug_info = get_debug_info diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/if_condition_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/if_condition_activity.py index 9f210ac81682..a8cb1da690e1 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/if_condition_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/if_condition_activity.py @@ -19,6 +19,9 @@ class IfConditionActivity(ControlActivity): All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -50,6 +53,7 @@ class IfConditionActivity(ControlActivity): } _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/if_condition_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/if_condition_activity_py3.py index 404aaa8bef4d..7921a2602807 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/if_condition_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/if_condition_activity_py3.py @@ -19,6 +19,9 @@ class IfConditionActivity(ControlActivity): All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -50,6 +53,7 @@ class IfConditionActivity(ControlActivity): } _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, @@ -60,8 +64,8 @@ class IfConditionActivity(ControlActivity): 'if_false_activities': {'key': 'typeProperties.ifFalseActivities', 'type': '[Activity]'}, } - def __init__(self, *, name: str, expression, description: str=None, depends_on=None, user_properties=None, if_true_activities=None, if_false_activities=None, **kwargs) -> None: - super(IfConditionActivity, self).__init__(name=name, description=description, depends_on=depends_on, user_properties=user_properties, **kwargs) + def __init__(self, *, name: str, expression, additional_properties=None, description: str=None, depends_on=None, user_properties=None, if_true_activities=None, if_false_activities=None, **kwargs) -> None: + super(IfConditionActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, **kwargs) self.expression = expression self.if_true_activities = if_true_activities self.if_false_activities = if_false_activities diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/lookup_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/lookup_activity.py index 4573d9369246..62584b2f704a 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/lookup_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/lookup_activity.py @@ -17,6 +17,9 @@ class LookupActivity(ExecutionActivity): All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -50,6 +53,7 @@ class LookupActivity(ExecutionActivity): } _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/lookup_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/lookup_activity_py3.py index 22ba9d129585..41061675ebbe 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/lookup_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/lookup_activity_py3.py @@ -17,6 +17,9 @@ class LookupActivity(ExecutionActivity): All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -50,6 +53,7 @@ class LookupActivity(ExecutionActivity): } _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, @@ -62,8 +66,8 @@ class LookupActivity(ExecutionActivity): 'first_row_only': {'key': 'typeProperties.firstRowOnly', 'type': 'object'}, } - def __init__(self, *, name: str, source, dataset, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, first_row_only=None, **kwargs) -> None: - super(LookupActivity, self).__init__(name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) + def __init__(self, *, name: str, source, dataset, additional_properties=None, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, first_row_only=None, **kwargs) -> None: + super(LookupActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) self.source = source self.dataset = dataset self.first_row_only = first_row_only diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_stored_procedure_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_stored_procedure_activity.py index 163339b67ec9..6f31002f32d1 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_stored_procedure_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_stored_procedure_activity.py @@ -17,6 +17,9 @@ class SqlServerStoredProcedureActivity(ExecutionActivity): All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -48,6 +51,7 @@ class SqlServerStoredProcedureActivity(ExecutionActivity): } _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_stored_procedure_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_stored_procedure_activity_py3.py index db487bd47b9f..477f0c6c775c 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_stored_procedure_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_stored_procedure_activity_py3.py @@ -17,6 +17,9 @@ class SqlServerStoredProcedureActivity(ExecutionActivity): All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -48,6 +51,7 @@ class SqlServerStoredProcedureActivity(ExecutionActivity): } _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, @@ -59,8 +63,8 @@ class SqlServerStoredProcedureActivity(ExecutionActivity): 'stored_procedure_parameters': {'key': 'typeProperties.storedProcedureParameters', 'type': '{StoredProcedureParameter}'}, } - def __init__(self, *, name: str, stored_procedure_name, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, stored_procedure_parameters=None, **kwargs) -> None: - super(SqlServerStoredProcedureActivity, self).__init__(name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) + def __init__(self, *, name: str, stored_procedure_name, additional_properties=None, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, stored_procedure_parameters=None, **kwargs) -> None: + super(SqlServerStoredProcedureActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) self.stored_procedure_name = stored_procedure_name self.stored_procedure_parameters = stored_procedure_parameters self.type = 'SqlServerStoredProcedure' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/until_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/until_activity.py index f9e7472c93f3..eede36501d6c 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/until_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/until_activity.py @@ -18,6 +18,9 @@ class UntilActivity(ControlActivity): All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -50,6 +53,7 @@ class UntilActivity(ControlActivity): } _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/until_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/until_activity_py3.py index 795f8ed5a814..40c03ce18591 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/until_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/until_activity_py3.py @@ -18,6 +18,9 @@ class UntilActivity(ControlActivity): All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -50,6 +53,7 @@ class UntilActivity(ControlActivity): } _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, @@ -60,8 +64,8 @@ class UntilActivity(ControlActivity): 'activities': {'key': 'typeProperties.activities', 'type': '[Activity]'}, } - def __init__(self, *, name: str, expression, activities, description: str=None, depends_on=None, user_properties=None, timeout=None, **kwargs) -> None: - super(UntilActivity, self).__init__(name=name, description=description, depends_on=depends_on, user_properties=user_properties, **kwargs) + def __init__(self, *, name: str, expression, activities, additional_properties=None, description: str=None, depends_on=None, user_properties=None, timeout=None, **kwargs) -> None: + super(UntilActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, **kwargs) self.expression = expression self.timeout = timeout self.activities = activities diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/wait_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/wait_activity.py index 987abc3eea78..91f3decc7473 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/wait_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/wait_activity.py @@ -17,6 +17,9 @@ class WaitActivity(ControlActivity): All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -38,6 +41,7 @@ class WaitActivity(ControlActivity): } _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/wait_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/wait_activity_py3.py index 31cc91ca1f3b..ff85c9d16733 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/wait_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/wait_activity_py3.py @@ -17,6 +17,9 @@ class WaitActivity(ControlActivity): All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -38,6 +41,7 @@ class WaitActivity(ControlActivity): } _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, @@ -46,7 +50,7 @@ class WaitActivity(ControlActivity): 'wait_time_in_seconds': {'key': 'typeProperties.waitTimeInSeconds', 'type': 'int'}, } - def __init__(self, *, name: str, wait_time_in_seconds: int, description: str=None, depends_on=None, user_properties=None, **kwargs) -> None: - super(WaitActivity, self).__init__(name=name, description=description, depends_on=depends_on, user_properties=user_properties, **kwargs) + def __init__(self, *, name: str, wait_time_in_seconds: int, additional_properties=None, description: str=None, depends_on=None, user_properties=None, **kwargs) -> None: + super(WaitActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, **kwargs) self.wait_time_in_seconds = wait_time_in_seconds self.type = 'Wait' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_activity.py index 8d175329d949..70264719d52e 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_activity.py @@ -17,6 +17,9 @@ class WebActivity(ExecutionActivity): All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -66,6 +69,7 @@ class WebActivity(ExecutionActivity): } _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_activity_py3.py index 7d06e9a72884..9a64114a00c6 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_activity_py3.py @@ -17,6 +17,9 @@ class WebActivity(ExecutionActivity): All required parameters must be populated in order to send to Azure. + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] :param name: Required. Activity name. :type name: str :param description: Activity description. @@ -66,6 +69,7 @@ class WebActivity(ExecutionActivity): } _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, @@ -82,8 +86,8 @@ class WebActivity(ExecutionActivity): 'linked_services': {'key': 'typeProperties.linkedServices', 'type': '[LinkedServiceReference]'}, } - def __init__(self, *, name: str, method, url, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, headers=None, body=None, authentication=None, datasets=None, linked_services=None, **kwargs) -> None: - super(WebActivity, self).__init__(name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) + def __init__(self, *, name: str, method, url, additional_properties=None, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, headers=None, body=None, authentication=None, datasets=None, linked_services=None, **kwargs) -> None: + super(WebActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) self.method = method self.url = url self.headers = headers From 86ac7db9f04b4dd2b2a8c121032b53eda01dd812 Mon Sep 17 00:00:00 2001 From: Azure SDK for Python bot Date: Wed, 1 Aug 2018 11:01:07 -0700 Subject: [PATCH 14/54] Generated from d1ad1388e3f654de7cea6f0f62b951174b2fc578 (#3061) change encryptedCredential type to string --- .../azure/mgmt/datafactory/models/__init__.py | 6 ++ .../azure_blob_storage_linked_service.py | 91 +++++++++++++++++++ .../azure_blob_storage_linked_service_py3.py | 91 +++++++++++++++++++ .../models/azure_storage_linked_service.py | 4 +- .../azure_storage_linked_service_py3.py | 6 +- .../azure_table_storage_linked_service.py | 70 ++++++++++++++ .../azure_table_storage_linked_service_py3.py | 70 ++++++++++++++ .../models/factory_git_hub_configuration.py | 11 +-- .../factory_git_hub_configuration_py3.py | 13 +-- .../mgmt/datafactory/models/linked_service.py | 6 +- .../datafactory/models/linked_service_py3.py | 6 +- 11 files changed, 350 insertions(+), 24 deletions(-) create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_storage_linked_service.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_storage_linked_service_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_table_storage_linked_service.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_table_storage_linked_service_py3.py diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py index 2397ab63e4c5..4143ff1e6621 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py @@ -138,6 +138,8 @@ from .azure_sql_database_linked_service_py3 import AzureSqlDatabaseLinkedService from .sql_server_linked_service_py3 import SqlServerLinkedService from .azure_sql_dw_linked_service_py3 import AzureSqlDWLinkedService + from .azure_table_storage_linked_service_py3 import AzureTableStorageLinkedService + from .azure_blob_storage_linked_service_py3 import AzureBlobStorageLinkedService from .azure_storage_linked_service_py3 import AzureStorageLinkedService from .responsys_object_dataset_py3 import ResponsysObjectDataset from .salesforce_marketing_cloud_object_dataset_py3 import SalesforceMarketingCloudObjectDataset @@ -471,6 +473,8 @@ from .azure_sql_database_linked_service import AzureSqlDatabaseLinkedService from .sql_server_linked_service import SqlServerLinkedService from .azure_sql_dw_linked_service import AzureSqlDWLinkedService + from .azure_table_storage_linked_service import AzureTableStorageLinkedService + from .azure_blob_storage_linked_service import AzureBlobStorageLinkedService from .azure_storage_linked_service import AzureStorageLinkedService from .responsys_object_dataset import ResponsysObjectDataset from .salesforce_marketing_cloud_object_dataset import SalesforceMarketingCloudObjectDataset @@ -872,6 +876,8 @@ 'AzureSqlDatabaseLinkedService', 'SqlServerLinkedService', 'AzureSqlDWLinkedService', + 'AzureTableStorageLinkedService', + 'AzureBlobStorageLinkedService', 'AzureStorageLinkedService', 'ResponsysObjectDataset', 'SalesforceMarketingCloudObjectDataset', diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_storage_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_storage_linked_service.py new file mode 100644 index 000000000000..d9a129530f53 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_storage_linked_service.py @@ -0,0 +1,91 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class AzureBlobStorageLinkedService(LinkedService): + """The azure blob storage linked service. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param connection_string: The connection string. It is mutually exclusive + with sasUri, serviceEndpoint property. Type: string, SecureString or + AzureKeyVaultSecretReference. + :type connection_string: object + :param sas_uri: SAS URI of the Azure Blob Storage resource. It is mutually + exclusive with connectionString, serviceEndpoint property. + :type sas_uri: ~azure.mgmt.datafactory.models.SecretBase + :param service_endpoint: Blob service endpoint of the Azure Blob Storage + resource. It is mutually exclusive with connectionString, sasUri property. + :type service_endpoint: str + :param service_principal_id: The ID of the service principal used to + authenticate against Azure SQL Data Warehouse. Type: string (or Expression + with resultType string). + :type service_principal_id: object + :param service_principal_key: The key of the service principal used to + authenticate against Azure SQL Data Warehouse. + :type service_principal_key: ~azure.mgmt.datafactory.models.SecretBase + :param tenant: The name or ID of the tenant to which the service principal + belongs. Type: string (or Expression with resultType string). + :type tenant: object + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: str + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, + 'sas_uri': {'key': 'typeProperties.sasUri', 'type': 'SecretBase'}, + 'service_endpoint': {'key': 'typeProperties.serviceEndpoint', 'type': 'str'}, + 'service_principal_id': {'key': 'typeProperties.servicePrincipalId', 'type': 'object'}, + 'service_principal_key': {'key': 'typeProperties.servicePrincipalKey', 'type': 'SecretBase'}, + 'tenant': {'key': 'typeProperties.tenant', 'type': 'object'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'str'}, + } + + def __init__(self, **kwargs): + super(AzureBlobStorageLinkedService, self).__init__(**kwargs) + self.connection_string = kwargs.get('connection_string', None) + self.sas_uri = kwargs.get('sas_uri', None) + self.service_endpoint = kwargs.get('service_endpoint', None) + self.service_principal_id = kwargs.get('service_principal_id', None) + self.service_principal_key = kwargs.get('service_principal_key', None) + self.tenant = kwargs.get('tenant', None) + self.encrypted_credential = kwargs.get('encrypted_credential', None) + self.type = 'AzureBlobStorage' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_storage_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_storage_linked_service_py3.py new file mode 100644 index 000000000000..39efbab9e83e --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_storage_linked_service_py3.py @@ -0,0 +1,91 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service_py3 import LinkedService + + +class AzureBlobStorageLinkedService(LinkedService): + """The azure blob storage linked service. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param connection_string: The connection string. It is mutually exclusive + with sasUri, serviceEndpoint property. Type: string, SecureString or + AzureKeyVaultSecretReference. + :type connection_string: object + :param sas_uri: SAS URI of the Azure Blob Storage resource. It is mutually + exclusive with connectionString, serviceEndpoint property. + :type sas_uri: ~azure.mgmt.datafactory.models.SecretBase + :param service_endpoint: Blob service endpoint of the Azure Blob Storage + resource. It is mutually exclusive with connectionString, sasUri property. + :type service_endpoint: str + :param service_principal_id: The ID of the service principal used to + authenticate against Azure SQL Data Warehouse. Type: string (or Expression + with resultType string). + :type service_principal_id: object + :param service_principal_key: The key of the service principal used to + authenticate against Azure SQL Data Warehouse. + :type service_principal_key: ~azure.mgmt.datafactory.models.SecretBase + :param tenant: The name or ID of the tenant to which the service principal + belongs. Type: string (or Expression with resultType string). + :type tenant: object + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: str + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, + 'sas_uri': {'key': 'typeProperties.sasUri', 'type': 'SecretBase'}, + 'service_endpoint': {'key': 'typeProperties.serviceEndpoint', 'type': 'str'}, + 'service_principal_id': {'key': 'typeProperties.servicePrincipalId', 'type': 'object'}, + 'service_principal_key': {'key': 'typeProperties.servicePrincipalKey', 'type': 'SecretBase'}, + 'tenant': {'key': 'typeProperties.tenant', 'type': 'object'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'str'}, + } + + def __init__(self, *, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, connection_string=None, sas_uri=None, service_endpoint: str=None, service_principal_id=None, service_principal_key=None, tenant=None, encrypted_credential: str=None, **kwargs) -> None: + super(AzureBlobStorageLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.connection_string = connection_string + self.sas_uri = sas_uri + self.service_endpoint = service_endpoint + self.service_principal_id = service_principal_id + self.service_principal_key = service_principal_key + self.tenant = tenant + self.encrypted_credential = encrypted_credential + self.type = 'AzureBlobStorage' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_storage_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_storage_linked_service.py index d21da53dc341..d59dc43d5260 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_storage_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_storage_linked_service.py @@ -43,7 +43,7 @@ class AzureStorageLinkedService(LinkedService): :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). - :type encrypted_credential: object + :type encrypted_credential: str """ _validation = { @@ -59,7 +59,7 @@ class AzureStorageLinkedService(LinkedService): 'type': {'key': 'type', 'type': 'str'}, 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, 'sas_uri': {'key': 'typeProperties.sasUri', 'type': 'SecretBase'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'str'}, } def __init__(self, **kwargs): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_storage_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_storage_linked_service_py3.py index 6852af1d13ac..e508de34ee21 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_storage_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_storage_linked_service_py3.py @@ -43,7 +43,7 @@ class AzureStorageLinkedService(LinkedService): :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). - :type encrypted_credential: object + :type encrypted_credential: str """ _validation = { @@ -59,10 +59,10 @@ class AzureStorageLinkedService(LinkedService): 'type': {'key': 'type', 'type': 'str'}, 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, 'sas_uri': {'key': 'typeProperties.sasUri', 'type': 'SecretBase'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'str'}, } - def __init__(self, *, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, connection_string=None, sas_uri=None, encrypted_credential=None, **kwargs) -> None: + def __init__(self, *, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, connection_string=None, sas_uri=None, encrypted_credential: str=None, **kwargs) -> None: super(AzureStorageLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) self.connection_string = connection_string self.sas_uri = sas_uri diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_table_storage_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_table_storage_linked_service.py new file mode 100644 index 000000000000..8fdd6e339123 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_table_storage_linked_service.py @@ -0,0 +1,70 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class AzureTableStorageLinkedService(LinkedService): + """The azure table storage linked service. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param connection_string: The connection string. It is mutually exclusive + with sasUri property. Type: string, SecureString or + AzureKeyVaultSecretReference. + :type connection_string: object + :param sas_uri: SAS URI of the Azure Storage resource. It is mutually + exclusive with connectionString property. + :type sas_uri: ~azure.mgmt.datafactory.models.SecretBase + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: str + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, + 'sas_uri': {'key': 'typeProperties.sasUri', 'type': 'SecretBase'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'str'}, + } + + def __init__(self, **kwargs): + super(AzureTableStorageLinkedService, self).__init__(**kwargs) + self.connection_string = kwargs.get('connection_string', None) + self.sas_uri = kwargs.get('sas_uri', None) + self.encrypted_credential = kwargs.get('encrypted_credential', None) + self.type = 'AzureTableStorage' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_table_storage_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_table_storage_linked_service_py3.py new file mode 100644 index 000000000000..2a1f3df9222d --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_table_storage_linked_service_py3.py @@ -0,0 +1,70 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service_py3 import LinkedService + + +class AzureTableStorageLinkedService(LinkedService): + """The azure table storage linked service. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param connection_string: The connection string. It is mutually exclusive + with sasUri property. Type: string, SecureString or + AzureKeyVaultSecretReference. + :type connection_string: object + :param sas_uri: SAS URI of the Azure Storage resource. It is mutually + exclusive with connectionString property. + :type sas_uri: ~azure.mgmt.datafactory.models.SecretBase + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: str + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, + 'sas_uri': {'key': 'typeProperties.sasUri', 'type': 'SecretBase'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'str'}, + } + + def __init__(self, *, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, connection_string=None, sas_uri=None, encrypted_credential: str=None, **kwargs) -> None: + super(AzureTableStorageLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.connection_string = connection_string + self.sas_uri = sas_uri + self.encrypted_credential = encrypted_credential + self.type = 'AzureTableStorage' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_git_hub_configuration.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_git_hub_configuration.py index 2c3f9b14840e..f870e4c10229 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_git_hub_configuration.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_git_hub_configuration.py @@ -15,9 +15,6 @@ class FactoryGitHubConfiguration(FactoryRepoConfiguration): """Factory's GitHub repo information. - Variables are only populated by the server, and will be ignored when - sending a request. - All required parameters must be populated in order to send to Azure. :param account_name: Required. Account name. @@ -32,8 +29,9 @@ class FactoryGitHubConfiguration(FactoryRepoConfiguration): :type last_commit_id: str :param type: Required. Constant filled by server. :type type: str - :ivar host_name: GitHub repo host name. - :vartype host_name: str + :param host_name: GitHub Enterprise host name. For example: + https://github.mydomain.com + :type host_name: str """ _validation = { @@ -42,7 +40,6 @@ class FactoryGitHubConfiguration(FactoryRepoConfiguration): 'collaboration_branch': {'required': True}, 'root_folder': {'required': True}, 'type': {'required': True}, - 'host_name': {'readonly': True}, } _attribute_map = { @@ -57,5 +54,5 @@ class FactoryGitHubConfiguration(FactoryRepoConfiguration): def __init__(self, **kwargs): super(FactoryGitHubConfiguration, self).__init__(**kwargs) - self.host_name = None + self.host_name = kwargs.get('host_name', None) self.type = 'FactoryGithubConfiguration' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_git_hub_configuration_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_git_hub_configuration_py3.py index 1a71ff9b4476..7f9b2f210f83 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_git_hub_configuration_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_git_hub_configuration_py3.py @@ -15,9 +15,6 @@ class FactoryGitHubConfiguration(FactoryRepoConfiguration): """Factory's GitHub repo information. - Variables are only populated by the server, and will be ignored when - sending a request. - All required parameters must be populated in order to send to Azure. :param account_name: Required. Account name. @@ -32,8 +29,9 @@ class FactoryGitHubConfiguration(FactoryRepoConfiguration): :type last_commit_id: str :param type: Required. Constant filled by server. :type type: str - :ivar host_name: GitHub repo host name. - :vartype host_name: str + :param host_name: GitHub Enterprise host name. For example: + https://github.mydomain.com + :type host_name: str """ _validation = { @@ -42,7 +40,6 @@ class FactoryGitHubConfiguration(FactoryRepoConfiguration): 'collaboration_branch': {'required': True}, 'root_folder': {'required': True}, 'type': {'required': True}, - 'host_name': {'readonly': True}, } _attribute_map = { @@ -55,7 +52,7 @@ class FactoryGitHubConfiguration(FactoryRepoConfiguration): 'host_name': {'key': 'hostName', 'type': 'str'}, } - def __init__(self, *, account_name: str, repository_name: str, collaboration_branch: str, root_folder: str, last_commit_id: str=None, **kwargs) -> None: + def __init__(self, *, account_name: str, repository_name: str, collaboration_branch: str, root_folder: str, last_commit_id: str=None, host_name: str=None, **kwargs) -> None: super(FactoryGitHubConfiguration, self).__init__(account_name=account_name, repository_name=repository_name, collaboration_branch=collaboration_branch, root_folder=root_folder, last_commit_id=last_commit_id, **kwargs) - self.host_name = None + self.host_name = host_name self.type = 'FactoryGithubConfiguration' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service.py index ecb6b42e145a..3dec3c0dcc24 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service.py @@ -43,7 +43,9 @@ class LinkedService(Model): OracleLinkedService, FileServerLinkedService, HDInsightLinkedService, DynamicsLinkedService, CosmosDbLinkedService, AzureKeyVaultLinkedService, AzureBatchLinkedService, AzureSqlDatabaseLinkedService, - SqlServerLinkedService, AzureSqlDWLinkedService, AzureStorageLinkedService + SqlServerLinkedService, AzureSqlDWLinkedService, + AzureTableStorageLinkedService, AzureBlobStorageLinkedService, + AzureStorageLinkedService All required parameters must be populated in order to send to Azure. @@ -79,7 +81,7 @@ class LinkedService(Model): } _subtype_map = { - 'type': {'Responsys': 'ResponsysLinkedService', 'AzureDatabricks': 'AzureDatabricksLinkedService', 'AzureDataLakeAnalytics': 'AzureDataLakeAnalyticsLinkedService', 'HDInsightOnDemand': 'HDInsightOnDemandLinkedService', 'SalesforceMarketingCloud': 'SalesforceMarketingCloudLinkedService', 'Netezza': 'NetezzaLinkedService', 'Vertica': 'VerticaLinkedService', 'Zoho': 'ZohoLinkedService', 'Xero': 'XeroLinkedService', 'Square': 'SquareLinkedService', 'Spark': 'SparkLinkedService', 'Shopify': 'ShopifyLinkedService', 'ServiceNow': 'ServiceNowLinkedService', 'QuickBooks': 'QuickBooksLinkedService', 'Presto': 'PrestoLinkedService', 'Phoenix': 'PhoenixLinkedService', 'Paypal': 'PaypalLinkedService', 'Marketo': 'MarketoLinkedService', 'MariaDB': 'MariaDBLinkedService', 'Magento': 'MagentoLinkedService', 'Jira': 'JiraLinkedService', 'Impala': 'ImpalaLinkedService', 'Hubspot': 'HubspotLinkedService', 'Hive': 'HiveLinkedService', 'HBase': 'HBaseLinkedService', 'Greenplum': 'GreenplumLinkedService', 'GoogleBigQuery': 'GoogleBigQueryLinkedService', 'Eloqua': 'EloquaLinkedService', 'Drill': 'DrillLinkedService', 'Couchbase': 'CouchbaseLinkedService', 'Concur': 'ConcurLinkedService', 'AzurePostgreSql': 'AzurePostgreSqlLinkedService', 'AmazonMWS': 'AmazonMWSLinkedService', 'SapHana': 'SapHanaLinkedService', 'SapBW': 'SapBWLinkedService', 'Sftp': 'SftpServerLinkedService', 'FtpServer': 'FtpServerLinkedService', 'HttpServer': 'HttpLinkedService', 'AzureSearch': 'AzureSearchLinkedService', 'CustomDataSource': 'CustomDataSourceLinkedService', 'AmazonRedshift': 'AmazonRedshiftLinkedService', 'AmazonS3': 'AmazonS3LinkedService', 'SapEcc': 'SapEccLinkedService', 'SapCloudForCustomer': 'SapCloudForCustomerLinkedService', 'Salesforce': 'SalesforceLinkedService', 'AzureDataLakeStore': 'AzureDataLakeStoreLinkedService', 'MongoDb': 'MongoDbLinkedService', 'Cassandra': 'CassandraLinkedService', 'Web': 'WebLinkedService', 'OData': 'ODataLinkedService', 'Hdfs': 'HdfsLinkedService', 'Odbc': 'OdbcLinkedService', 'AzureML': 'AzureMLLinkedService', 'Teradata': 'TeradataLinkedService', 'Db2': 'Db2LinkedService', 'Sybase': 'SybaseLinkedService', 'PostgreSql': 'PostgreSqlLinkedService', 'MySql': 'MySqlLinkedService', 'AzureMySql': 'AzureMySqlLinkedService', 'Oracle': 'OracleLinkedService', 'FileServer': 'FileServerLinkedService', 'HDInsight': 'HDInsightLinkedService', 'Dynamics': 'DynamicsLinkedService', 'CosmosDb': 'CosmosDbLinkedService', 'AzureKeyVault': 'AzureKeyVaultLinkedService', 'AzureBatch': 'AzureBatchLinkedService', 'AzureSqlDatabase': 'AzureSqlDatabaseLinkedService', 'SqlServer': 'SqlServerLinkedService', 'AzureSqlDW': 'AzureSqlDWLinkedService', 'AzureStorage': 'AzureStorageLinkedService'} + 'type': {'Responsys': 'ResponsysLinkedService', 'AzureDatabricks': 'AzureDatabricksLinkedService', 'AzureDataLakeAnalytics': 'AzureDataLakeAnalyticsLinkedService', 'HDInsightOnDemand': 'HDInsightOnDemandLinkedService', 'SalesforceMarketingCloud': 'SalesforceMarketingCloudLinkedService', 'Netezza': 'NetezzaLinkedService', 'Vertica': 'VerticaLinkedService', 'Zoho': 'ZohoLinkedService', 'Xero': 'XeroLinkedService', 'Square': 'SquareLinkedService', 'Spark': 'SparkLinkedService', 'Shopify': 'ShopifyLinkedService', 'ServiceNow': 'ServiceNowLinkedService', 'QuickBooks': 'QuickBooksLinkedService', 'Presto': 'PrestoLinkedService', 'Phoenix': 'PhoenixLinkedService', 'Paypal': 'PaypalLinkedService', 'Marketo': 'MarketoLinkedService', 'MariaDB': 'MariaDBLinkedService', 'Magento': 'MagentoLinkedService', 'Jira': 'JiraLinkedService', 'Impala': 'ImpalaLinkedService', 'Hubspot': 'HubspotLinkedService', 'Hive': 'HiveLinkedService', 'HBase': 'HBaseLinkedService', 'Greenplum': 'GreenplumLinkedService', 'GoogleBigQuery': 'GoogleBigQueryLinkedService', 'Eloqua': 'EloquaLinkedService', 'Drill': 'DrillLinkedService', 'Couchbase': 'CouchbaseLinkedService', 'Concur': 'ConcurLinkedService', 'AzurePostgreSql': 'AzurePostgreSqlLinkedService', 'AmazonMWS': 'AmazonMWSLinkedService', 'SapHana': 'SapHanaLinkedService', 'SapBW': 'SapBWLinkedService', 'Sftp': 'SftpServerLinkedService', 'FtpServer': 'FtpServerLinkedService', 'HttpServer': 'HttpLinkedService', 'AzureSearch': 'AzureSearchLinkedService', 'CustomDataSource': 'CustomDataSourceLinkedService', 'AmazonRedshift': 'AmazonRedshiftLinkedService', 'AmazonS3': 'AmazonS3LinkedService', 'SapEcc': 'SapEccLinkedService', 'SapCloudForCustomer': 'SapCloudForCustomerLinkedService', 'Salesforce': 'SalesforceLinkedService', 'AzureDataLakeStore': 'AzureDataLakeStoreLinkedService', 'MongoDb': 'MongoDbLinkedService', 'Cassandra': 'CassandraLinkedService', 'Web': 'WebLinkedService', 'OData': 'ODataLinkedService', 'Hdfs': 'HdfsLinkedService', 'Odbc': 'OdbcLinkedService', 'AzureML': 'AzureMLLinkedService', 'Teradata': 'TeradataLinkedService', 'Db2': 'Db2LinkedService', 'Sybase': 'SybaseLinkedService', 'PostgreSql': 'PostgreSqlLinkedService', 'MySql': 'MySqlLinkedService', 'AzureMySql': 'AzureMySqlLinkedService', 'Oracle': 'OracleLinkedService', 'FileServer': 'FileServerLinkedService', 'HDInsight': 'HDInsightLinkedService', 'Dynamics': 'DynamicsLinkedService', 'CosmosDb': 'CosmosDbLinkedService', 'AzureKeyVault': 'AzureKeyVaultLinkedService', 'AzureBatch': 'AzureBatchLinkedService', 'AzureSqlDatabase': 'AzureSqlDatabaseLinkedService', 'SqlServer': 'SqlServerLinkedService', 'AzureSqlDW': 'AzureSqlDWLinkedService', 'AzureTableStorage': 'AzureTableStorageLinkedService', 'AzureBlobStorage': 'AzureBlobStorageLinkedService', 'AzureStorage': 'AzureStorageLinkedService'} } def __init__(self, **kwargs): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service_py3.py index 39119635cba5..885af520bcda 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service_py3.py @@ -43,7 +43,9 @@ class LinkedService(Model): OracleLinkedService, FileServerLinkedService, HDInsightLinkedService, DynamicsLinkedService, CosmosDbLinkedService, AzureKeyVaultLinkedService, AzureBatchLinkedService, AzureSqlDatabaseLinkedService, - SqlServerLinkedService, AzureSqlDWLinkedService, AzureStorageLinkedService + SqlServerLinkedService, AzureSqlDWLinkedService, + AzureTableStorageLinkedService, AzureBlobStorageLinkedService, + AzureStorageLinkedService All required parameters must be populated in order to send to Azure. @@ -79,7 +81,7 @@ class LinkedService(Model): } _subtype_map = { - 'type': {'Responsys': 'ResponsysLinkedService', 'AzureDatabricks': 'AzureDatabricksLinkedService', 'AzureDataLakeAnalytics': 'AzureDataLakeAnalyticsLinkedService', 'HDInsightOnDemand': 'HDInsightOnDemandLinkedService', 'SalesforceMarketingCloud': 'SalesforceMarketingCloudLinkedService', 'Netezza': 'NetezzaLinkedService', 'Vertica': 'VerticaLinkedService', 'Zoho': 'ZohoLinkedService', 'Xero': 'XeroLinkedService', 'Square': 'SquareLinkedService', 'Spark': 'SparkLinkedService', 'Shopify': 'ShopifyLinkedService', 'ServiceNow': 'ServiceNowLinkedService', 'QuickBooks': 'QuickBooksLinkedService', 'Presto': 'PrestoLinkedService', 'Phoenix': 'PhoenixLinkedService', 'Paypal': 'PaypalLinkedService', 'Marketo': 'MarketoLinkedService', 'MariaDB': 'MariaDBLinkedService', 'Magento': 'MagentoLinkedService', 'Jira': 'JiraLinkedService', 'Impala': 'ImpalaLinkedService', 'Hubspot': 'HubspotLinkedService', 'Hive': 'HiveLinkedService', 'HBase': 'HBaseLinkedService', 'Greenplum': 'GreenplumLinkedService', 'GoogleBigQuery': 'GoogleBigQueryLinkedService', 'Eloqua': 'EloquaLinkedService', 'Drill': 'DrillLinkedService', 'Couchbase': 'CouchbaseLinkedService', 'Concur': 'ConcurLinkedService', 'AzurePostgreSql': 'AzurePostgreSqlLinkedService', 'AmazonMWS': 'AmazonMWSLinkedService', 'SapHana': 'SapHanaLinkedService', 'SapBW': 'SapBWLinkedService', 'Sftp': 'SftpServerLinkedService', 'FtpServer': 'FtpServerLinkedService', 'HttpServer': 'HttpLinkedService', 'AzureSearch': 'AzureSearchLinkedService', 'CustomDataSource': 'CustomDataSourceLinkedService', 'AmazonRedshift': 'AmazonRedshiftLinkedService', 'AmazonS3': 'AmazonS3LinkedService', 'SapEcc': 'SapEccLinkedService', 'SapCloudForCustomer': 'SapCloudForCustomerLinkedService', 'Salesforce': 'SalesforceLinkedService', 'AzureDataLakeStore': 'AzureDataLakeStoreLinkedService', 'MongoDb': 'MongoDbLinkedService', 'Cassandra': 'CassandraLinkedService', 'Web': 'WebLinkedService', 'OData': 'ODataLinkedService', 'Hdfs': 'HdfsLinkedService', 'Odbc': 'OdbcLinkedService', 'AzureML': 'AzureMLLinkedService', 'Teradata': 'TeradataLinkedService', 'Db2': 'Db2LinkedService', 'Sybase': 'SybaseLinkedService', 'PostgreSql': 'PostgreSqlLinkedService', 'MySql': 'MySqlLinkedService', 'AzureMySql': 'AzureMySqlLinkedService', 'Oracle': 'OracleLinkedService', 'FileServer': 'FileServerLinkedService', 'HDInsight': 'HDInsightLinkedService', 'Dynamics': 'DynamicsLinkedService', 'CosmosDb': 'CosmosDbLinkedService', 'AzureKeyVault': 'AzureKeyVaultLinkedService', 'AzureBatch': 'AzureBatchLinkedService', 'AzureSqlDatabase': 'AzureSqlDatabaseLinkedService', 'SqlServer': 'SqlServerLinkedService', 'AzureSqlDW': 'AzureSqlDWLinkedService', 'AzureStorage': 'AzureStorageLinkedService'} + 'type': {'Responsys': 'ResponsysLinkedService', 'AzureDatabricks': 'AzureDatabricksLinkedService', 'AzureDataLakeAnalytics': 'AzureDataLakeAnalyticsLinkedService', 'HDInsightOnDemand': 'HDInsightOnDemandLinkedService', 'SalesforceMarketingCloud': 'SalesforceMarketingCloudLinkedService', 'Netezza': 'NetezzaLinkedService', 'Vertica': 'VerticaLinkedService', 'Zoho': 'ZohoLinkedService', 'Xero': 'XeroLinkedService', 'Square': 'SquareLinkedService', 'Spark': 'SparkLinkedService', 'Shopify': 'ShopifyLinkedService', 'ServiceNow': 'ServiceNowLinkedService', 'QuickBooks': 'QuickBooksLinkedService', 'Presto': 'PrestoLinkedService', 'Phoenix': 'PhoenixLinkedService', 'Paypal': 'PaypalLinkedService', 'Marketo': 'MarketoLinkedService', 'MariaDB': 'MariaDBLinkedService', 'Magento': 'MagentoLinkedService', 'Jira': 'JiraLinkedService', 'Impala': 'ImpalaLinkedService', 'Hubspot': 'HubspotLinkedService', 'Hive': 'HiveLinkedService', 'HBase': 'HBaseLinkedService', 'Greenplum': 'GreenplumLinkedService', 'GoogleBigQuery': 'GoogleBigQueryLinkedService', 'Eloqua': 'EloquaLinkedService', 'Drill': 'DrillLinkedService', 'Couchbase': 'CouchbaseLinkedService', 'Concur': 'ConcurLinkedService', 'AzurePostgreSql': 'AzurePostgreSqlLinkedService', 'AmazonMWS': 'AmazonMWSLinkedService', 'SapHana': 'SapHanaLinkedService', 'SapBW': 'SapBWLinkedService', 'Sftp': 'SftpServerLinkedService', 'FtpServer': 'FtpServerLinkedService', 'HttpServer': 'HttpLinkedService', 'AzureSearch': 'AzureSearchLinkedService', 'CustomDataSource': 'CustomDataSourceLinkedService', 'AmazonRedshift': 'AmazonRedshiftLinkedService', 'AmazonS3': 'AmazonS3LinkedService', 'SapEcc': 'SapEccLinkedService', 'SapCloudForCustomer': 'SapCloudForCustomerLinkedService', 'Salesforce': 'SalesforceLinkedService', 'AzureDataLakeStore': 'AzureDataLakeStoreLinkedService', 'MongoDb': 'MongoDbLinkedService', 'Cassandra': 'CassandraLinkedService', 'Web': 'WebLinkedService', 'OData': 'ODataLinkedService', 'Hdfs': 'HdfsLinkedService', 'Odbc': 'OdbcLinkedService', 'AzureML': 'AzureMLLinkedService', 'Teradata': 'TeradataLinkedService', 'Db2': 'Db2LinkedService', 'Sybase': 'SybaseLinkedService', 'PostgreSql': 'PostgreSqlLinkedService', 'MySql': 'MySqlLinkedService', 'AzureMySql': 'AzureMySqlLinkedService', 'Oracle': 'OracleLinkedService', 'FileServer': 'FileServerLinkedService', 'HDInsight': 'HDInsightLinkedService', 'Dynamics': 'DynamicsLinkedService', 'CosmosDb': 'CosmosDbLinkedService', 'AzureKeyVault': 'AzureKeyVaultLinkedService', 'AzureBatch': 'AzureBatchLinkedService', 'AzureSqlDatabase': 'AzureSqlDatabaseLinkedService', 'SqlServer': 'SqlServerLinkedService', 'AzureSqlDW': 'AzureSqlDWLinkedService', 'AzureTableStorage': 'AzureTableStorageLinkedService', 'AzureBlobStorage': 'AzureBlobStorageLinkedService', 'AzureStorage': 'AzureStorageLinkedService'} } def __init__(self, *, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, **kwargs) -> None: From 81ef8ac76c8c65a20aa85df16e0b0aa73b27d9b2 Mon Sep 17 00:00:00 2001 From: Azure SDK for Python bot Date: Wed, 22 Aug 2018 09:06:46 -0700 Subject: [PATCH 15/54] [AutoPR datafactory/resource-manager] Add two API-s to integration runtime (#3168) * Generated from ac46263d530ef85bae2d611e77ae63f2797897f4 Add two API-s to integration runtime * Generated from 43e82243c77750188200ef16df385cecd49ff652 Fixed model validation error for the new APIs * Generated from 5c8e79e5018783b5e2c0c119c72cd64dc0194fba Improve the properties description per to review comments --- .../azure/mgmt/datafactory/models/__init__.py | 3 + ...eate_linked_integration_runtime_request.py | 43 +++++ ..._linked_integration_runtime_request_py3.py | 43 +++++ .../data_factory_management_client_enums.py | 1 + .../self_hosted_integration_runtime_node.py | 2 +- ...elf_hosted_integration_runtime_node_py3.py | 2 +- .../self_hosted_integration_runtime_status.py | 6 + ...f_hosted_integration_runtime_status_py3.py | 6 + .../operations/activity_runs_operations.py | 6 +- .../operations/datasets_operations.py | 24 ++- .../operations/factories_operations.py | 43 +++-- .../integration_runtime_nodes_operations.py | 87 +++++++++- .../integration_runtimes_operations.py | 162 +++++++++++++----- .../operations/linked_services_operations.py | 24 ++- .../mgmt/datafactory/operations/operations.py | 7 +- .../operations/pipeline_runs_operations.py | 17 +- .../operations/pipelines_operations.py | 30 ++-- .../operations/trigger_runs_operations.py | 6 +- .../operations/triggers_operations.py | 34 ++-- 19 files changed, 386 insertions(+), 160 deletions(-) create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/create_linked_integration_runtime_request.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/create_linked_integration_runtime_request_py3.py diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py index 4143ff1e6621..1c0dbc75b996 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py @@ -29,6 +29,7 @@ from .update_integration_runtime_request_py3 import UpdateIntegrationRuntimeRequest from .update_integration_runtime_node_request_py3 import UpdateIntegrationRuntimeNodeRequest from .linked_integration_runtime_request_py3 import LinkedIntegrationRuntimeRequest + from .create_linked_integration_runtime_request_py3 import CreateLinkedIntegrationRuntimeRequest from .parameter_specification_py3 import ParameterSpecification from .linked_service_py3 import LinkedService from .linked_service_resource_py3 import LinkedServiceResource @@ -364,6 +365,7 @@ from .update_integration_runtime_request import UpdateIntegrationRuntimeRequest from .update_integration_runtime_node_request import UpdateIntegrationRuntimeNodeRequest from .linked_integration_runtime_request import LinkedIntegrationRuntimeRequest + from .create_linked_integration_runtime_request import CreateLinkedIntegrationRuntimeRequest from .parameter_specification import ParameterSpecification from .linked_service import LinkedService from .linked_service_resource import LinkedServiceResource @@ -767,6 +769,7 @@ 'UpdateIntegrationRuntimeRequest', 'UpdateIntegrationRuntimeNodeRequest', 'LinkedIntegrationRuntimeRequest', + 'CreateLinkedIntegrationRuntimeRequest', 'ParameterSpecification', 'LinkedService', 'LinkedServiceResource', diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/create_linked_integration_runtime_request.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/create_linked_integration_runtime_request.py new file mode 100644 index 000000000000..0e7002dcf68a --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/create_linked_integration_runtime_request.py @@ -0,0 +1,43 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class CreateLinkedIntegrationRuntimeRequest(Model): + """The linked integration runtime information. + + :param name: The name of the linked integration runtime. + :type name: str + :param subscription_id: The ID of the subscription that the linked + integration runtime belongs to. + :type subscription_id: str + :param data_factory_name: The name of the data factory that the linked + integration runtime belongs to. + :type data_factory_name: str + :param data_factory_location: The location of the data factory that the + linked integration runtime belongs to. + :type data_factory_location: str + """ + + _attribute_map = { + 'name': {'key': 'name', 'type': 'str'}, + 'subscription_id': {'key': 'subscriptionId', 'type': 'str'}, + 'data_factory_name': {'key': 'dataFactoryName', 'type': 'str'}, + 'data_factory_location': {'key': 'dataFactoryLocation', 'type': 'str'}, + } + + def __init__(self, **kwargs): + super(CreateLinkedIntegrationRuntimeRequest, self).__init__(**kwargs) + self.name = kwargs.get('name', None) + self.subscription_id = kwargs.get('subscription_id', None) + self.data_factory_name = kwargs.get('data_factory_name', None) + self.data_factory_location = kwargs.get('data_factory_location', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/create_linked_integration_runtime_request_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/create_linked_integration_runtime_request_py3.py new file mode 100644 index 000000000000..aad7d6fa5ac0 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/create_linked_integration_runtime_request_py3.py @@ -0,0 +1,43 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class CreateLinkedIntegrationRuntimeRequest(Model): + """The linked integration runtime information. + + :param name: The name of the linked integration runtime. + :type name: str + :param subscription_id: The ID of the subscription that the linked + integration runtime belongs to. + :type subscription_id: str + :param data_factory_name: The name of the data factory that the linked + integration runtime belongs to. + :type data_factory_name: str + :param data_factory_location: The location of the data factory that the + linked integration runtime belongs to. + :type data_factory_location: str + """ + + _attribute_map = { + 'name': {'key': 'name', 'type': 'str'}, + 'subscription_id': {'key': 'subscriptionId', 'type': 'str'}, + 'data_factory_name': {'key': 'dataFactoryName', 'type': 'str'}, + 'data_factory_location': {'key': 'dataFactoryLocation', 'type': 'str'}, + } + + def __init__(self, *, name: str=None, subscription_id: str=None, data_factory_name: str=None, data_factory_location: str=None, **kwargs) -> None: + super(CreateLinkedIntegrationRuntimeRequest, self).__init__(**kwargs) + self.name = name + self.subscription_id = subscription_id + self.data_factory_name = data_factory_name + self.data_factory_location = data_factory_location diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_factory_management_client_enums.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_factory_management_client_enums.py index 363e6cfb9840..086348af981d 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_factory_management_client_enums.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_factory_management_client_enums.py @@ -402,6 +402,7 @@ class SelfHostedIntegrationRuntimeNodeStatus(str, Enum): class IntegrationRuntimeUpdateResult(str, Enum): + none = "None" succeed = "Succeed" fail = "Fail" diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime_node.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime_node.py index 84b8586a3c8f..1491a80dc19a 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime_node.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime_node.py @@ -53,7 +53,7 @@ class SelfHostedIntegrationRuntimeNode(Model): :ivar last_stop_time: The integration runtime node last stop time. :vartype last_stop_time: datetime :ivar last_update_result: The result of the last integration runtime node - update. Possible values include: 'Succeed', 'Fail' + update. Possible values include: 'None', 'Succeed', 'Fail' :vartype last_update_result: str or ~azure.mgmt.datafactory.models.IntegrationRuntimeUpdateResult :ivar last_start_update_time: The last time for the integration runtime diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime_node_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime_node_py3.py index 1fd742f9ffcb..59b703737a5d 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime_node_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime_node_py3.py @@ -53,7 +53,7 @@ class SelfHostedIntegrationRuntimeNode(Model): :ivar last_stop_time: The integration runtime node last stop time. :vartype last_stop_time: datetime :ivar last_update_result: The result of the last integration runtime node - update. Possible values include: 'Succeed', 'Fail' + update. Possible values include: 'None', 'Succeed', 'Fail' :vartype last_update_result: str or ~azure.mgmt.datafactory.models.IntegrationRuntimeUpdateResult :ivar last_start_update_time: The last time for the integration runtime diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime_status.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime_status.py index c1ad1df7deda..5dd9995987d9 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime_status.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime_status.py @@ -77,6 +77,9 @@ class SelfHostedIntegrationRuntimeStatus(IntegrationRuntimeStatus): :vartype pushed_version: str :ivar latest_version: The latest version on download center. :vartype latest_version: str + :ivar auto_update_eta: The estimated time when the self-hosted integration + runtime will be updated. + :vartype auto_update_eta: datetime """ _validation = { @@ -96,6 +99,7 @@ class SelfHostedIntegrationRuntimeStatus(IntegrationRuntimeStatus): 'version_status': {'readonly': True}, 'pushed_version': {'readonly': True}, 'latest_version': {'readonly': True}, + 'auto_update_eta': {'readonly': True}, } _attribute_map = { @@ -118,6 +122,7 @@ class SelfHostedIntegrationRuntimeStatus(IntegrationRuntimeStatus): 'links': {'key': 'typeProperties.links', 'type': '[LinkedIntegrationRuntime]'}, 'pushed_version': {'key': 'typeProperties.pushedVersion', 'type': 'str'}, 'latest_version': {'key': 'typeProperties.latestVersion', 'type': 'str'}, + 'auto_update_eta': {'key': 'typeProperties.autoUpdateETA', 'type': 'iso-8601'}, } def __init__(self, **kwargs): @@ -137,4 +142,5 @@ def __init__(self, **kwargs): self.links = kwargs.get('links', None) self.pushed_version = None self.latest_version = None + self.auto_update_eta = None self.type = 'SelfHosted' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime_status_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime_status_py3.py index 100cc77982ff..acad7661fc15 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime_status_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_hosted_integration_runtime_status_py3.py @@ -77,6 +77,9 @@ class SelfHostedIntegrationRuntimeStatus(IntegrationRuntimeStatus): :vartype pushed_version: str :ivar latest_version: The latest version on download center. :vartype latest_version: str + :ivar auto_update_eta: The estimated time when the self-hosted integration + runtime will be updated. + :vartype auto_update_eta: datetime """ _validation = { @@ -96,6 +99,7 @@ class SelfHostedIntegrationRuntimeStatus(IntegrationRuntimeStatus): 'version_status': {'readonly': True}, 'pushed_version': {'readonly': True}, 'latest_version': {'readonly': True}, + 'auto_update_eta': {'readonly': True}, } _attribute_map = { @@ -118,6 +122,7 @@ class SelfHostedIntegrationRuntimeStatus(IntegrationRuntimeStatus): 'links': {'key': 'typeProperties.links', 'type': '[LinkedIntegrationRuntime]'}, 'pushed_version': {'key': 'typeProperties.pushedVersion', 'type': 'str'}, 'latest_version': {'key': 'typeProperties.latestVersion', 'type': 'str'}, + 'auto_update_eta': {'key': 'typeProperties.autoUpdateETA', 'type': 'iso-8601'}, } def __init__(self, *, additional_properties=None, nodes=None, links=None, **kwargs) -> None: @@ -137,4 +142,5 @@ def __init__(self, *, additional_properties=None, nodes=None, links=None, **kwar self.links = links self.pushed_version = None self.latest_version = None + self.auto_update_eta = None self.type = 'SelfHosted' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/activity_runs_operations.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/activity_runs_operations.py index ec8288ad562a..f338a1a9c835 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/activity_runs_operations.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/activity_runs_operations.py @@ -76,6 +76,7 @@ def query_by_pipeline_run( # Construct headers header_parameters = {} + header_parameters['Accept'] = 'application/json' header_parameters['Content-Type'] = 'application/json; charset=utf-8' if self.config.generate_client_request_id: header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) @@ -88,9 +89,8 @@ def query_by_pipeline_run( body_content = self._serialize.body(filter_parameters, 'RunFilterParameters') # Construct and send request - request = self._client.post(url, query_parameters) - response = self._client.send( - request, header_parameters, body_content, stream=False, **operation_config) + request = self._client.post(url, query_parameters, header_parameters, body_content) + response = self._client.send(request, stream=False, **operation_config) if response.status_code not in [200]: exp = CloudError(response) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/datasets_operations.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/datasets_operations.py index d2768372d196..278815d03479 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/datasets_operations.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/datasets_operations.py @@ -77,7 +77,7 @@ def internal_paging(next_link=None, raw=False): # Construct headers header_parameters = {} - header_parameters['Content-Type'] = 'application/json; charset=utf-8' + header_parameters['Accept'] = 'application/json' if self.config.generate_client_request_id: header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) if custom_headers: @@ -86,9 +86,8 @@ def internal_paging(next_link=None, raw=False): header_parameters['accept-language'] = self._serialize.header("self.config.accept_language", self.config.accept_language, 'str') # Construct and send request - request = self._client.get(url, query_parameters) - response = self._client.send( - request, header_parameters, stream=False, **operation_config) + request = self._client.get(url, query_parameters, header_parameters) + response = self._client.send(request, stream=False, **operation_config) if response.status_code not in [200]: exp = CloudError(response) @@ -152,6 +151,7 @@ def create_or_update( # Construct headers header_parameters = {} + header_parameters['Accept'] = 'application/json' header_parameters['Content-Type'] = 'application/json; charset=utf-8' if self.config.generate_client_request_id: header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) @@ -166,9 +166,8 @@ def create_or_update( body_content = self._serialize.body(dataset, 'DatasetResource') # Construct and send request - request = self._client.put(url, query_parameters) - response = self._client.send( - request, header_parameters, body_content, stream=False, **operation_config) + request = self._client.put(url, query_parameters, header_parameters, body_content) + response = self._client.send(request, stream=False, **operation_config) if response.status_code not in [200]: exp = CloudError(response) @@ -227,7 +226,7 @@ def get( # Construct headers header_parameters = {} - header_parameters['Content-Type'] = 'application/json; charset=utf-8' + header_parameters['Accept'] = 'application/json' if self.config.generate_client_request_id: header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) if custom_headers: @@ -238,8 +237,8 @@ def get( header_parameters['accept-language'] = self._serialize.header("self.config.accept_language", self.config.accept_language, 'str') # Construct and send request - request = self._client.get(url, query_parameters) - response = self._client.send(request, header_parameters, stream=False, **operation_config) + request = self._client.get(url, query_parameters, header_parameters) + response = self._client.send(request, stream=False, **operation_config) if response.status_code not in [200, 304]: exp = CloudError(response) @@ -293,7 +292,6 @@ def delete( # Construct headers header_parameters = {} - header_parameters['Content-Type'] = 'application/json; charset=utf-8' if self.config.generate_client_request_id: header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) if custom_headers: @@ -302,8 +300,8 @@ def delete( header_parameters['accept-language'] = self._serialize.header("self.config.accept_language", self.config.accept_language, 'str') # Construct and send request - request = self._client.delete(url, query_parameters) - response = self._client.send(request, header_parameters, stream=False, **operation_config) + request = self._client.delete(url, query_parameters, header_parameters) + response = self._client.send(request, stream=False, **operation_config) if response.status_code not in [200, 204]: exp = CloudError(response) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/factories_operations.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/factories_operations.py index d77c7fa31166..587cb30a87d1 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/factories_operations.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/factories_operations.py @@ -71,7 +71,7 @@ def internal_paging(next_link=None, raw=False): # Construct headers header_parameters = {} - header_parameters['Content-Type'] = 'application/json; charset=utf-8' + header_parameters['Accept'] = 'application/json' if self.config.generate_client_request_id: header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) if custom_headers: @@ -80,9 +80,8 @@ def internal_paging(next_link=None, raw=False): header_parameters['accept-language'] = self._serialize.header("self.config.accept_language", self.config.accept_language, 'str') # Construct and send request - request = self._client.get(url, query_parameters) - response = self._client.send( - request, header_parameters, stream=False, **operation_config) + request = self._client.get(url, query_parameters, header_parameters) + response = self._client.send(request, stream=False, **operation_config) if response.status_code not in [200]: exp = CloudError(response) @@ -139,6 +138,7 @@ def configure_factory_repo( # Construct headers header_parameters = {} + header_parameters['Accept'] = 'application/json' header_parameters['Content-Type'] = 'application/json; charset=utf-8' if self.config.generate_client_request_id: header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) @@ -151,9 +151,8 @@ def configure_factory_repo( body_content = self._serialize.body(factory_repo_update, 'FactoryRepoUpdate') # Construct and send request - request = self._client.post(url, query_parameters) - response = self._client.send( - request, header_parameters, body_content, stream=False, **operation_config) + request = self._client.post(url, query_parameters, header_parameters, body_content) + response = self._client.send(request, stream=False, **operation_config) if response.status_code not in [200]: exp = CloudError(response) @@ -209,7 +208,7 @@ def internal_paging(next_link=None, raw=False): # Construct headers header_parameters = {} - header_parameters['Content-Type'] = 'application/json; charset=utf-8' + header_parameters['Accept'] = 'application/json' if self.config.generate_client_request_id: header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) if custom_headers: @@ -218,9 +217,8 @@ def internal_paging(next_link=None, raw=False): header_parameters['accept-language'] = self._serialize.header("self.config.accept_language", self.config.accept_language, 'str') # Construct and send request - request = self._client.get(url, query_parameters) - response = self._client.send( - request, header_parameters, stream=False, **operation_config) + request = self._client.get(url, query_parameters, header_parameters) + response = self._client.send(request, stream=False, **operation_config) if response.status_code not in [200]: exp = CloudError(response) @@ -279,6 +277,7 @@ def create_or_update( # Construct headers header_parameters = {} + header_parameters['Accept'] = 'application/json' header_parameters['Content-Type'] = 'application/json; charset=utf-8' if self.config.generate_client_request_id: header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) @@ -293,9 +292,8 @@ def create_or_update( body_content = self._serialize.body(factory, 'Factory') # Construct and send request - request = self._client.put(url, query_parameters) - response = self._client.send( - request, header_parameters, body_content, stream=False, **operation_config) + request = self._client.put(url, query_parameters, header_parameters, body_content) + response = self._client.send(request, stream=False, **operation_config) if response.status_code not in [200]: exp = CloudError(response) @@ -353,6 +351,7 @@ def update( # Construct headers header_parameters = {} + header_parameters['Accept'] = 'application/json' header_parameters['Content-Type'] = 'application/json; charset=utf-8' if self.config.generate_client_request_id: header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) @@ -365,9 +364,8 @@ def update( body_content = self._serialize.body(factory_update_parameters, 'FactoryUpdateParameters') # Construct and send request - request = self._client.patch(url, query_parameters) - response = self._client.send( - request, header_parameters, body_content, stream=False, **operation_config) + request = self._client.patch(url, query_parameters, header_parameters, body_content) + response = self._client.send(request, stream=False, **operation_config) if response.status_code not in [200]: exp = CloudError(response) @@ -423,7 +421,7 @@ def get( # Construct headers header_parameters = {} - header_parameters['Content-Type'] = 'application/json; charset=utf-8' + header_parameters['Accept'] = 'application/json' if self.config.generate_client_request_id: header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) if custom_headers: @@ -434,8 +432,8 @@ def get( header_parameters['accept-language'] = self._serialize.header("self.config.accept_language", self.config.accept_language, 'str') # Construct and send request - request = self._client.get(url, query_parameters) - response = self._client.send(request, header_parameters, stream=False, **operation_config) + request = self._client.get(url, query_parameters, header_parameters) + response = self._client.send(request, stream=False, **operation_config) if response.status_code not in [200, 304]: exp = CloudError(response) @@ -486,7 +484,6 @@ def delete( # Construct headers header_parameters = {} - header_parameters['Content-Type'] = 'application/json; charset=utf-8' if self.config.generate_client_request_id: header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) if custom_headers: @@ -495,8 +492,8 @@ def delete( header_parameters['accept-language'] = self._serialize.header("self.config.accept_language", self.config.accept_language, 'str') # Construct and send request - request = self._client.delete(url, query_parameters) - response = self._client.send(request, header_parameters, stream=False, **operation_config) + request = self._client.delete(url, query_parameters, header_parameters) + response = self._client.send(request, stream=False, **operation_config) if response.status_code not in [200, 204]: exp = CloudError(response) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/integration_runtime_nodes_operations.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/integration_runtime_nodes_operations.py index e909517a9cbc..81467b9e3385 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/integration_runtime_nodes_operations.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/integration_runtime_nodes_operations.py @@ -37,6 +37,76 @@ def __init__(self, client, config, serializer, deserializer): self.config = config + def get( + self, resource_group_name, factory_name, integration_runtime_name, node_name, custom_headers=None, raw=False, **operation_config): + """Gets a self-hosted integration runtime node. + + :param resource_group_name: The resource group name. + :type resource_group_name: str + :param factory_name: The factory name. + :type factory_name: str + :param integration_runtime_name: The integration runtime name. + :type integration_runtime_name: str + :param node_name: The integration runtime node name. + :type node_name: str + :param dict custom_headers: headers that will be added to the request + :param bool raw: returns the direct response alongside the + deserialized response + :param operation_config: :ref:`Operation configuration + overrides`. + :return: SelfHostedIntegrationRuntimeNode or ClientRawResponse if + raw=true + :rtype: + ~azure.mgmt.datafactory.models.SelfHostedIntegrationRuntimeNode or + ~msrest.pipeline.ClientRawResponse + :raises: :class:`CloudError` + """ + # Construct URL + url = self.get.metadata['url'] + path_format_arguments = { + 'subscriptionId': self._serialize.url("self.config.subscription_id", self.config.subscription_id, 'str'), + 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str', max_length=90, min_length=1, pattern=r'^[-\w\._\(\)]+$'), + 'factoryName': self._serialize.url("factory_name", factory_name, 'str', max_length=63, min_length=3, pattern=r'^[A-Za-z0-9]+(?:-[A-Za-z0-9]+)*$'), + 'integrationRuntimeName': self._serialize.url("integration_runtime_name", integration_runtime_name, 'str', max_length=63, min_length=3, pattern=r'^[A-Za-z0-9]+(?:-[A-Za-z0-9]+)*$'), + 'nodeName': self._serialize.url("node_name", node_name, 'str', max_length=150, min_length=1, pattern=r'^[a-z0-9A-Z][a-z0-9A-Z_-]{0,149}$') + } + url = self._client.format_url(url, **path_format_arguments) + + # Construct parameters + query_parameters = {} + query_parameters['api-version'] = self._serialize.query("self.api_version", self.api_version, 'str') + + # Construct headers + header_parameters = {} + header_parameters['Accept'] = 'application/json' + if self.config.generate_client_request_id: + header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) + if custom_headers: + header_parameters.update(custom_headers) + if self.config.accept_language is not None: + header_parameters['accept-language'] = self._serialize.header("self.config.accept_language", self.config.accept_language, 'str') + + # Construct and send request + request = self._client.get(url, query_parameters, header_parameters) + response = self._client.send(request, stream=False, **operation_config) + + if response.status_code not in [200]: + exp = CloudError(response) + exp.request_id = response.headers.get('x-ms-request-id') + raise exp + + deserialized = None + + if response.status_code == 200: + deserialized = self._deserialize('SelfHostedIntegrationRuntimeNode', response) + + if raw: + client_raw_response = ClientRawResponse(deserialized, response) + return client_raw_response + + return deserialized + get.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/integrationRuntimes/{integrationRuntimeName}/nodes/{nodeName}'} + def delete( self, resource_group_name, factory_name, integration_runtime_name, node_name, custom_headers=None, raw=False, **operation_config): """Deletes a self-hosted integration runtime node. @@ -75,7 +145,6 @@ def delete( # Construct headers header_parameters = {} - header_parameters['Content-Type'] = 'application/json; charset=utf-8' if self.config.generate_client_request_id: header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) if custom_headers: @@ -84,8 +153,8 @@ def delete( header_parameters['accept-language'] = self._serialize.header("self.config.accept_language", self.config.accept_language, 'str') # Construct and send request - request = self._client.delete(url, query_parameters) - response = self._client.send(request, header_parameters, stream=False, **operation_config) + request = self._client.delete(url, query_parameters, header_parameters) + response = self._client.send(request, stream=False, **operation_config) if response.status_code not in [200, 204]: exp = CloudError(response) @@ -144,6 +213,7 @@ def update( # Construct headers header_parameters = {} + header_parameters['Accept'] = 'application/json' header_parameters['Content-Type'] = 'application/json; charset=utf-8' if self.config.generate_client_request_id: header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) @@ -156,9 +226,8 @@ def update( body_content = self._serialize.body(update_integration_runtime_node_request, 'UpdateIntegrationRuntimeNodeRequest') # Construct and send request - request = self._client.patch(url, query_parameters) - response = self._client.send( - request, header_parameters, body_content, stream=False, **operation_config) + request = self._client.patch(url, query_parameters, header_parameters, body_content) + response = self._client.send(request, stream=False, **operation_config) if response.status_code not in [200]: exp = CloudError(response) @@ -217,7 +286,7 @@ def get_ip_address( # Construct headers header_parameters = {} - header_parameters['Content-Type'] = 'application/json; charset=utf-8' + header_parameters['Accept'] = 'application/json' if self.config.generate_client_request_id: header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) if custom_headers: @@ -226,8 +295,8 @@ def get_ip_address( header_parameters['accept-language'] = self._serialize.header("self.config.accept_language", self.config.accept_language, 'str') # Construct and send request - request = self._client.post(url, query_parameters) - response = self._client.send(request, header_parameters, stream=False, **operation_config) + request = self._client.post(url, query_parameters, header_parameters) + response = self._client.send(request, stream=False, **operation_config) if response.status_code not in [200]: exp = CloudError(response) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/integration_runtimes_operations.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/integration_runtimes_operations.py index 1593ce82058b..b61c8c4e5715 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/integration_runtimes_operations.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/integration_runtimes_operations.py @@ -79,7 +79,7 @@ def internal_paging(next_link=None, raw=False): # Construct headers header_parameters = {} - header_parameters['Content-Type'] = 'application/json; charset=utf-8' + header_parameters['Accept'] = 'application/json' if self.config.generate_client_request_id: header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) if custom_headers: @@ -88,9 +88,8 @@ def internal_paging(next_link=None, raw=False): header_parameters['accept-language'] = self._serialize.header("self.config.accept_language", self.config.accept_language, 'str') # Construct and send request - request = self._client.get(url, query_parameters) - response = self._client.send( - request, header_parameters, stream=False, **operation_config) + request = self._client.get(url, query_parameters, header_parameters) + response = self._client.send(request, stream=False, **operation_config) if response.status_code not in [200]: exp = CloudError(response) @@ -154,6 +153,7 @@ def create_or_update( # Construct headers header_parameters = {} + header_parameters['Accept'] = 'application/json' header_parameters['Content-Type'] = 'application/json; charset=utf-8' if self.config.generate_client_request_id: header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) @@ -168,9 +168,8 @@ def create_or_update( body_content = self._serialize.body(integration_runtime, 'IntegrationRuntimeResource') # Construct and send request - request = self._client.put(url, query_parameters) - response = self._client.send( - request, header_parameters, body_content, stream=False, **operation_config) + request = self._client.put(url, query_parameters, header_parameters, body_content) + response = self._client.send(request, stream=False, **operation_config) if response.status_code not in [200]: exp = CloudError(response) @@ -229,7 +228,7 @@ def get( # Construct headers header_parameters = {} - header_parameters['Content-Type'] = 'application/json; charset=utf-8' + header_parameters['Accept'] = 'application/json' if self.config.generate_client_request_id: header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) if custom_headers: @@ -240,8 +239,8 @@ def get( header_parameters['accept-language'] = self._serialize.header("self.config.accept_language", self.config.accept_language, 'str') # Construct and send request - request = self._client.get(url, query_parameters) - response = self._client.send(request, header_parameters, stream=False, **operation_config) + request = self._client.get(url, query_parameters, header_parameters) + response = self._client.send(request, stream=False, **operation_config) if response.status_code not in [200, 304]: exp = CloudError(response) @@ -308,6 +307,7 @@ def update( # Construct headers header_parameters = {} + header_parameters['Accept'] = 'application/json' header_parameters['Content-Type'] = 'application/json; charset=utf-8' if self.config.generate_client_request_id: header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) @@ -320,9 +320,8 @@ def update( body_content = self._serialize.body(update_integration_runtime_request, 'UpdateIntegrationRuntimeRequest') # Construct and send request - request = self._client.patch(url, query_parameters) - response = self._client.send( - request, header_parameters, body_content, stream=False, **operation_config) + request = self._client.patch(url, query_parameters, header_parameters, body_content) + response = self._client.send(request, stream=False, **operation_config) if response.status_code not in [200]: exp = CloudError(response) @@ -376,7 +375,6 @@ def delete( # Construct headers header_parameters = {} - header_parameters['Content-Type'] = 'application/json; charset=utf-8' if self.config.generate_client_request_id: header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) if custom_headers: @@ -385,8 +383,8 @@ def delete( header_parameters['accept-language'] = self._serialize.header("self.config.accept_language", self.config.accept_language, 'str') # Construct and send request - request = self._client.delete(url, query_parameters) - response = self._client.send(request, header_parameters, stream=False, **operation_config) + request = self._client.delete(url, query_parameters, header_parameters) + response = self._client.send(request, stream=False, **operation_config) if response.status_code not in [200, 204]: exp = CloudError(response) @@ -436,7 +434,7 @@ def get_status( # Construct headers header_parameters = {} - header_parameters['Content-Type'] = 'application/json; charset=utf-8' + header_parameters['Accept'] = 'application/json' if self.config.generate_client_request_id: header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) if custom_headers: @@ -445,8 +443,8 @@ def get_status( header_parameters['accept-language'] = self._serialize.header("self.config.accept_language", self.config.accept_language, 'str') # Construct and send request - request = self._client.post(url, query_parameters) - response = self._client.send(request, header_parameters, stream=False, **operation_config) + request = self._client.post(url, query_parameters, header_parameters) + response = self._client.send(request, stream=False, **operation_config) if response.status_code not in [200]: exp = CloudError(response) @@ -504,7 +502,7 @@ def get_connection_info( # Construct headers header_parameters = {} - header_parameters['Content-Type'] = 'application/json; charset=utf-8' + header_parameters['Accept'] = 'application/json' if self.config.generate_client_request_id: header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) if custom_headers: @@ -513,8 +511,8 @@ def get_connection_info( header_parameters['accept-language'] = self._serialize.header("self.config.accept_language", self.config.accept_language, 'str') # Construct and send request - request = self._client.post(url, query_parameters) - response = self._client.send(request, header_parameters, stream=False, **operation_config) + request = self._client.post(url, query_parameters, header_parameters) + response = self._client.send(request, stream=False, **operation_config) if response.status_code not in [200]: exp = CloudError(response) @@ -575,6 +573,7 @@ def regenerate_auth_key( # Construct headers header_parameters = {} + header_parameters['Accept'] = 'application/json' header_parameters['Content-Type'] = 'application/json; charset=utf-8' if self.config.generate_client_request_id: header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) @@ -587,9 +586,8 @@ def regenerate_auth_key( body_content = self._serialize.body(regenerate_key_parameters, 'IntegrationRuntimeRegenerateKeyParameters') # Construct and send request - request = self._client.post(url, query_parameters) - response = self._client.send( - request, header_parameters, body_content, stream=False, **operation_config) + request = self._client.post(url, query_parameters, header_parameters, body_content) + response = self._client.send(request, stream=False, **operation_config) if response.status_code not in [200]: exp = CloudError(response) @@ -644,7 +642,7 @@ def list_auth_keys( # Construct headers header_parameters = {} - header_parameters['Content-Type'] = 'application/json; charset=utf-8' + header_parameters['Accept'] = 'application/json' if self.config.generate_client_request_id: header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) if custom_headers: @@ -653,8 +651,8 @@ def list_auth_keys( header_parameters['accept-language'] = self._serialize.header("self.config.accept_language", self.config.accept_language, 'str') # Construct and send request - request = self._client.post(url, query_parameters) - response = self._client.send(request, header_parameters, stream=False, **operation_config) + request = self._client.post(url, query_parameters, header_parameters) + response = self._client.send(request, stream=False, **operation_config) if response.status_code not in [200]: exp = CloudError(response) @@ -692,7 +690,7 @@ def _start_initial( # Construct headers header_parameters = {} - header_parameters['Content-Type'] = 'application/json; charset=utf-8' + header_parameters['Accept'] = 'application/json' if self.config.generate_client_request_id: header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) if custom_headers: @@ -701,8 +699,8 @@ def _start_initial( header_parameters['accept-language'] = self._serialize.header("self.config.accept_language", self.config.accept_language, 'str') # Construct and send request - request = self._client.post(url, query_parameters) - response = self._client.send(request, header_parameters, stream=False, **operation_config) + request = self._client.post(url, query_parameters, header_parameters) + response = self._client.send(request, stream=False, **operation_config) if response.status_code not in [200, 202]: exp = CloudError(response) @@ -790,7 +788,6 @@ def _stop_initial( # Construct headers header_parameters = {} - header_parameters['Content-Type'] = 'application/json; charset=utf-8' if self.config.generate_client_request_id: header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) if custom_headers: @@ -799,8 +796,8 @@ def _stop_initial( header_parameters['accept-language'] = self._serialize.header("self.config.accept_language", self.config.accept_language, 'str') # Construct and send request - request = self._client.post(url, query_parameters) - response = self._client.send(request, header_parameters, stream=False, **operation_config) + request = self._client.post(url, query_parameters, header_parameters) + response = self._client.send(request, stream=False, **operation_config) if response.status_code not in [200, 202]: exp = CloudError(response) @@ -895,7 +892,6 @@ def sync_credentials( # Construct headers header_parameters = {} - header_parameters['Content-Type'] = 'application/json; charset=utf-8' if self.config.generate_client_request_id: header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) if custom_headers: @@ -904,8 +900,8 @@ def sync_credentials( header_parameters['accept-language'] = self._serialize.header("self.config.accept_language", self.config.accept_language, 'str') # Construct and send request - request = self._client.post(url, query_parameters) - response = self._client.send(request, header_parameters, stream=False, **operation_config) + request = self._client.post(url, query_parameters, header_parameters) + response = self._client.send(request, stream=False, **operation_config) if response.status_code not in [200]: exp = CloudError(response) @@ -956,7 +952,7 @@ def get_monitoring_data( # Construct headers header_parameters = {} - header_parameters['Content-Type'] = 'application/json; charset=utf-8' + header_parameters['Accept'] = 'application/json' if self.config.generate_client_request_id: header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) if custom_headers: @@ -965,8 +961,8 @@ def get_monitoring_data( header_parameters['accept-language'] = self._serialize.header("self.config.accept_language", self.config.accept_language, 'str') # Construct and send request - request = self._client.post(url, query_parameters) - response = self._client.send(request, header_parameters, stream=False, **operation_config) + request = self._client.post(url, query_parameters, header_parameters) + response = self._client.send(request, stream=False, **operation_config) if response.status_code not in [200]: exp = CloudError(response) @@ -1020,7 +1016,6 @@ def upgrade( # Construct headers header_parameters = {} - header_parameters['Content-Type'] = 'application/json; charset=utf-8' if self.config.generate_client_request_id: header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) if custom_headers: @@ -1029,8 +1024,8 @@ def upgrade( header_parameters['accept-language'] = self._serialize.header("self.config.accept_language", self.config.accept_language, 'str') # Construct and send request - request = self._client.post(url, query_parameters) - response = self._client.send(request, header_parameters, stream=False, **operation_config) + request = self._client.post(url, query_parameters, header_parameters) + response = self._client.send(request, stream=False, **operation_config) if response.status_code not in [200]: exp = CloudError(response) @@ -1095,9 +1090,8 @@ def remove_links( body_content = self._serialize.body(linked_integration_runtime_request, 'LinkedIntegrationRuntimeRequest') # Construct and send request - request = self._client.post(url, query_parameters) - response = self._client.send( - request, header_parameters, body_content, stream=False, **operation_config) + request = self._client.post(url, query_parameters, header_parameters, body_content) + response = self._client.send(request, stream=False, **operation_config) if response.status_code not in [200]: exp = CloudError(response) @@ -1108,3 +1102,79 @@ def remove_links( client_raw_response = ClientRawResponse(None, response) return client_raw_response remove_links.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/integrationRuntimes/{integrationRuntimeName}/removeLinks'} + + def create_linked_integration_runtime( + self, resource_group_name, factory_name, integration_runtime_name, create_linked_integration_runtime_request, custom_headers=None, raw=False, **operation_config): + """Create a linked integration runtime entry in a shared integration + runtime. + + :param resource_group_name: The resource group name. + :type resource_group_name: str + :param factory_name: The factory name. + :type factory_name: str + :param integration_runtime_name: The integration runtime name. + :type integration_runtime_name: str + :param create_linked_integration_runtime_request: The linked + integration runtime properties. + :type create_linked_integration_runtime_request: + ~azure.mgmt.datafactory.models.CreateLinkedIntegrationRuntimeRequest + :param dict custom_headers: headers that will be added to the request + :param bool raw: returns the direct response alongside the + deserialized response + :param operation_config: :ref:`Operation configuration + overrides`. + :return: IntegrationRuntimeStatusResponse or ClientRawResponse if + raw=true + :rtype: + ~azure.mgmt.datafactory.models.IntegrationRuntimeStatusResponse or + ~msrest.pipeline.ClientRawResponse + :raises: :class:`CloudError` + """ + # Construct URL + url = self.create_linked_integration_runtime.metadata['url'] + path_format_arguments = { + 'subscriptionId': self._serialize.url("self.config.subscription_id", self.config.subscription_id, 'str'), + 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str', max_length=90, min_length=1, pattern=r'^[-\w\._\(\)]+$'), + 'factoryName': self._serialize.url("factory_name", factory_name, 'str', max_length=63, min_length=3, pattern=r'^[A-Za-z0-9]+(?:-[A-Za-z0-9]+)*$'), + 'integrationRuntimeName': self._serialize.url("integration_runtime_name", integration_runtime_name, 'str', max_length=63, min_length=3, pattern=r'^[A-Za-z0-9]+(?:-[A-Za-z0-9]+)*$') + } + url = self._client.format_url(url, **path_format_arguments) + + # Construct parameters + query_parameters = {} + query_parameters['api-version'] = self._serialize.query("self.api_version", self.api_version, 'str') + + # Construct headers + header_parameters = {} + header_parameters['Accept'] = 'application/json' + header_parameters['Content-Type'] = 'application/json; charset=utf-8' + if self.config.generate_client_request_id: + header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) + if custom_headers: + header_parameters.update(custom_headers) + if self.config.accept_language is not None: + header_parameters['accept-language'] = self._serialize.header("self.config.accept_language", self.config.accept_language, 'str') + + # Construct body + body_content = self._serialize.body(create_linked_integration_runtime_request, 'CreateLinkedIntegrationRuntimeRequest') + + # Construct and send request + request = self._client.post(url, query_parameters, header_parameters, body_content) + response = self._client.send(request, stream=False, **operation_config) + + if response.status_code not in [200]: + exp = CloudError(response) + exp.request_id = response.headers.get('x-ms-request-id') + raise exp + + deserialized = None + + if response.status_code == 200: + deserialized = self._deserialize('IntegrationRuntimeStatusResponse', response) + + if raw: + client_raw_response = ClientRawResponse(deserialized, response) + return client_raw_response + + return deserialized + create_linked_integration_runtime.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/integrationRuntimes/{integrationRuntimeName}/linkedIntegrationRuntime'} diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/linked_services_operations.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/linked_services_operations.py index 98ef184787ba..e6878336df91 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/linked_services_operations.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/linked_services_operations.py @@ -77,7 +77,7 @@ def internal_paging(next_link=None, raw=False): # Construct headers header_parameters = {} - header_parameters['Content-Type'] = 'application/json; charset=utf-8' + header_parameters['Accept'] = 'application/json' if self.config.generate_client_request_id: header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) if custom_headers: @@ -86,9 +86,8 @@ def internal_paging(next_link=None, raw=False): header_parameters['accept-language'] = self._serialize.header("self.config.accept_language", self.config.accept_language, 'str') # Construct and send request - request = self._client.get(url, query_parameters) - response = self._client.send( - request, header_parameters, stream=False, **operation_config) + request = self._client.get(url, query_parameters, header_parameters) + response = self._client.send(request, stream=False, **operation_config) if response.status_code not in [200]: exp = CloudError(response) @@ -152,6 +151,7 @@ def create_or_update( # Construct headers header_parameters = {} + header_parameters['Accept'] = 'application/json' header_parameters['Content-Type'] = 'application/json; charset=utf-8' if self.config.generate_client_request_id: header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) @@ -166,9 +166,8 @@ def create_or_update( body_content = self._serialize.body(linked_service, 'LinkedServiceResource') # Construct and send request - request = self._client.put(url, query_parameters) - response = self._client.send( - request, header_parameters, body_content, stream=False, **operation_config) + request = self._client.put(url, query_parameters, header_parameters, body_content) + response = self._client.send(request, stream=False, **operation_config) if response.status_code not in [200]: exp = CloudError(response) @@ -227,7 +226,7 @@ def get( # Construct headers header_parameters = {} - header_parameters['Content-Type'] = 'application/json; charset=utf-8' + header_parameters['Accept'] = 'application/json' if self.config.generate_client_request_id: header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) if custom_headers: @@ -238,8 +237,8 @@ def get( header_parameters['accept-language'] = self._serialize.header("self.config.accept_language", self.config.accept_language, 'str') # Construct and send request - request = self._client.get(url, query_parameters) - response = self._client.send(request, header_parameters, stream=False, **operation_config) + request = self._client.get(url, query_parameters, header_parameters) + response = self._client.send(request, stream=False, **operation_config) if response.status_code not in [200, 304]: exp = CloudError(response) @@ -293,7 +292,6 @@ def delete( # Construct headers header_parameters = {} - header_parameters['Content-Type'] = 'application/json; charset=utf-8' if self.config.generate_client_request_id: header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) if custom_headers: @@ -302,8 +300,8 @@ def delete( header_parameters['accept-language'] = self._serialize.header("self.config.accept_language", self.config.accept_language, 'str') # Construct and send request - request = self._client.delete(url, query_parameters) - response = self._client.send(request, header_parameters, stream=False, **operation_config) + request = self._client.delete(url, query_parameters, header_parameters) + response = self._client.send(request, stream=False, **operation_config) if response.status_code not in [200, 204]: exp = CloudError(response) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/operations.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/operations.py index 8af242714177..2273e12d5ada 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/operations.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/operations.py @@ -67,7 +67,7 @@ def internal_paging(next_link=None, raw=False): # Construct headers header_parameters = {} - header_parameters['Content-Type'] = 'application/json; charset=utf-8' + header_parameters['Accept'] = 'application/json' if self.config.generate_client_request_id: header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) if custom_headers: @@ -76,9 +76,8 @@ def internal_paging(next_link=None, raw=False): header_parameters['accept-language'] = self._serialize.header("self.config.accept_language", self.config.accept_language, 'str') # Construct and send request - request = self._client.get(url, query_parameters) - response = self._client.send( - request, header_parameters, stream=False, **operation_config) + request = self._client.get(url, query_parameters, header_parameters) + response = self._client.send(request, stream=False, **operation_config) if response.status_code not in [200]: exp = CloudError(response) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/pipeline_runs_operations.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/pipeline_runs_operations.py index 6fcc2316ce16..28799443c0d3 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/pipeline_runs_operations.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/pipeline_runs_operations.py @@ -73,6 +73,7 @@ def query_by_factory( # Construct headers header_parameters = {} + header_parameters['Accept'] = 'application/json' header_parameters['Content-Type'] = 'application/json; charset=utf-8' if self.config.generate_client_request_id: header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) @@ -85,9 +86,8 @@ def query_by_factory( body_content = self._serialize.body(filter_parameters, 'RunFilterParameters') # Construct and send request - request = self._client.post(url, query_parameters) - response = self._client.send( - request, header_parameters, body_content, stream=False, **operation_config) + request = self._client.post(url, query_parameters, header_parameters, body_content) + response = self._client.send(request, stream=False, **operation_config) if response.status_code not in [200]: exp = CloudError(response) @@ -142,7 +142,7 @@ def get( # Construct headers header_parameters = {} - header_parameters['Content-Type'] = 'application/json; charset=utf-8' + header_parameters['Accept'] = 'application/json' if self.config.generate_client_request_id: header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) if custom_headers: @@ -151,8 +151,8 @@ def get( header_parameters['accept-language'] = self._serialize.header("self.config.accept_language", self.config.accept_language, 'str') # Construct and send request - request = self._client.get(url, query_parameters) - response = self._client.send(request, header_parameters, stream=False, **operation_config) + request = self._client.get(url, query_parameters, header_parameters) + response = self._client.send(request, stream=False, **operation_config) if response.status_code not in [200]: exp = CloudError(response) @@ -206,7 +206,6 @@ def cancel( # Construct headers header_parameters = {} - header_parameters['Content-Type'] = 'application/json; charset=utf-8' if self.config.generate_client_request_id: header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) if custom_headers: @@ -215,8 +214,8 @@ def cancel( header_parameters['accept-language'] = self._serialize.header("self.config.accept_language", self.config.accept_language, 'str') # Construct and send request - request = self._client.post(url, query_parameters) - response = self._client.send(request, header_parameters, stream=False, **operation_config) + request = self._client.post(url, query_parameters, header_parameters) + response = self._client.send(request, stream=False, **operation_config) if response.status_code not in [200]: exp = CloudError(response) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/pipelines_operations.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/pipelines_operations.py index 61cf8f7bcda8..b00d9a12e180 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/pipelines_operations.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/pipelines_operations.py @@ -77,7 +77,7 @@ def internal_paging(next_link=None, raw=False): # Construct headers header_parameters = {} - header_parameters['Content-Type'] = 'application/json; charset=utf-8' + header_parameters['Accept'] = 'application/json' if self.config.generate_client_request_id: header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) if custom_headers: @@ -86,9 +86,8 @@ def internal_paging(next_link=None, raw=False): header_parameters['accept-language'] = self._serialize.header("self.config.accept_language", self.config.accept_language, 'str') # Construct and send request - request = self._client.get(url, query_parameters) - response = self._client.send( - request, header_parameters, stream=False, **operation_config) + request = self._client.get(url, query_parameters, header_parameters) + response = self._client.send(request, stream=False, **operation_config) if response.status_code not in [200]: exp = CloudError(response) @@ -150,6 +149,7 @@ def create_or_update( # Construct headers header_parameters = {} + header_parameters['Accept'] = 'application/json' header_parameters['Content-Type'] = 'application/json; charset=utf-8' if self.config.generate_client_request_id: header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) @@ -164,9 +164,8 @@ def create_or_update( body_content = self._serialize.body(pipeline, 'PipelineResource') # Construct and send request - request = self._client.put(url, query_parameters) - response = self._client.send( - request, header_parameters, body_content, stream=False, **operation_config) + request = self._client.put(url, query_parameters, header_parameters, body_content) + response = self._client.send(request, stream=False, **operation_config) if response.status_code not in [200]: exp = CloudError(response) @@ -225,7 +224,7 @@ def get( # Construct headers header_parameters = {} - header_parameters['Content-Type'] = 'application/json; charset=utf-8' + header_parameters['Accept'] = 'application/json' if self.config.generate_client_request_id: header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) if custom_headers: @@ -236,8 +235,8 @@ def get( header_parameters['accept-language'] = self._serialize.header("self.config.accept_language", self.config.accept_language, 'str') # Construct and send request - request = self._client.get(url, query_parameters) - response = self._client.send(request, header_parameters, stream=False, **operation_config) + request = self._client.get(url, query_parameters, header_parameters) + response = self._client.send(request, stream=False, **operation_config) if response.status_code not in [200, 304]: exp = CloudError(response) @@ -291,7 +290,6 @@ def delete( # Construct headers header_parameters = {} - header_parameters['Content-Type'] = 'application/json; charset=utf-8' if self.config.generate_client_request_id: header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) if custom_headers: @@ -300,8 +298,8 @@ def delete( header_parameters['accept-language'] = self._serialize.header("self.config.accept_language", self.config.accept_language, 'str') # Construct and send request - request = self._client.delete(url, query_parameters) - response = self._client.send(request, header_parameters, stream=False, **operation_config) + request = self._client.delete(url, query_parameters, header_parameters) + response = self._client.send(request, stream=False, **operation_config) if response.status_code not in [200, 204]: exp = CloudError(response) @@ -358,6 +356,7 @@ def create_run( # Construct headers header_parameters = {} + header_parameters['Accept'] = 'application/json' header_parameters['Content-Type'] = 'application/json; charset=utf-8' if self.config.generate_client_request_id: header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) @@ -373,9 +372,8 @@ def create_run( body_content = None # Construct and send request - request = self._client.post(url, query_parameters) - response = self._client.send( - request, header_parameters, body_content, stream=False, **operation_config) + request = self._client.post(url, query_parameters, header_parameters, body_content) + response = self._client.send(request, stream=False, **operation_config) if response.status_code not in [200]: exp = CloudError(response) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/trigger_runs_operations.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/trigger_runs_operations.py index f56548720d87..51e9b0ac37a3 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/trigger_runs_operations.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/trigger_runs_operations.py @@ -73,6 +73,7 @@ def query_by_factory( # Construct headers header_parameters = {} + header_parameters['Accept'] = 'application/json' header_parameters['Content-Type'] = 'application/json; charset=utf-8' if self.config.generate_client_request_id: header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) @@ -85,9 +86,8 @@ def query_by_factory( body_content = self._serialize.body(filter_parameters, 'RunFilterParameters') # Construct and send request - request = self._client.post(url, query_parameters) - response = self._client.send( - request, header_parameters, body_content, stream=False, **operation_config) + request = self._client.post(url, query_parameters, header_parameters, body_content) + response = self._client.send(request, stream=False, **operation_config) if response.status_code not in [200]: exp = CloudError(response) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/triggers_operations.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/triggers_operations.py index 9c8359038228..f80cfcb2870b 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/triggers_operations.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/triggers_operations.py @@ -79,7 +79,7 @@ def internal_paging(next_link=None, raw=False): # Construct headers header_parameters = {} - header_parameters['Content-Type'] = 'application/json; charset=utf-8' + header_parameters['Accept'] = 'application/json' if self.config.generate_client_request_id: header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) if custom_headers: @@ -88,9 +88,8 @@ def internal_paging(next_link=None, raw=False): header_parameters['accept-language'] = self._serialize.header("self.config.accept_language", self.config.accept_language, 'str') # Construct and send request - request = self._client.get(url, query_parameters) - response = self._client.send( - request, header_parameters, stream=False, **operation_config) + request = self._client.get(url, query_parameters, header_parameters) + response = self._client.send(request, stream=False, **operation_config) if response.status_code not in [200]: exp = CloudError(response) @@ -154,6 +153,7 @@ def create_or_update( # Construct headers header_parameters = {} + header_parameters['Accept'] = 'application/json' header_parameters['Content-Type'] = 'application/json; charset=utf-8' if self.config.generate_client_request_id: header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) @@ -168,9 +168,8 @@ def create_or_update( body_content = self._serialize.body(trigger, 'TriggerResource') # Construct and send request - request = self._client.put(url, query_parameters) - response = self._client.send( - request, header_parameters, body_content, stream=False, **operation_config) + request = self._client.put(url, query_parameters, header_parameters, body_content) + response = self._client.send(request, stream=False, **operation_config) if response.status_code not in [200]: exp = CloudError(response) @@ -229,7 +228,7 @@ def get( # Construct headers header_parameters = {} - header_parameters['Content-Type'] = 'application/json; charset=utf-8' + header_parameters['Accept'] = 'application/json' if self.config.generate_client_request_id: header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) if custom_headers: @@ -240,8 +239,8 @@ def get( header_parameters['accept-language'] = self._serialize.header("self.config.accept_language", self.config.accept_language, 'str') # Construct and send request - request = self._client.get(url, query_parameters) - response = self._client.send(request, header_parameters, stream=False, **operation_config) + request = self._client.get(url, query_parameters, header_parameters) + response = self._client.send(request, stream=False, **operation_config) if response.status_code not in [200, 304]: exp = CloudError(response) @@ -295,7 +294,6 @@ def delete( # Construct headers header_parameters = {} - header_parameters['Content-Type'] = 'application/json; charset=utf-8' if self.config.generate_client_request_id: header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) if custom_headers: @@ -304,8 +302,8 @@ def delete( header_parameters['accept-language'] = self._serialize.header("self.config.accept_language", self.config.accept_language, 'str') # Construct and send request - request = self._client.delete(url, query_parameters) - response = self._client.send(request, header_parameters, stream=False, **operation_config) + request = self._client.delete(url, query_parameters, header_parameters) + response = self._client.send(request, stream=False, **operation_config) if response.status_code not in [200, 204]: exp = CloudError(response) @@ -336,7 +334,6 @@ def _start_initial( # Construct headers header_parameters = {} - header_parameters['Content-Type'] = 'application/json; charset=utf-8' if self.config.generate_client_request_id: header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) if custom_headers: @@ -345,8 +342,8 @@ def _start_initial( header_parameters['accept-language'] = self._serialize.header("self.config.accept_language", self.config.accept_language, 'str') # Construct and send request - request = self._client.post(url, query_parameters) - response = self._client.send(request, header_parameters, stream=False, **operation_config) + request = self._client.post(url, query_parameters, header_parameters) + response = self._client.send(request, stream=False, **operation_config) if response.status_code not in [200]: exp = CloudError(response) @@ -420,7 +417,6 @@ def _stop_initial( # Construct headers header_parameters = {} - header_parameters['Content-Type'] = 'application/json; charset=utf-8' if self.config.generate_client_request_id: header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) if custom_headers: @@ -429,8 +425,8 @@ def _stop_initial( header_parameters['accept-language'] = self._serialize.header("self.config.accept_language", self.config.accept_language, 'str') # Construct and send request - request = self._client.post(url, query_parameters) - response = self._client.send(request, header_parameters, stream=False, **operation_config) + request = self._client.post(url, query_parameters, header_parameters) + response = self._client.send(request, stream=False, **operation_config) if response.status_code not in [200]: exp = CloudError(response) From 48e19f3fca69548802491ed4dec7b3aff3b138e2 Mon Sep 17 00:00:00 2001 From: Azure SDK for Python bot Date: Tue, 28 Aug 2018 16:22:37 -0700 Subject: [PATCH 16/54] [AutoPR datafactory/resource-manager] [Datafactory] Add folders to Pipeline and Dataset (#3138) * Generated from 89afa6ac13c9d99503cc21f4ac43988a6fe89bef Update Dataset.json Adding folders to Dataset definition * Generated from c0f87ba44c116462174353c59ac1339902d21a0d Update Dataset.json * Generated from c0f87ba44c116462174353c59ac1339902d21a0d Update Dataset.json --- .../azure/mgmt/datafactory/models/__init__.py | 6 ++++ .../models/amazon_mws_object_dataset.py | 4 +++ .../models/amazon_mws_object_dataset_py3.py | 8 +++-- .../datafactory/models/amazon_s3_dataset.py | 4 +++ .../models/amazon_s3_dataset_py3.py | 8 +++-- .../datafactory/models/azure_blob_dataset.py | 4 +++ .../models/azure_blob_dataset_py3.py | 8 +++-- .../models/azure_data_lake_store_dataset.py | 4 +++ .../azure_data_lake_store_dataset_py3.py | 8 +++-- .../models/azure_my_sql_table_dataset.py | 4 +++ .../models/azure_my_sql_table_dataset_py3.py | 8 +++-- .../models/azure_postgre_sql_table_dataset.py | 4 +++ .../azure_postgre_sql_table_dataset_py3.py | 8 +++-- .../models/azure_search_index_dataset.py | 4 +++ .../models/azure_search_index_dataset_py3.py | 8 +++-- .../models/azure_sql_dw_table_dataset.py | 4 +++ .../models/azure_sql_dw_table_dataset_py3.py | 8 +++-- .../models/azure_sql_table_dataset.py | 4 +++ .../models/azure_sql_table_dataset_py3.py | 8 +++-- .../datafactory/models/azure_table_dataset.py | 4 +++ .../models/azure_table_dataset_py3.py | 8 +++-- .../models/cassandra_table_dataset.py | 4 +++ .../models/cassandra_table_dataset_py3.py | 8 +++-- .../models/concur_object_dataset.py | 4 +++ .../models/concur_object_dataset_py3.py | 8 +++-- .../models/couchbase_table_dataset.py | 4 +++ .../models/couchbase_table_dataset_py3.py | 8 +++-- .../mgmt/datafactory/models/custom_dataset.py | 4 +++ .../datafactory/models/custom_dataset_py3.py | 8 +++-- .../azure/mgmt/datafactory/models/dataset.py | 5 ++++ .../mgmt/datafactory/models/dataset_folder.py | 29 +++++++++++++++++++ .../datafactory/models/dataset_folder_py3.py | 29 +++++++++++++++++++ .../mgmt/datafactory/models/dataset_py3.py | 7 ++++- .../models/document_db_collection_dataset.py | 4 +++ .../document_db_collection_dataset_py3.py | 8 +++-- .../datafactory/models/drill_table_dataset.py | 4 +++ .../models/drill_table_dataset_py3.py | 8 +++-- .../models/dynamics_entity_dataset.py | 4 +++ .../models/dynamics_entity_dataset_py3.py | 8 +++-- .../models/eloqua_object_dataset.py | 4 +++ .../models/eloqua_object_dataset_py3.py | 8 +++-- .../datafactory/models/file_share_dataset.py | 4 +++ .../models/file_share_dataset_py3.py | 8 +++-- .../models/google_big_query_object_dataset.py | 4 +++ .../google_big_query_object_dataset_py3.py | 8 +++-- .../models/greenplum_table_dataset.py | 4 +++ .../models/greenplum_table_dataset_py3.py | 8 +++-- .../models/hbase_object_dataset.py | 4 +++ .../models/hbase_object_dataset_py3.py | 8 +++-- .../datafactory/models/hive_object_dataset.py | 4 +++ .../models/hive_object_dataset_py3.py | 8 +++-- .../mgmt/datafactory/models/http_dataset.py | 4 +++ .../datafactory/models/http_dataset_py3.py | 8 +++-- .../models/hubspot_object_dataset.py | 4 +++ .../models/hubspot_object_dataset_py3.py | 8 +++-- .../models/impala_object_dataset.py | 4 +++ .../models/impala_object_dataset_py3.py | 8 +++-- .../datafactory/models/jira_object_dataset.py | 4 +++ .../models/jira_object_dataset_py3.py | 8 +++-- .../models/magento_object_dataset.py | 4 +++ .../models/magento_object_dataset_py3.py | 8 +++-- .../models/maria_db_table_dataset.py | 4 +++ .../models/maria_db_table_dataset_py3.py | 8 +++-- .../models/marketo_object_dataset.py | 4 +++ .../models/marketo_object_dataset_py3.py | 8 +++-- .../models/mongo_db_collection_dataset.py | 4 +++ .../models/mongo_db_collection_dataset_py3.py | 8 +++-- .../models/netezza_table_dataset.py | 4 +++ .../models/netezza_table_dataset_py3.py | 8 +++-- .../models/odata_resource_dataset.py | 4 +++ .../models/odata_resource_dataset_py3.py | 8 +++-- .../models/oracle_table_dataset.py | 4 +++ .../models/oracle_table_dataset_py3.py | 8 +++-- .../models/paypal_object_dataset.py | 4 +++ .../models/paypal_object_dataset_py3.py | 8 +++-- .../models/phoenix_object_dataset.py | 4 +++ .../models/phoenix_object_dataset_py3.py | 8 +++-- .../datafactory/models/pipeline_folder.py | 29 +++++++++++++++++++ .../datafactory/models/pipeline_folder_py3.py | 29 +++++++++++++++++++ .../datafactory/models/pipeline_resource.py | 5 ++++ .../models/pipeline_resource_py3.py | 7 ++++- .../models/presto_object_dataset.py | 4 +++ .../models/presto_object_dataset_py3.py | 8 +++-- .../models/quick_books_object_dataset.py | 4 +++ .../models/quick_books_object_dataset_py3.py | 8 +++-- .../models/relational_table_dataset.py | 4 +++ .../models/relational_table_dataset_py3.py | 8 +++-- .../models/responsys_object_dataset.py | 4 +++ .../models/responsys_object_dataset_py3.py | 8 +++-- ...lesforce_marketing_cloud_object_dataset.py | 4 +++ ...orce_marketing_cloud_object_dataset_py3.py | 8 +++-- .../models/salesforce_object_dataset.py | 4 +++ .../models/salesforce_object_dataset_py3.py | 8 +++-- ...sap_cloud_for_customer_resource_dataset.py | 4 +++ ...cloud_for_customer_resource_dataset_py3.py | 8 +++-- .../models/sap_ecc_resource_dataset.py | 4 +++ .../models/sap_ecc_resource_dataset_py3.py | 8 +++-- .../models/service_now_object_dataset.py | 4 +++ .../models/service_now_object_dataset_py3.py | 8 +++-- .../models/shopify_object_dataset.py | 4 +++ .../models/shopify_object_dataset_py3.py | 8 +++-- .../models/spark_object_dataset.py | 4 +++ .../models/spark_object_dataset_py3.py | 8 +++-- .../models/sql_server_table_dataset.py | 4 +++ .../models/sql_server_table_dataset_py3.py | 8 +++-- .../models/square_object_dataset.py | 4 +++ .../models/square_object_dataset_py3.py | 8 +++-- .../models/vertica_table_dataset.py | 4 +++ .../models/vertica_table_dataset_py3.py | 8 +++-- .../datafactory/models/web_table_dataset.py | 4 +++ .../models/web_table_dataset_py3.py | 8 +++-- .../datafactory/models/xero_object_dataset.py | 4 +++ .../models/xero_object_dataset_py3.py | 8 +++-- .../datafactory/models/zoho_object_dataset.py | 4 +++ .../models/zoho_object_dataset_py3.py | 8 +++-- 115 files changed, 674 insertions(+), 108 deletions(-) create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_folder.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_folder_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_folder.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_folder_py3.py diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py index 1c0dbc75b996..e20345f7ab5b 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py @@ -33,11 +33,13 @@ from .parameter_specification_py3 import ParameterSpecification from .linked_service_py3 import LinkedService from .linked_service_resource_py3 import LinkedServiceResource + from .dataset_folder_py3 import DatasetFolder from .dataset_py3 import Dataset from .dataset_resource_py3 import DatasetResource from .activity_dependency_py3 import ActivityDependency from .user_property_py3 import UserProperty from .activity_py3 import Activity + from .pipeline_folder_py3 import PipelineFolder from .pipeline_resource_py3 import PipelineResource from .trigger_py3 import Trigger from .trigger_resource_py3 import TriggerResource @@ -369,11 +371,13 @@ from .parameter_specification import ParameterSpecification from .linked_service import LinkedService from .linked_service_resource import LinkedServiceResource + from .dataset_folder import DatasetFolder from .dataset import Dataset from .dataset_resource import DatasetResource from .activity_dependency import ActivityDependency from .user_property import UserProperty from .activity import Activity + from .pipeline_folder import PipelineFolder from .pipeline_resource import PipelineResource from .trigger import Trigger from .trigger_resource import TriggerResource @@ -773,11 +777,13 @@ 'ParameterSpecification', 'LinkedService', 'LinkedServiceResource', + 'DatasetFolder', 'Dataset', 'DatasetResource', 'ActivityDependency', 'UserProperty', 'Activity', + 'PipelineFolder', 'PipelineResource', 'Trigger', 'TriggerResource', diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_object_dataset.py index 7062a5534fca..07eaa902b276 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_object_dataset.py @@ -34,6 +34,9 @@ class AmazonMWSObjectDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str """ @@ -50,6 +53,7 @@ class AmazonMWSObjectDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, } diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_object_dataset_py3.py index 7250b12f0fb9..3180f8952185 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_object_dataset_py3.py @@ -34,6 +34,9 @@ class AmazonMWSObjectDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str """ @@ -50,9 +53,10 @@ class AmazonMWSObjectDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: - super(AmazonMWSObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + super(AmazonMWSObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.type = 'AmazonMWSObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_s3_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_s3_dataset.py index f2bd2a93dff8..87540eb0eb11 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_s3_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_s3_dataset.py @@ -34,6 +34,9 @@ class AmazonS3Dataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str :param bucket_name: Required. The name of the Amazon S3 bucket. Type: @@ -68,6 +71,7 @@ class AmazonS3Dataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, 'bucket_name': {'key': 'typeProperties.bucketName', 'type': 'object'}, 'key': {'key': 'typeProperties.key', 'type': 'object'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_s3_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_s3_dataset_py3.py index b9cc4fc4782b..c4b602da1303 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_s3_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_s3_dataset_py3.py @@ -34,6 +34,9 @@ class AmazonS3Dataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str :param bucket_name: Required. The name of the Amazon S3 bucket. Type: @@ -68,6 +71,7 @@ class AmazonS3Dataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, 'bucket_name': {'key': 'typeProperties.bucketName', 'type': 'object'}, 'key': {'key': 'typeProperties.key', 'type': 'object'}, @@ -77,8 +81,8 @@ class AmazonS3Dataset(Dataset): 'compression': {'key': 'typeProperties.compression', 'type': 'DatasetCompression'}, } - def __init__(self, *, linked_service_name, bucket_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, key=None, prefix=None, version=None, format=None, compression=None, **kwargs) -> None: - super(AmazonS3Dataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + def __init__(self, *, linked_service_name, bucket_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, key=None, prefix=None, version=None, format=None, compression=None, **kwargs) -> None: + super(AmazonS3Dataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.bucket_name = bucket_name self.key = key self.prefix = prefix diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_dataset.py index fa15f8a9ad1e..dfa31040d658 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_dataset.py @@ -34,6 +34,9 @@ class AzureBlobDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str :param folder_path: The path of the Azure Blob storage. Type: string (or @@ -63,6 +66,7 @@ class AzureBlobDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, 'folder_path': {'key': 'typeProperties.folderPath', 'type': 'object'}, 'table_root_location': {'key': 'typeProperties.tableRootLocation', 'type': 'object'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_dataset_py3.py index fe28cffcac03..adef2dbd6def 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_dataset_py3.py @@ -34,6 +34,9 @@ class AzureBlobDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str :param folder_path: The path of the Azure Blob storage. Type: string (or @@ -63,6 +66,7 @@ class AzureBlobDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, 'folder_path': {'key': 'typeProperties.folderPath', 'type': 'object'}, 'table_root_location': {'key': 'typeProperties.tableRootLocation', 'type': 'object'}, @@ -71,8 +75,8 @@ class AzureBlobDataset(Dataset): 'compression': {'key': 'typeProperties.compression', 'type': 'DatasetCompression'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder_path=None, table_root_location=None, file_name=None, format=None, compression=None, **kwargs) -> None: - super(AzureBlobDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, folder_path=None, table_root_location=None, file_name=None, format=None, compression=None, **kwargs) -> None: + super(AzureBlobDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.folder_path = folder_path self.table_root_location = table_root_location self.file_name = file_name diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_dataset.py index be589a7803cc..0d3d9edf45d9 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_dataset.py @@ -34,6 +34,9 @@ class AzureDataLakeStoreDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str :param folder_path: Required. Path to the folder in the Azure Data Lake @@ -62,6 +65,7 @@ class AzureDataLakeStoreDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, 'folder_path': {'key': 'typeProperties.folderPath', 'type': 'object'}, 'file_name': {'key': 'typeProperties.fileName', 'type': 'object'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_dataset_py3.py index 5053cbd9a7a6..8d108c944dcb 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_dataset_py3.py @@ -34,6 +34,9 @@ class AzureDataLakeStoreDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str :param folder_path: Required. Path to the folder in the Azure Data Lake @@ -62,6 +65,7 @@ class AzureDataLakeStoreDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, 'folder_path': {'key': 'typeProperties.folderPath', 'type': 'object'}, 'file_name': {'key': 'typeProperties.fileName', 'type': 'object'}, @@ -69,8 +73,8 @@ class AzureDataLakeStoreDataset(Dataset): 'compression': {'key': 'typeProperties.compression', 'type': 'DatasetCompression'}, } - def __init__(self, *, linked_service_name, folder_path, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, file_name=None, format=None, compression=None, **kwargs) -> None: - super(AzureDataLakeStoreDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + def __init__(self, *, linked_service_name, folder_path, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, file_name=None, format=None, compression=None, **kwargs) -> None: + super(AzureDataLakeStoreDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.folder_path = folder_path self.file_name = file_name self.format = format diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_table_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_table_dataset.py index d342c3528e04..bbea45836145 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_table_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_table_dataset.py @@ -34,6 +34,9 @@ class AzureMySqlTableDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str :param table_name: The Azure MySQL database table name. Type: string (or @@ -53,6 +56,7 @@ class AzureMySqlTableDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_table_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_table_dataset_py3.py index 3807f7484878..1fc1fef3201f 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_table_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_table_dataset_py3.py @@ -34,6 +34,9 @@ class AzureMySqlTableDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str :param table_name: The Azure MySQL database table name. Type: string (or @@ -53,11 +56,12 @@ class AzureMySqlTableDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, table_name=None, **kwargs) -> None: - super(AzureMySqlTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: + super(AzureMySqlTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.table_name = table_name self.type = 'AzureMySqlTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_table_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_table_dataset.py index 0ebd39c8c3ab..33e74887c012 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_table_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_table_dataset.py @@ -34,6 +34,9 @@ class AzurePostgreSqlTableDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str """ @@ -50,6 +53,7 @@ class AzurePostgreSqlTableDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, } diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_table_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_table_dataset_py3.py index 9f3a62d269f2..0a6d551868f7 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_table_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_table_dataset_py3.py @@ -34,6 +34,9 @@ class AzurePostgreSqlTableDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str """ @@ -50,9 +53,10 @@ class AzurePostgreSqlTableDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: - super(AzurePostgreSqlTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + super(AzurePostgreSqlTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.type = 'AzurePostgreSqlTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_search_index_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_search_index_dataset.py index afe7efe8f66d..de9f38cf6793 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_search_index_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_search_index_dataset.py @@ -34,6 +34,9 @@ class AzureSearchIndexDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str :param index_name: Required. The name of the Azure Search Index. Type: @@ -54,6 +57,7 @@ class AzureSearchIndexDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, 'index_name': {'key': 'typeProperties.indexName', 'type': 'object'}, } diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_search_index_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_search_index_dataset_py3.py index b949c1878a80..6d57a2d3faad 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_search_index_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_search_index_dataset_py3.py @@ -34,6 +34,9 @@ class AzureSearchIndexDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str :param index_name: Required. The name of the Azure Search Index. Type: @@ -54,11 +57,12 @@ class AzureSearchIndexDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, 'index_name': {'key': 'typeProperties.indexName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, index_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: - super(AzureSearchIndexDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + def __init__(self, *, linked_service_name, index_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + super(AzureSearchIndexDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.index_name = index_name self.type = 'AzureSearchIndex' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_dw_table_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_dw_table_dataset.py index c66b72f6ef06..75dcf5778ba3 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_dw_table_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_dw_table_dataset.py @@ -34,6 +34,9 @@ class AzureSqlDWTableDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str :param table_name: Required. The table name of the Azure SQL Data @@ -54,6 +57,7 @@ class AzureSqlDWTableDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_dw_table_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_dw_table_dataset_py3.py index 145d409e2b76..516d9fc80d69 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_dw_table_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_dw_table_dataset_py3.py @@ -34,6 +34,9 @@ class AzureSqlDWTableDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str :param table_name: Required. The table name of the Azure SQL Data @@ -54,11 +57,12 @@ class AzureSqlDWTableDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, table_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: - super(AzureSqlDWTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + def __init__(self, *, linked_service_name, table_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + super(AzureSqlDWTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.table_name = table_name self.type = 'AzureSqlDWTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_table_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_table_dataset.py index 8b6f061514d8..6e7cd0acc419 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_table_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_table_dataset.py @@ -34,6 +34,9 @@ class AzureSqlTableDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str :param table_name: Required. The table name of the Azure SQL database. @@ -54,6 +57,7 @@ class AzureSqlTableDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_table_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_table_dataset_py3.py index 8a226a2e673b..1769551763da 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_table_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_table_dataset_py3.py @@ -34,6 +34,9 @@ class AzureSqlTableDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str :param table_name: Required. The table name of the Azure SQL database. @@ -54,11 +57,12 @@ class AzureSqlTableDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, table_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: - super(AzureSqlTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + def __init__(self, *, linked_service_name, table_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + super(AzureSqlTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.table_name = table_name self.type = 'AzureSqlTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_table_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_table_dataset.py index 810d53a4fdca..dae0a5b3bf02 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_table_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_table_dataset.py @@ -34,6 +34,9 @@ class AzureTableDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str :param table_name: Required. The table name of the Azure Table storage. @@ -54,6 +57,7 @@ class AzureTableDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_table_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_table_dataset_py3.py index b0c1cd13e693..74f0ae0b7096 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_table_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_table_dataset_py3.py @@ -34,6 +34,9 @@ class AzureTableDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str :param table_name: Required. The table name of the Azure Table storage. @@ -54,11 +57,12 @@ class AzureTableDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, table_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: - super(AzureTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + def __init__(self, *, linked_service_name, table_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + super(AzureTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.table_name = table_name self.type = 'AzureTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cassandra_table_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cassandra_table_dataset.py index 91b5e11aa8f2..b356830de832 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cassandra_table_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cassandra_table_dataset.py @@ -34,6 +34,9 @@ class CassandraTableDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str :param table_name: The table name of the Cassandra database. Type: string @@ -56,6 +59,7 @@ class CassandraTableDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, 'keyspace': {'key': 'typeProperties.keyspace', 'type': 'object'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cassandra_table_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cassandra_table_dataset_py3.py index 6927e69a6e0d..914c8b38cb62 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cassandra_table_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cassandra_table_dataset_py3.py @@ -34,6 +34,9 @@ class CassandraTableDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str :param table_name: The table name of the Cassandra database. Type: string @@ -56,13 +59,14 @@ class CassandraTableDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, 'keyspace': {'key': 'typeProperties.keyspace', 'type': 'object'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, table_name=None, keyspace=None, **kwargs) -> None: - super(CassandraTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, table_name=None, keyspace=None, **kwargs) -> None: + super(CassandraTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.table_name = table_name self.keyspace = keyspace self.type = 'CassandraTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_object_dataset.py index 00ee6540daa1..f0bf7c705098 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_object_dataset.py @@ -34,6 +34,9 @@ class ConcurObjectDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str """ @@ -50,6 +53,7 @@ class ConcurObjectDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, } diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_object_dataset_py3.py index b77695bdc037..6ca127fab236 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_object_dataset_py3.py @@ -34,6 +34,9 @@ class ConcurObjectDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str """ @@ -50,9 +53,10 @@ class ConcurObjectDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: - super(ConcurObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + super(ConcurObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.type = 'ConcurObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_table_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_table_dataset.py index 21e4a5d39363..88053e34fe9d 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_table_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_table_dataset.py @@ -34,6 +34,9 @@ class CouchbaseTableDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str """ @@ -50,6 +53,7 @@ class CouchbaseTableDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, } diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_table_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_table_dataset_py3.py index aab909106680..00bf158a6f84 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_table_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_table_dataset_py3.py @@ -34,6 +34,9 @@ class CouchbaseTableDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str """ @@ -50,9 +53,10 @@ class CouchbaseTableDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: - super(CouchbaseTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + super(CouchbaseTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.type = 'CouchbaseTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_dataset.py index 4dc34f7442a9..b74f0e44cf8e 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_dataset.py @@ -34,6 +34,9 @@ class CustomDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str :param type_properties: Required. Custom dataset properties. @@ -53,6 +56,7 @@ class CustomDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, 'type_properties': {'key': 'typeProperties', 'type': 'object'}, } diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_dataset_py3.py index 8070049fa8cd..65927670ca6c 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_dataset_py3.py @@ -34,6 +34,9 @@ class CustomDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str :param type_properties: Required. Custom dataset properties. @@ -53,11 +56,12 @@ class CustomDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, 'type_properties': {'key': 'typeProperties', 'type': 'object'}, } - def __init__(self, *, linked_service_name, type_properties, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: - super(CustomDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + def __init__(self, *, linked_service_name, type_properties, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + super(CustomDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.type_properties = type_properties self.type = 'CustomDataset' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset.py index 4ebebafdda73..626015fd5d11 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset.py @@ -57,6 +57,9 @@ class Dataset(Model): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str """ @@ -73,6 +76,7 @@ class Dataset(Model): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, } @@ -88,4 +92,5 @@ def __init__(self, **kwargs): self.linked_service_name = kwargs.get('linked_service_name', None) self.parameters = kwargs.get('parameters', None) self.annotations = kwargs.get('annotations', None) + self.folder = kwargs.get('folder', None) self.type = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_folder.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_folder.py new file mode 100644 index 000000000000..882c84a1e84c --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_folder.py @@ -0,0 +1,29 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class DatasetFolder(Model): + """The folder that this Dataset is in. If not specified, Dataset will appear + at the root level. + + :param name: The name of the folder that this Dataset is in. + :type name: str + """ + + _attribute_map = { + 'name': {'key': 'name', 'type': 'str'}, + } + + def __init__(self, **kwargs): + super(DatasetFolder, self).__init__(**kwargs) + self.name = kwargs.get('name', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_folder_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_folder_py3.py new file mode 100644 index 000000000000..ea7fc313f967 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_folder_py3.py @@ -0,0 +1,29 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class DatasetFolder(Model): + """The folder that this Dataset is in. If not specified, Dataset will appear + at the root level. + + :param name: The name of the folder that this Dataset is in. + :type name: str + """ + + _attribute_map = { + 'name': {'key': 'name', 'type': 'str'}, + } + + def __init__(self, *, name: str=None, **kwargs) -> None: + super(DatasetFolder, self).__init__(**kwargs) + self.name = name diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_py3.py index 65f43811d7a7..3d67a1328cd4 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_py3.py @@ -57,6 +57,9 @@ class Dataset(Model): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str """ @@ -73,6 +76,7 @@ class Dataset(Model): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, } @@ -80,7 +84,7 @@ class Dataset(Model): 'type': {'ResponsysObject': 'ResponsysObjectDataset', 'SalesforceMarketingCloudObject': 'SalesforceMarketingCloudObjectDataset', 'VerticaTable': 'VerticaTableDataset', 'NetezzaTable': 'NetezzaTableDataset', 'ZohoObject': 'ZohoObjectDataset', 'XeroObject': 'XeroObjectDataset', 'SquareObject': 'SquareObjectDataset', 'SparkObject': 'SparkObjectDataset', 'ShopifyObject': 'ShopifyObjectDataset', 'ServiceNowObject': 'ServiceNowObjectDataset', 'QuickBooksObject': 'QuickBooksObjectDataset', 'PrestoObject': 'PrestoObjectDataset', 'PhoenixObject': 'PhoenixObjectDataset', 'PaypalObject': 'PaypalObjectDataset', 'MarketoObject': 'MarketoObjectDataset', 'MariaDBTable': 'MariaDBTableDataset', 'MagentoObject': 'MagentoObjectDataset', 'JiraObject': 'JiraObjectDataset', 'ImpalaObject': 'ImpalaObjectDataset', 'HubspotObject': 'HubspotObjectDataset', 'HiveObject': 'HiveObjectDataset', 'HBaseObject': 'HBaseObjectDataset', 'GreenplumTable': 'GreenplumTableDataset', 'GoogleBigQueryObject': 'GoogleBigQueryObjectDataset', 'EloquaObject': 'EloquaObjectDataset', 'DrillTable': 'DrillTableDataset', 'CouchbaseTable': 'CouchbaseTableDataset', 'ConcurObject': 'ConcurObjectDataset', 'AzurePostgreSqlTable': 'AzurePostgreSqlTableDataset', 'AmazonMWSObject': 'AmazonMWSObjectDataset', 'HttpFile': 'HttpDataset', 'AzureSearchIndex': 'AzureSearchIndexDataset', 'WebTable': 'WebTableDataset', 'SqlServerTable': 'SqlServerTableDataset', 'SapEccResource': 'SapEccResourceDataset', 'SapCloudForCustomerResource': 'SapCloudForCustomerResourceDataset', 'SalesforceObject': 'SalesforceObjectDataset', 'RelationalTable': 'RelationalTableDataset', 'AzureMySqlTable': 'AzureMySqlTableDataset', 'OracleTable': 'OracleTableDataset', 'ODataResource': 'ODataResourceDataset', 'MongoDbCollection': 'MongoDbCollectionDataset', 'FileShare': 'FileShareDataset', 'AzureDataLakeStoreFile': 'AzureDataLakeStoreDataset', 'DynamicsEntity': 'DynamicsEntityDataset', 'DocumentDbCollection': 'DocumentDbCollectionDataset', 'CustomDataset': 'CustomDataset', 'CassandraTable': 'CassandraTableDataset', 'AzureSqlDWTable': 'AzureSqlDWTableDataset', 'AzureSqlTable': 'AzureSqlTableDataset', 'AzureTable': 'AzureTableDataset', 'AzureBlob': 'AzureBlobDataset', 'AmazonS3Object': 'AmazonS3Dataset'} } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: super(Dataset, self).__init__(**kwargs) self.additional_properties = additional_properties self.description = description @@ -88,4 +92,5 @@ def __init__(self, *, linked_service_name, additional_properties=None, descripti self.linked_service_name = linked_service_name self.parameters = parameters self.annotations = annotations + self.folder = folder self.type = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/document_db_collection_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/document_db_collection_dataset.py index 58f055d3693e..84a9e50bd82d 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/document_db_collection_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/document_db_collection_dataset.py @@ -34,6 +34,9 @@ class DocumentDbCollectionDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str :param collection_name: Required. Document Database collection name. Type: @@ -54,6 +57,7 @@ class DocumentDbCollectionDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, 'collection_name': {'key': 'typeProperties.collectionName', 'type': 'object'}, } diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/document_db_collection_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/document_db_collection_dataset_py3.py index e30288172785..c3566bf9e0f8 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/document_db_collection_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/document_db_collection_dataset_py3.py @@ -34,6 +34,9 @@ class DocumentDbCollectionDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str :param collection_name: Required. Document Database collection name. Type: @@ -54,11 +57,12 @@ class DocumentDbCollectionDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, 'collection_name': {'key': 'typeProperties.collectionName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, collection_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: - super(DocumentDbCollectionDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + def __init__(self, *, linked_service_name, collection_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + super(DocumentDbCollectionDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.collection_name = collection_name self.type = 'DocumentDbCollection' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_table_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_table_dataset.py index 21e7e392ec7b..59802dcfcd77 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_table_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_table_dataset.py @@ -34,6 +34,9 @@ class DrillTableDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str """ @@ -50,6 +53,7 @@ class DrillTableDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, } diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_table_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_table_dataset_py3.py index cdff7f8360ca..be4da0ea660d 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_table_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_table_dataset_py3.py @@ -34,6 +34,9 @@ class DrillTableDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str """ @@ -50,9 +53,10 @@ class DrillTableDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: - super(DrillTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + super(DrillTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.type = 'DrillTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_entity_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_entity_dataset.py index 23eb34e5c6b2..2048d589759c 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_entity_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_entity_dataset.py @@ -34,6 +34,9 @@ class DynamicsEntityDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str :param entity_name: The logical name of the entity. Type: string (or @@ -53,6 +56,7 @@ class DynamicsEntityDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, 'entity_name': {'key': 'typeProperties.entityName', 'type': 'object'}, } diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_entity_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_entity_dataset_py3.py index 4a5e6f3f542b..e87e511d062a 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_entity_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_entity_dataset_py3.py @@ -34,6 +34,9 @@ class DynamicsEntityDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str :param entity_name: The logical name of the entity. Type: string (or @@ -53,11 +56,12 @@ class DynamicsEntityDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, 'entity_name': {'key': 'typeProperties.entityName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, entity_name=None, **kwargs) -> None: - super(DynamicsEntityDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, entity_name=None, **kwargs) -> None: + super(DynamicsEntityDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.entity_name = entity_name self.type = 'DynamicsEntity' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/eloqua_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/eloqua_object_dataset.py index 95d41cef4d5e..5ff8abf3bf32 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/eloqua_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/eloqua_object_dataset.py @@ -34,6 +34,9 @@ class EloquaObjectDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str """ @@ -50,6 +53,7 @@ class EloquaObjectDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, } diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/eloqua_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/eloqua_object_dataset_py3.py index fe8be1261fcd..bd4e833d1eea 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/eloqua_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/eloqua_object_dataset_py3.py @@ -34,6 +34,9 @@ class EloquaObjectDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str """ @@ -50,9 +53,10 @@ class EloquaObjectDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: - super(EloquaObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + super(EloquaObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.type = 'EloquaObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_share_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_share_dataset.py index 0ca1dc0fd79a..d5cbf4ac6822 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_share_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_share_dataset.py @@ -34,6 +34,9 @@ class FileShareDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str :param folder_path: The path of the on-premises file system. Type: string @@ -64,6 +67,7 @@ class FileShareDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, 'folder_path': {'key': 'typeProperties.folderPath', 'type': 'object'}, 'file_name': {'key': 'typeProperties.fileName', 'type': 'object'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_share_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_share_dataset_py3.py index 08524cbc2a4b..fe980181543e 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_share_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_share_dataset_py3.py @@ -34,6 +34,9 @@ class FileShareDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str :param folder_path: The path of the on-premises file system. Type: string @@ -64,6 +67,7 @@ class FileShareDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, 'folder_path': {'key': 'typeProperties.folderPath', 'type': 'object'}, 'file_name': {'key': 'typeProperties.fileName', 'type': 'object'}, @@ -72,8 +76,8 @@ class FileShareDataset(Dataset): 'compression': {'key': 'typeProperties.compression', 'type': 'DatasetCompression'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder_path=None, file_name=None, format=None, file_filter=None, compression=None, **kwargs) -> None: - super(FileShareDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, folder_path=None, file_name=None, format=None, file_filter=None, compression=None, **kwargs) -> None: + super(FileShareDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.folder_path = folder_path self.file_name = file_name self.format = format diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_big_query_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_big_query_object_dataset.py index e089870c4d1a..2a739203a74b 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_big_query_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_big_query_object_dataset.py @@ -34,6 +34,9 @@ class GoogleBigQueryObjectDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str """ @@ -50,6 +53,7 @@ class GoogleBigQueryObjectDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, } diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_big_query_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_big_query_object_dataset_py3.py index d51dfe5de0f9..57423a6bc421 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_big_query_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_big_query_object_dataset_py3.py @@ -34,6 +34,9 @@ class GoogleBigQueryObjectDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str """ @@ -50,9 +53,10 @@ class GoogleBigQueryObjectDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: - super(GoogleBigQueryObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + super(GoogleBigQueryObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.type = 'GoogleBigQueryObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_table_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_table_dataset.py index 7c031f285b00..7a5bbdb32063 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_table_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_table_dataset.py @@ -34,6 +34,9 @@ class GreenplumTableDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str """ @@ -50,6 +53,7 @@ class GreenplumTableDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, } diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_table_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_table_dataset_py3.py index a266e22dae9c..a693ddcd4e41 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_table_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_table_dataset_py3.py @@ -34,6 +34,9 @@ class GreenplumTableDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str """ @@ -50,9 +53,10 @@ class GreenplumTableDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: - super(GreenplumTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + super(GreenplumTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.type = 'GreenplumTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hbase_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hbase_object_dataset.py index 6e78856a1468..8500e647d220 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hbase_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hbase_object_dataset.py @@ -34,6 +34,9 @@ class HBaseObjectDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str """ @@ -50,6 +53,7 @@ class HBaseObjectDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, } diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hbase_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hbase_object_dataset_py3.py index a4386af6240f..d52489d391c0 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hbase_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hbase_object_dataset_py3.py @@ -34,6 +34,9 @@ class HBaseObjectDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str """ @@ -50,9 +53,10 @@ class HBaseObjectDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: - super(HBaseObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + super(HBaseObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.type = 'HBaseObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hive_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hive_object_dataset.py index cfc5620fd776..848031ff4115 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hive_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hive_object_dataset.py @@ -34,6 +34,9 @@ class HiveObjectDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str """ @@ -50,6 +53,7 @@ class HiveObjectDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, } diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hive_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hive_object_dataset_py3.py index d4bebbbcfe7b..944cab17444e 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hive_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hive_object_dataset_py3.py @@ -34,6 +34,9 @@ class HiveObjectDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str """ @@ -50,9 +53,10 @@ class HiveObjectDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: - super(HiveObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + super(HiveObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.type = 'HiveObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/http_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/http_dataset.py index e2b2325b5294..532cd654ccef 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/http_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/http_dataset.py @@ -34,6 +34,9 @@ class HttpDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str :param relative_url: The relative URL based on the URL in the @@ -70,6 +73,7 @@ class HttpDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, 'relative_url': {'key': 'typeProperties.relativeUrl', 'type': 'object'}, 'request_method': {'key': 'typeProperties.requestMethod', 'type': 'object'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/http_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/http_dataset_py3.py index 05c340ab447a..6cacff526753 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/http_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/http_dataset_py3.py @@ -34,6 +34,9 @@ class HttpDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str :param relative_url: The relative URL based on the URL in the @@ -70,6 +73,7 @@ class HttpDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, 'relative_url': {'key': 'typeProperties.relativeUrl', 'type': 'object'}, 'request_method': {'key': 'typeProperties.requestMethod', 'type': 'object'}, @@ -79,8 +83,8 @@ class HttpDataset(Dataset): 'compression': {'key': 'typeProperties.compression', 'type': 'DatasetCompression'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, relative_url=None, request_method=None, request_body=None, additional_headers=None, format=None, compression=None, **kwargs) -> None: - super(HttpDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, relative_url=None, request_method=None, request_body=None, additional_headers=None, format=None, compression=None, **kwargs) -> None: + super(HttpDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.relative_url = relative_url self.request_method = request_method self.request_body = request_body diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_object_dataset.py index a03b68a4f38b..d3a53c067ede 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_object_dataset.py @@ -34,6 +34,9 @@ class HubspotObjectDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str """ @@ -50,6 +53,7 @@ class HubspotObjectDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, } diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_object_dataset_py3.py index f0e8d1a74809..c5efaf9cad56 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_object_dataset_py3.py @@ -34,6 +34,9 @@ class HubspotObjectDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str """ @@ -50,9 +53,10 @@ class HubspotObjectDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: - super(HubspotObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + super(HubspotObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.type = 'HubspotObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/impala_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/impala_object_dataset.py index 634a63f99be7..62e709ebebc3 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/impala_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/impala_object_dataset.py @@ -34,6 +34,9 @@ class ImpalaObjectDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str """ @@ -50,6 +53,7 @@ class ImpalaObjectDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, } diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/impala_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/impala_object_dataset_py3.py index 964bfbd71748..a17b4a08fb26 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/impala_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/impala_object_dataset_py3.py @@ -34,6 +34,9 @@ class ImpalaObjectDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str """ @@ -50,9 +53,10 @@ class ImpalaObjectDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: - super(ImpalaObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + super(ImpalaObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.type = 'ImpalaObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_object_dataset.py index d777d905f3b0..464a640719c4 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_object_dataset.py @@ -34,6 +34,9 @@ class JiraObjectDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str """ @@ -50,6 +53,7 @@ class JiraObjectDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, } diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_object_dataset_py3.py index e0c3007bca89..955eabebb92c 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_object_dataset_py3.py @@ -34,6 +34,9 @@ class JiraObjectDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str """ @@ -50,9 +53,10 @@ class JiraObjectDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: - super(JiraObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + super(JiraObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.type = 'JiraObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/magento_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/magento_object_dataset.py index 2291b8c47796..913ccbb68326 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/magento_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/magento_object_dataset.py @@ -34,6 +34,9 @@ class MagentoObjectDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str """ @@ -50,6 +53,7 @@ class MagentoObjectDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, } diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/magento_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/magento_object_dataset_py3.py index ab3c0bf4f39c..08639eae595c 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/magento_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/magento_object_dataset_py3.py @@ -34,6 +34,9 @@ class MagentoObjectDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str """ @@ -50,9 +53,10 @@ class MagentoObjectDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: - super(MagentoObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + super(MagentoObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.type = 'MagentoObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_table_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_table_dataset.py index 3c4d6d2874a3..5ef7db26acdd 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_table_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_table_dataset.py @@ -34,6 +34,9 @@ class MariaDBTableDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str """ @@ -50,6 +53,7 @@ class MariaDBTableDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, } diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_table_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_table_dataset_py3.py index e557822e5fca..6b9b1149e4f3 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_table_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_table_dataset_py3.py @@ -34,6 +34,9 @@ class MariaDBTableDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str """ @@ -50,9 +53,10 @@ class MariaDBTableDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: - super(MariaDBTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + super(MariaDBTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.type = 'MariaDBTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/marketo_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/marketo_object_dataset.py index ad553cde51b5..ee2b58603215 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/marketo_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/marketo_object_dataset.py @@ -34,6 +34,9 @@ class MarketoObjectDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str """ @@ -50,6 +53,7 @@ class MarketoObjectDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, } diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/marketo_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/marketo_object_dataset_py3.py index dfb08a716949..9218191b80c1 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/marketo_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/marketo_object_dataset_py3.py @@ -34,6 +34,9 @@ class MarketoObjectDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str """ @@ -50,9 +53,10 @@ class MarketoObjectDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: - super(MarketoObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + super(MarketoObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.type = 'MarketoObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/mongo_db_collection_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/mongo_db_collection_dataset.py index c76dc730105a..733d6454492c 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/mongo_db_collection_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/mongo_db_collection_dataset.py @@ -34,6 +34,9 @@ class MongoDbCollectionDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str :param collection_name: Required. The table name of the MongoDB database. @@ -54,6 +57,7 @@ class MongoDbCollectionDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, 'collection_name': {'key': 'typeProperties.collectionName', 'type': 'object'}, } diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/mongo_db_collection_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/mongo_db_collection_dataset_py3.py index dac4b49da87a..7b9de8741d64 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/mongo_db_collection_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/mongo_db_collection_dataset_py3.py @@ -34,6 +34,9 @@ class MongoDbCollectionDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str :param collection_name: Required. The table name of the MongoDB database. @@ -54,11 +57,12 @@ class MongoDbCollectionDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, 'collection_name': {'key': 'typeProperties.collectionName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, collection_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: - super(MongoDbCollectionDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + def __init__(self, *, linked_service_name, collection_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + super(MongoDbCollectionDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.collection_name = collection_name self.type = 'MongoDbCollection' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/netezza_table_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/netezza_table_dataset.py index 399eb3872869..ddbfe1b223b8 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/netezza_table_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/netezza_table_dataset.py @@ -34,6 +34,9 @@ class NetezzaTableDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str """ @@ -50,6 +53,7 @@ class NetezzaTableDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, } diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/netezza_table_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/netezza_table_dataset_py3.py index 28de2143add2..4540fae23f87 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/netezza_table_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/netezza_table_dataset_py3.py @@ -34,6 +34,9 @@ class NetezzaTableDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str """ @@ -50,9 +53,10 @@ class NetezzaTableDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: - super(NetezzaTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + super(NetezzaTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.type = 'NetezzaTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odata_resource_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odata_resource_dataset.py index 9e23abc0dd0a..e1c429b6ad62 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odata_resource_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odata_resource_dataset.py @@ -34,6 +34,9 @@ class ODataResourceDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str :param path: The OData resource path. Type: string (or Expression with @@ -53,6 +56,7 @@ class ODataResourceDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, 'path': {'key': 'typeProperties.path', 'type': 'object'}, } diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odata_resource_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odata_resource_dataset_py3.py index 9cd01c414218..ea103df22f30 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odata_resource_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odata_resource_dataset_py3.py @@ -34,6 +34,9 @@ class ODataResourceDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str :param path: The OData resource path. Type: string (or Expression with @@ -53,11 +56,12 @@ class ODataResourceDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, 'path': {'key': 'typeProperties.path', 'type': 'object'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, path=None, **kwargs) -> None: - super(ODataResourceDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, path=None, **kwargs) -> None: + super(ODataResourceDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.path = path self.type = 'ODataResource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_table_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_table_dataset.py index 78215f783ae3..7362e0bd42cf 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_table_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_table_dataset.py @@ -34,6 +34,9 @@ class OracleTableDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str :param table_name: Required. The table name of the on-premises Oracle @@ -54,6 +57,7 @@ class OracleTableDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_table_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_table_dataset_py3.py index 0c12187802c9..af1e9443e1cb 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_table_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_table_dataset_py3.py @@ -34,6 +34,9 @@ class OracleTableDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str :param table_name: Required. The table name of the on-premises Oracle @@ -54,11 +57,12 @@ class OracleTableDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, table_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: - super(OracleTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + def __init__(self, *, linked_service_name, table_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + super(OracleTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.table_name = table_name self.type = 'OracleTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_object_dataset.py index 6d48d7fb99c3..0fd71aabe0c7 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_object_dataset.py @@ -34,6 +34,9 @@ class PaypalObjectDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str """ @@ -50,6 +53,7 @@ class PaypalObjectDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, } diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_object_dataset_py3.py index cb0e92035dfd..f44966955b76 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_object_dataset_py3.py @@ -34,6 +34,9 @@ class PaypalObjectDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str """ @@ -50,9 +53,10 @@ class PaypalObjectDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: - super(PaypalObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + super(PaypalObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.type = 'PaypalObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/phoenix_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/phoenix_object_dataset.py index b1cd7cf37049..4573ed10a086 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/phoenix_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/phoenix_object_dataset.py @@ -34,6 +34,9 @@ class PhoenixObjectDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str """ @@ -50,6 +53,7 @@ class PhoenixObjectDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, } diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/phoenix_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/phoenix_object_dataset_py3.py index 73cb4ee2184d..b10bf84d2f52 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/phoenix_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/phoenix_object_dataset_py3.py @@ -34,6 +34,9 @@ class PhoenixObjectDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str """ @@ -50,9 +53,10 @@ class PhoenixObjectDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: - super(PhoenixObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + super(PhoenixObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.type = 'PhoenixObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_folder.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_folder.py new file mode 100644 index 000000000000..bebc05cb1824 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_folder.py @@ -0,0 +1,29 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class PipelineFolder(Model): + """The folder that this Pipeline is in. If not specified, Pipeline will appear + at the root level. + + :param name: The name of the folder that this Pipeline is in. + :type name: str + """ + + _attribute_map = { + 'name': {'key': 'name', 'type': 'str'}, + } + + def __init__(self, **kwargs): + super(PipelineFolder, self).__init__(**kwargs) + self.name = kwargs.get('name', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_folder_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_folder_py3.py new file mode 100644 index 000000000000..02c9b8dbbff1 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_folder_py3.py @@ -0,0 +1,29 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class PipelineFolder(Model): + """The folder that this Pipeline is in. If not specified, Pipeline will appear + at the root level. + + :param name: The name of the folder that this Pipeline is in. + :type name: str + """ + + _attribute_map = { + 'name': {'key': 'name', 'type': 'str'}, + } + + def __init__(self, *, name: str=None, **kwargs) -> None: + super(PipelineFolder, self).__init__(**kwargs) + self.name = name diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_resource.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_resource.py index 2531c35a3838..00d1ec8cf81a 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_resource.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_resource.py @@ -41,6 +41,9 @@ class PipelineResource(SubResource): :param annotations: List of tags that can be used for describing the Pipeline. :type annotations: list[object] + :param folder: The folder that this Pipeline is in. If not specified, + Pipeline will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.PipelineFolder """ _validation = { @@ -62,6 +65,7 @@ class PipelineResource(SubResource): 'parameters': {'key': 'properties.parameters', 'type': '{ParameterSpecification}'}, 'concurrency': {'key': 'properties.concurrency', 'type': 'int'}, 'annotations': {'key': 'properties.annotations', 'type': '[object]'}, + 'folder': {'key': 'properties.folder', 'type': 'PipelineFolder'}, } def __init__(self, **kwargs): @@ -72,3 +76,4 @@ def __init__(self, **kwargs): self.parameters = kwargs.get('parameters', None) self.concurrency = kwargs.get('concurrency', None) self.annotations = kwargs.get('annotations', None) + self.folder = kwargs.get('folder', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_resource_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_resource_py3.py index 2b3bf380782a..44eb05b05522 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_resource_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_resource_py3.py @@ -41,6 +41,9 @@ class PipelineResource(SubResource): :param annotations: List of tags that can be used for describing the Pipeline. :type annotations: list[object] + :param folder: The folder that this Pipeline is in. If not specified, + Pipeline will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.PipelineFolder """ _validation = { @@ -62,9 +65,10 @@ class PipelineResource(SubResource): 'parameters': {'key': 'properties.parameters', 'type': '{ParameterSpecification}'}, 'concurrency': {'key': 'properties.concurrency', 'type': 'int'}, 'annotations': {'key': 'properties.annotations', 'type': '[object]'}, + 'folder': {'key': 'properties.folder', 'type': 'PipelineFolder'}, } - def __init__(self, *, additional_properties=None, description: str=None, activities=None, parameters=None, concurrency: int=None, annotations=None, **kwargs) -> None: + def __init__(self, *, additional_properties=None, description: str=None, activities=None, parameters=None, concurrency: int=None, annotations=None, folder=None, **kwargs) -> None: super(PipelineResource, self).__init__(**kwargs) self.additional_properties = additional_properties self.description = description @@ -72,3 +76,4 @@ def __init__(self, *, additional_properties=None, description: str=None, activit self.parameters = parameters self.concurrency = concurrency self.annotations = annotations + self.folder = folder diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/presto_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/presto_object_dataset.py index 863d8851701d..e2215ca65bbe 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/presto_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/presto_object_dataset.py @@ -34,6 +34,9 @@ class PrestoObjectDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str """ @@ -50,6 +53,7 @@ class PrestoObjectDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, } diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/presto_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/presto_object_dataset_py3.py index dfb69ded9c4c..1cb012b082e0 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/presto_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/presto_object_dataset_py3.py @@ -34,6 +34,9 @@ class PrestoObjectDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str """ @@ -50,9 +53,10 @@ class PrestoObjectDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: - super(PrestoObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + super(PrestoObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.type = 'PrestoObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_object_dataset.py index 56e68336c469..7b9a358c5adf 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_object_dataset.py @@ -34,6 +34,9 @@ class QuickBooksObjectDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str """ @@ -50,6 +53,7 @@ class QuickBooksObjectDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, } diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_object_dataset_py3.py index 7897881b5cbb..63ea660c5330 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_object_dataset_py3.py @@ -34,6 +34,9 @@ class QuickBooksObjectDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str """ @@ -50,9 +53,10 @@ class QuickBooksObjectDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: - super(QuickBooksObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + super(QuickBooksObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.type = 'QuickBooksObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/relational_table_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/relational_table_dataset.py index 1037b936b32f..b0c91815e554 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/relational_table_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/relational_table_dataset.py @@ -34,6 +34,9 @@ class RelationalTableDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str :param table_name: The relational table name. Type: string (or Expression @@ -53,6 +56,7 @@ class RelationalTableDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/relational_table_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/relational_table_dataset_py3.py index 0b492686356d..638b0b8a3eef 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/relational_table_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/relational_table_dataset_py3.py @@ -34,6 +34,9 @@ class RelationalTableDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str :param table_name: The relational table name. Type: string (or Expression @@ -53,11 +56,12 @@ class RelationalTableDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, table_name=None, **kwargs) -> None: - super(RelationalTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: + super(RelationalTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.table_name = table_name self.type = 'RelationalTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_object_dataset.py index eef47fbf10a8..84bc7035f7f4 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_object_dataset.py @@ -34,6 +34,9 @@ class ResponsysObjectDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str """ @@ -50,6 +53,7 @@ class ResponsysObjectDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, } diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_object_dataset_py3.py index ffdd38dea188..682db13e5c4a 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_object_dataset_py3.py @@ -34,6 +34,9 @@ class ResponsysObjectDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str """ @@ -50,9 +53,10 @@ class ResponsysObjectDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: - super(ResponsysObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + super(ResponsysObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.type = 'ResponsysObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_marketing_cloud_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_marketing_cloud_object_dataset.py index 7d8a49518ca6..32b02df94ea6 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_marketing_cloud_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_marketing_cloud_object_dataset.py @@ -34,6 +34,9 @@ class SalesforceMarketingCloudObjectDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str """ @@ -50,6 +53,7 @@ class SalesforceMarketingCloudObjectDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, } diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_marketing_cloud_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_marketing_cloud_object_dataset_py3.py index 3180cbd37ea1..6e260fac8a66 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_marketing_cloud_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_marketing_cloud_object_dataset_py3.py @@ -34,6 +34,9 @@ class SalesforceMarketingCloudObjectDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str """ @@ -50,9 +53,10 @@ class SalesforceMarketingCloudObjectDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: - super(SalesforceMarketingCloudObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + super(SalesforceMarketingCloudObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.type = 'SalesforceMarketingCloudObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_object_dataset.py index 8ea94a9f7cd1..a55974c6428c 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_object_dataset.py @@ -34,6 +34,9 @@ class SalesforceObjectDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str :param object_api_name: The Salesforce object API name. Type: string (or @@ -53,6 +56,7 @@ class SalesforceObjectDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, 'object_api_name': {'key': 'typeProperties.objectApiName', 'type': 'object'}, } diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_object_dataset_py3.py index d31a9754f738..fa55f29420dd 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_object_dataset_py3.py @@ -34,6 +34,9 @@ class SalesforceObjectDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str :param object_api_name: The Salesforce object API name. Type: string (or @@ -53,11 +56,12 @@ class SalesforceObjectDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, 'object_api_name': {'key': 'typeProperties.objectApiName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, object_api_name=None, **kwargs) -> None: - super(SalesforceObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, object_api_name=None, **kwargs) -> None: + super(SalesforceObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.object_api_name = object_api_name self.type = 'SalesforceObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_resource_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_resource_dataset.py index 65655e9ce061..b975a3c99cc1 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_resource_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_resource_dataset.py @@ -34,6 +34,9 @@ class SapCloudForCustomerResourceDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str :param path: Required. The path of the SAP Cloud for Customer OData @@ -54,6 +57,7 @@ class SapCloudForCustomerResourceDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, 'path': {'key': 'typeProperties.path', 'type': 'object'}, } diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_resource_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_resource_dataset_py3.py index fc52d1b3bdaf..09ab4693c1fb 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_resource_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_resource_dataset_py3.py @@ -34,6 +34,9 @@ class SapCloudForCustomerResourceDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str :param path: Required. The path of the SAP Cloud for Customer OData @@ -54,11 +57,12 @@ class SapCloudForCustomerResourceDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, 'path': {'key': 'typeProperties.path', 'type': 'object'}, } - def __init__(self, *, linked_service_name, path, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: - super(SapCloudForCustomerResourceDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + def __init__(self, *, linked_service_name, path, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + super(SapCloudForCustomerResourceDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.path = path self.type = 'SapCloudForCustomerResource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_ecc_resource_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_ecc_resource_dataset.py index befc8ea726f9..f26232ea34f6 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_ecc_resource_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_ecc_resource_dataset.py @@ -34,6 +34,9 @@ class SapEccResourceDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str :param path: Required. The path of the SAP ECC OData entity. Type: string @@ -54,6 +57,7 @@ class SapEccResourceDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, 'path': {'key': 'typeProperties.path', 'type': 'str'}, } diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_ecc_resource_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_ecc_resource_dataset_py3.py index 6836e6206041..aced6ccc0f9e 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_ecc_resource_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_ecc_resource_dataset_py3.py @@ -34,6 +34,9 @@ class SapEccResourceDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str :param path: Required. The path of the SAP ECC OData entity. Type: string @@ -54,11 +57,12 @@ class SapEccResourceDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, 'path': {'key': 'typeProperties.path', 'type': 'str'}, } - def __init__(self, *, linked_service_name, path: str, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: - super(SapEccResourceDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + def __init__(self, *, linked_service_name, path: str, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + super(SapEccResourceDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.path = path self.type = 'SapEccResource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_object_dataset.py index df9d7a90f3f8..382d005ad5c5 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_object_dataset.py @@ -34,6 +34,9 @@ class ServiceNowObjectDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str """ @@ -50,6 +53,7 @@ class ServiceNowObjectDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, } diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_object_dataset_py3.py index 97a484fb8826..82d0177ff286 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_object_dataset_py3.py @@ -34,6 +34,9 @@ class ServiceNowObjectDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str """ @@ -50,9 +53,10 @@ class ServiceNowObjectDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: - super(ServiceNowObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + super(ServiceNowObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.type = 'ServiceNowObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_object_dataset.py index 2c64c3d43d8f..a70be7708fb2 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_object_dataset.py @@ -34,6 +34,9 @@ class ShopifyObjectDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str """ @@ -50,6 +53,7 @@ class ShopifyObjectDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, } diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_object_dataset_py3.py index f83ac708a93d..548d10f44028 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_object_dataset_py3.py @@ -34,6 +34,9 @@ class ShopifyObjectDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str """ @@ -50,9 +53,10 @@ class ShopifyObjectDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: - super(ShopifyObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + super(ShopifyObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.type = 'ShopifyObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/spark_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/spark_object_dataset.py index 84a7776e3091..52d31be1f06e 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/spark_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/spark_object_dataset.py @@ -34,6 +34,9 @@ class SparkObjectDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str """ @@ -50,6 +53,7 @@ class SparkObjectDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, } diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/spark_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/spark_object_dataset_py3.py index c12d454f68a7..ceb7634d9237 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/spark_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/spark_object_dataset_py3.py @@ -34,6 +34,9 @@ class SparkObjectDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str """ @@ -50,9 +53,10 @@ class SparkObjectDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: - super(SparkObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + super(SparkObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.type = 'SparkObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_table_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_table_dataset.py index 3e058a5d9506..99afdbf99ce0 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_table_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_table_dataset.py @@ -34,6 +34,9 @@ class SqlServerTableDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str :param table_name: Required. The table name of the SQL Server dataset. @@ -54,6 +57,7 @@ class SqlServerTableDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_table_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_table_dataset_py3.py index 4a2f80144cda..9567fc1e6bb8 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_table_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_table_dataset_py3.py @@ -34,6 +34,9 @@ class SqlServerTableDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str :param table_name: Required. The table name of the SQL Server dataset. @@ -54,11 +57,12 @@ class SqlServerTableDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, table_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: - super(SqlServerTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + def __init__(self, *, linked_service_name, table_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + super(SqlServerTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.table_name = table_name self.type = 'SqlServerTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_object_dataset.py index 789fec1809d2..77d8b7f3f3c3 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_object_dataset.py @@ -34,6 +34,9 @@ class SquareObjectDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str """ @@ -50,6 +53,7 @@ class SquareObjectDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, } diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_object_dataset_py3.py index 879b34c83620..d7ea66c2d81a 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_object_dataset_py3.py @@ -34,6 +34,9 @@ class SquareObjectDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str """ @@ -50,9 +53,10 @@ class SquareObjectDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: - super(SquareObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + super(SquareObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.type = 'SquareObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/vertica_table_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/vertica_table_dataset.py index 98a70f28cee9..755a0225e1f9 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/vertica_table_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/vertica_table_dataset.py @@ -34,6 +34,9 @@ class VerticaTableDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str """ @@ -50,6 +53,7 @@ class VerticaTableDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, } diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/vertica_table_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/vertica_table_dataset_py3.py index 1dd56dfd46ee..cd6f7186855c 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/vertica_table_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/vertica_table_dataset_py3.py @@ -34,6 +34,9 @@ class VerticaTableDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str """ @@ -50,9 +53,10 @@ class VerticaTableDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: - super(VerticaTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + super(VerticaTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.type = 'VerticaTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_table_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_table_dataset.py index 671773cfa70f..a9b30245d2da 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_table_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_table_dataset.py @@ -34,6 +34,9 @@ class WebTableDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str :param index: Required. The zero-based index of the table in the web page. @@ -57,6 +60,7 @@ class WebTableDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, 'index': {'key': 'typeProperties.index', 'type': 'object'}, 'path': {'key': 'typeProperties.path', 'type': 'object'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_table_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_table_dataset_py3.py index 7a361289553c..d5ea9d5f66ba 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_table_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_table_dataset_py3.py @@ -34,6 +34,9 @@ class WebTableDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str :param index: Required. The zero-based index of the table in the web page. @@ -57,13 +60,14 @@ class WebTableDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, 'index': {'key': 'typeProperties.index', 'type': 'object'}, 'path': {'key': 'typeProperties.path', 'type': 'object'}, } - def __init__(self, *, linked_service_name, index, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, path=None, **kwargs) -> None: - super(WebTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + def __init__(self, *, linked_service_name, index, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, path=None, **kwargs) -> None: + super(WebTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.index = index self.path = path self.type = 'WebTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_object_dataset.py index 6ef72bc09878..f3efb679137d 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_object_dataset.py @@ -34,6 +34,9 @@ class XeroObjectDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str """ @@ -50,6 +53,7 @@ class XeroObjectDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, } diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_object_dataset_py3.py index 685be7b0dffa..3c5c970f2310 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_object_dataset_py3.py @@ -34,6 +34,9 @@ class XeroObjectDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str """ @@ -50,9 +53,10 @@ class XeroObjectDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: - super(XeroObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + super(XeroObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.type = 'XeroObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/zoho_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/zoho_object_dataset.py index 26ab79ab4301..fb988d8e5e04 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/zoho_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/zoho_object_dataset.py @@ -34,6 +34,9 @@ class ZohoObjectDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str """ @@ -50,6 +53,7 @@ class ZohoObjectDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, } diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/zoho_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/zoho_object_dataset_py3.py index 40a579665206..32cac965c535 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/zoho_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/zoho_object_dataset_py3.py @@ -34,6 +34,9 @@ class ZohoObjectDataset(Dataset): :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str """ @@ -50,9 +53,10 @@ class ZohoObjectDataset(Dataset): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, **kwargs) -> None: - super(ZohoObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, **kwargs) + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + super(ZohoObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.type = 'ZohoObject' From 51d00f50a661afb623b5a0615381bc5868d005bf Mon Sep 17 00:00:00 2001 From: Azure SDK for Python bot Date: Tue, 28 Aug 2018 19:41:19 -0700 Subject: [PATCH 17/54] =?UTF-8?q?[AutoPR=20datafactory/resource-manager]?= =?UTF-8?q?=20revert=20tumbling=20window=20dependsOn=20property=20under=20?= =?UTF-8?q?preview,=20and=20apply=20sa=E2=80=A6=20(#2956)?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit * Generated from da77fb49b0bf7a12fd028b3d66a8ffe18bdbcf83 revert tumbling window dependsOn property under preview, and apply same changes to stable * Generated from 12537688f45a6f7fcf8d336becf41b51230a365f update description * Generated from 12537688f45a6f7fcf8d336becf41b51230a365f update description --- .../azure/mgmt/datafactory/models/__init__.py | 15 ++++++ .../models/dependency_reference.py | 42 ++++++++++++++++ .../models/dependency_reference_py3.py | 42 ++++++++++++++++ ...dency_tumbling_window_trigger_reference.py | 46 +++++++++++++++++ ...y_tumbling_window_trigger_reference_py3.py | 46 +++++++++++++++++ .../models/trigger_dependency_reference.py | 46 +++++++++++++++++ .../trigger_dependency_reference_py3.py | 46 +++++++++++++++++ .../datafactory/models/trigger_reference.py | 44 ++++++++++++++++ .../models/trigger_reference_py3.py | 44 ++++++++++++++++ .../models/tumbling_window_trigger.py | 5 ++ ...ing_window_trigger_dependency_reference.py | 50 +++++++++++++++++++ ...window_trigger_dependency_reference_py3.py | 50 +++++++++++++++++++ .../models/tumbling_window_trigger_py3.py | 7 ++- 13 files changed, 482 insertions(+), 1 deletion(-) create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/dependency_reference.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/dependency_reference_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_dependency_tumbling_window_trigger_reference.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_dependency_tumbling_window_trigger_reference_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger_dependency_reference.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger_dependency_reference_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger_reference.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger_reference_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/tumbling_window_trigger_dependency_reference.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/tumbling_window_trigger_dependency_reference_py3.py diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py index e20345f7ab5b..e2a2058cf443 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py @@ -208,6 +208,11 @@ from .azure_table_dataset_py3 import AzureTableDataset from .azure_blob_dataset_py3 import AzureBlobDataset from .amazon_s3_dataset_py3 import AmazonS3Dataset + from .self_dependency_tumbling_window_trigger_reference_py3 import SelfDependencyTumblingWindowTriggerReference + from .trigger_reference_py3 import TriggerReference + from .tumbling_window_trigger_dependency_reference_py3 import TumblingWindowTriggerDependencyReference + from .trigger_dependency_reference_py3 import TriggerDependencyReference + from .dependency_reference_py3 import DependencyReference from .retry_policy_py3 import RetryPolicy from .tumbling_window_trigger_py3 import TumblingWindowTrigger from .blob_events_trigger_py3 import BlobEventsTrigger @@ -546,6 +551,11 @@ from .azure_table_dataset import AzureTableDataset from .azure_blob_dataset import AzureBlobDataset from .amazon_s3_dataset import AmazonS3Dataset + from .self_dependency_tumbling_window_trigger_reference import SelfDependencyTumblingWindowTriggerReference + from .trigger_reference import TriggerReference + from .tumbling_window_trigger_dependency_reference import TumblingWindowTriggerDependencyReference + from .trigger_dependency_reference import TriggerDependencyReference + from .dependency_reference import DependencyReference from .retry_policy import RetryPolicy from .tumbling_window_trigger import TumblingWindowTrigger from .blob_events_trigger import BlobEventsTrigger @@ -952,6 +962,11 @@ 'AzureTableDataset', 'AzureBlobDataset', 'AmazonS3Dataset', + 'SelfDependencyTumblingWindowTriggerReference', + 'TriggerReference', + 'TumblingWindowTriggerDependencyReference', + 'TriggerDependencyReference', + 'DependencyReference', 'RetryPolicy', 'TumblingWindowTrigger', 'BlobEventsTrigger', diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dependency_reference.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dependency_reference.py new file mode 100644 index 000000000000..89e750df8f0d --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dependency_reference.py @@ -0,0 +1,42 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class DependencyReference(Model): + """Referenced dependency. + + You probably want to use the sub-classes and not this class directly. Known + sub-classes are: SelfDependencyTumblingWindowTriggerReference, + TriggerDependencyReference + + All required parameters must be populated in order to send to Azure. + + :param type: Required. Constant filled by server. + :type type: str + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'type': {'key': 'type', 'type': 'str'}, + } + + _subtype_map = { + 'type': {'SelfDependencyTumblingWindowTriggerReference': 'SelfDependencyTumblingWindowTriggerReference', 'TriggerDependencyReference': 'TriggerDependencyReference'} + } + + def __init__(self, **kwargs): + super(DependencyReference, self).__init__(**kwargs) + self.type = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dependency_reference_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dependency_reference_py3.py new file mode 100644 index 000000000000..1b0647b74991 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dependency_reference_py3.py @@ -0,0 +1,42 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class DependencyReference(Model): + """Referenced dependency. + + You probably want to use the sub-classes and not this class directly. Known + sub-classes are: SelfDependencyTumblingWindowTriggerReference, + TriggerDependencyReference + + All required parameters must be populated in order to send to Azure. + + :param type: Required. Constant filled by server. + :type type: str + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'type': {'key': 'type', 'type': 'str'}, + } + + _subtype_map = { + 'type': {'SelfDependencyTumblingWindowTriggerReference': 'SelfDependencyTumblingWindowTriggerReference', 'TriggerDependencyReference': 'TriggerDependencyReference'} + } + + def __init__(self, **kwargs) -> None: + super(DependencyReference, self).__init__(**kwargs) + self.type = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_dependency_tumbling_window_trigger_reference.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_dependency_tumbling_window_trigger_reference.py new file mode 100644 index 000000000000..fc56f8e8a799 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_dependency_tumbling_window_trigger_reference.py @@ -0,0 +1,46 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dependency_reference import DependencyReference + + +class SelfDependencyTumblingWindowTriggerReference(DependencyReference): + """Self referenced tumbling window trigger dependency. + + All required parameters must be populated in order to send to Azure. + + :param type: Required. Constant filled by server. + :type type: str + :param offset: Required. Timespan applied to the start time of a tumbling + window when evaluating dependency. + :type offset: str + :param size: The size of the window when evaluating the dependency. If + undefined the frequency of the tumbling window will be used. + :type size: str + """ + + _validation = { + 'type': {'required': True}, + 'offset': {'required': True, 'max_length': 15, 'min_length': 8, 'pattern': r'((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9]))'}, + 'size': {'max_length': 15, 'min_length': 8, 'pattern': r'((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9]))'}, + } + + _attribute_map = { + 'type': {'key': 'type', 'type': 'str'}, + 'offset': {'key': 'offset', 'type': 'str'}, + 'size': {'key': 'size', 'type': 'str'}, + } + + def __init__(self, **kwargs): + super(SelfDependencyTumblingWindowTriggerReference, self).__init__(**kwargs) + self.offset = kwargs.get('offset', None) + self.size = kwargs.get('size', None) + self.type = 'SelfDependencyTumblingWindowTriggerReference' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_dependency_tumbling_window_trigger_reference_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_dependency_tumbling_window_trigger_reference_py3.py new file mode 100644 index 000000000000..1dd1e575c2e8 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/self_dependency_tumbling_window_trigger_reference_py3.py @@ -0,0 +1,46 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dependency_reference_py3 import DependencyReference + + +class SelfDependencyTumblingWindowTriggerReference(DependencyReference): + """Self referenced tumbling window trigger dependency. + + All required parameters must be populated in order to send to Azure. + + :param type: Required. Constant filled by server. + :type type: str + :param offset: Required. Timespan applied to the start time of a tumbling + window when evaluating dependency. + :type offset: str + :param size: The size of the window when evaluating the dependency. If + undefined the frequency of the tumbling window will be used. + :type size: str + """ + + _validation = { + 'type': {'required': True}, + 'offset': {'required': True, 'max_length': 15, 'min_length': 8, 'pattern': r'((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9]))'}, + 'size': {'max_length': 15, 'min_length': 8, 'pattern': r'((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9]))'}, + } + + _attribute_map = { + 'type': {'key': 'type', 'type': 'str'}, + 'offset': {'key': 'offset', 'type': 'str'}, + 'size': {'key': 'size', 'type': 'str'}, + } + + def __init__(self, *, offset: str, size: str=None, **kwargs) -> None: + super(SelfDependencyTumblingWindowTriggerReference, self).__init__(**kwargs) + self.offset = offset + self.size = size + self.type = 'SelfDependencyTumblingWindowTriggerReference' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger_dependency_reference.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger_dependency_reference.py new file mode 100644 index 000000000000..089aa9a3e5fc --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger_dependency_reference.py @@ -0,0 +1,46 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dependency_reference import DependencyReference + + +class TriggerDependencyReference(DependencyReference): + """Trigger referenced dependency. + + You probably want to use the sub-classes and not this class directly. Known + sub-classes are: TumblingWindowTriggerDependencyReference + + All required parameters must be populated in order to send to Azure. + + :param type: Required. Constant filled by server. + :type type: str + :param reference_trigger: Required. Referenced trigger. + :type reference_trigger: ~azure.mgmt.datafactory.models.TriggerReference + """ + + _validation = { + 'type': {'required': True}, + 'reference_trigger': {'required': True}, + } + + _attribute_map = { + 'type': {'key': 'type', 'type': 'str'}, + 'reference_trigger': {'key': 'referenceTrigger', 'type': 'TriggerReference'}, + } + + _subtype_map = { + 'type': {'TumblingWindowTriggerDependencyReference': 'TumblingWindowTriggerDependencyReference'} + } + + def __init__(self, **kwargs): + super(TriggerDependencyReference, self).__init__(**kwargs) + self.reference_trigger = kwargs.get('reference_trigger', None) + self.type = 'TriggerDependencyReference' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger_dependency_reference_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger_dependency_reference_py3.py new file mode 100644 index 000000000000..716a0d926f8b --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger_dependency_reference_py3.py @@ -0,0 +1,46 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dependency_reference_py3 import DependencyReference + + +class TriggerDependencyReference(DependencyReference): + """Trigger referenced dependency. + + You probably want to use the sub-classes and not this class directly. Known + sub-classes are: TumblingWindowTriggerDependencyReference + + All required parameters must be populated in order to send to Azure. + + :param type: Required. Constant filled by server. + :type type: str + :param reference_trigger: Required. Referenced trigger. + :type reference_trigger: ~azure.mgmt.datafactory.models.TriggerReference + """ + + _validation = { + 'type': {'required': True}, + 'reference_trigger': {'required': True}, + } + + _attribute_map = { + 'type': {'key': 'type', 'type': 'str'}, + 'reference_trigger': {'key': 'referenceTrigger', 'type': 'TriggerReference'}, + } + + _subtype_map = { + 'type': {'TumblingWindowTriggerDependencyReference': 'TumblingWindowTriggerDependencyReference'} + } + + def __init__(self, *, reference_trigger, **kwargs) -> None: + super(TriggerDependencyReference, self).__init__(**kwargs) + self.reference_trigger = reference_trigger + self.type = 'TriggerDependencyReference' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger_reference.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger_reference.py new file mode 100644 index 000000000000..a4f952dac85f --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger_reference.py @@ -0,0 +1,44 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class TriggerReference(Model): + """Trigger reference type. + + Variables are only populated by the server, and will be ignored when + sending a request. + + All required parameters must be populated in order to send to Azure. + + :ivar type: Required. Trigger reference type. Default value: + "TriggerReference" . + :vartype type: str + :param reference_name: Required. Reference trigger name. + :type reference_name: str + """ + + _validation = { + 'type': {'required': True, 'constant': True}, + 'reference_name': {'required': True}, + } + + _attribute_map = { + 'type': {'key': 'type', 'type': 'str'}, + 'reference_name': {'key': 'referenceName', 'type': 'str'}, + } + + type = "TriggerReference" + + def __init__(self, **kwargs): + super(TriggerReference, self).__init__(**kwargs) + self.reference_name = kwargs.get('reference_name', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger_reference_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger_reference_py3.py new file mode 100644 index 000000000000..805e407e80a7 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger_reference_py3.py @@ -0,0 +1,44 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class TriggerReference(Model): + """Trigger reference type. + + Variables are only populated by the server, and will be ignored when + sending a request. + + All required parameters must be populated in order to send to Azure. + + :ivar type: Required. Trigger reference type. Default value: + "TriggerReference" . + :vartype type: str + :param reference_name: Required. Reference trigger name. + :type reference_name: str + """ + + _validation = { + 'type': {'required': True, 'constant': True}, + 'reference_name': {'required': True}, + } + + _attribute_map = { + 'type': {'key': 'type', 'type': 'str'}, + 'reference_name': {'key': 'referenceName', 'type': 'str'}, + } + + type = "TriggerReference" + + def __init__(self, *, reference_name: str, **kwargs) -> None: + super(TriggerReference, self).__init__(**kwargs) + self.reference_name = reference_name diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/tumbling_window_trigger.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/tumbling_window_trigger.py index ec6d09496677..ce46a4aac7e2 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/tumbling_window_trigger.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/tumbling_window_trigger.py @@ -63,6 +63,9 @@ class TumblingWindowTrigger(Trigger): :param retry_policy: Retry policy that will be applied for failed pipeline runs. :type retry_policy: ~azure.mgmt.datafactory.models.RetryPolicy + :param depends_on: Triggers that this trigger depends on. Only tumbling + window triggers are supported. + :type depends_on: list[~azure.mgmt.datafactory.models.DependencyReference] """ _validation = { @@ -88,6 +91,7 @@ class TumblingWindowTrigger(Trigger): 'delay': {'key': 'typeProperties.delay', 'type': 'object'}, 'max_concurrency': {'key': 'typeProperties.maxConcurrency', 'type': 'int'}, 'retry_policy': {'key': 'typeProperties.retryPolicy', 'type': 'RetryPolicy'}, + 'depends_on': {'key': 'typeProperties.dependsOn', 'type': '[DependencyReference]'}, } def __init__(self, **kwargs): @@ -100,4 +104,5 @@ def __init__(self, **kwargs): self.delay = kwargs.get('delay', None) self.max_concurrency = kwargs.get('max_concurrency', None) self.retry_policy = kwargs.get('retry_policy', None) + self.depends_on = kwargs.get('depends_on', None) self.type = 'TumblingWindowTrigger' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/tumbling_window_trigger_dependency_reference.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/tumbling_window_trigger_dependency_reference.py new file mode 100644 index 000000000000..89dcefbc8c09 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/tumbling_window_trigger_dependency_reference.py @@ -0,0 +1,50 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .trigger_dependency_reference import TriggerDependencyReference + + +class TumblingWindowTriggerDependencyReference(TriggerDependencyReference): + """Referenced tumbling window trigger dependency. + + All required parameters must be populated in order to send to Azure. + + :param type: Required. Constant filled by server. + :type type: str + :param reference_trigger: Required. Referenced trigger. + :type reference_trigger: ~azure.mgmt.datafactory.models.TriggerReference + :param offset: Timespan applied to the start time of a tumbling window + when evaluating dependency. + :type offset: str + :param size: The size of the window when evaluating the dependency. If + undefined the frequency of the tumbling window will be used. + :type size: str + """ + + _validation = { + 'type': {'required': True}, + 'reference_trigger': {'required': True}, + 'offset': {'max_length': 15, 'min_length': 8, 'pattern': r'((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9]))'}, + 'size': {'max_length': 15, 'min_length': 8, 'pattern': r'((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9]))'}, + } + + _attribute_map = { + 'type': {'key': 'type', 'type': 'str'}, + 'reference_trigger': {'key': 'referenceTrigger', 'type': 'TriggerReference'}, + 'offset': {'key': 'offset', 'type': 'str'}, + 'size': {'key': 'size', 'type': 'str'}, + } + + def __init__(self, **kwargs): + super(TumblingWindowTriggerDependencyReference, self).__init__(**kwargs) + self.offset = kwargs.get('offset', None) + self.size = kwargs.get('size', None) + self.type = 'TumblingWindowTriggerDependencyReference' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/tumbling_window_trigger_dependency_reference_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/tumbling_window_trigger_dependency_reference_py3.py new file mode 100644 index 000000000000..648f25e59937 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/tumbling_window_trigger_dependency_reference_py3.py @@ -0,0 +1,50 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .trigger_dependency_reference_py3 import TriggerDependencyReference + + +class TumblingWindowTriggerDependencyReference(TriggerDependencyReference): + """Referenced tumbling window trigger dependency. + + All required parameters must be populated in order to send to Azure. + + :param type: Required. Constant filled by server. + :type type: str + :param reference_trigger: Required. Referenced trigger. + :type reference_trigger: ~azure.mgmt.datafactory.models.TriggerReference + :param offset: Timespan applied to the start time of a tumbling window + when evaluating dependency. + :type offset: str + :param size: The size of the window when evaluating the dependency. If + undefined the frequency of the tumbling window will be used. + :type size: str + """ + + _validation = { + 'type': {'required': True}, + 'reference_trigger': {'required': True}, + 'offset': {'max_length': 15, 'min_length': 8, 'pattern': r'((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9]))'}, + 'size': {'max_length': 15, 'min_length': 8, 'pattern': r'((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9]))'}, + } + + _attribute_map = { + 'type': {'key': 'type', 'type': 'str'}, + 'reference_trigger': {'key': 'referenceTrigger', 'type': 'TriggerReference'}, + 'offset': {'key': 'offset', 'type': 'str'}, + 'size': {'key': 'size', 'type': 'str'}, + } + + def __init__(self, *, reference_trigger, offset: str=None, size: str=None, **kwargs) -> None: + super(TumblingWindowTriggerDependencyReference, self).__init__(reference_trigger=reference_trigger, **kwargs) + self.offset = offset + self.size = size + self.type = 'TumblingWindowTriggerDependencyReference' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/tumbling_window_trigger_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/tumbling_window_trigger_py3.py index 9db68bc272d0..bc3114f08edd 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/tumbling_window_trigger_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/tumbling_window_trigger_py3.py @@ -63,6 +63,9 @@ class TumblingWindowTrigger(Trigger): :param retry_policy: Retry policy that will be applied for failed pipeline runs. :type retry_policy: ~azure.mgmt.datafactory.models.RetryPolicy + :param depends_on: Triggers that this trigger depends on. Only tumbling + window triggers are supported. + :type depends_on: list[~azure.mgmt.datafactory.models.DependencyReference] """ _validation = { @@ -88,9 +91,10 @@ class TumblingWindowTrigger(Trigger): 'delay': {'key': 'typeProperties.delay', 'type': 'object'}, 'max_concurrency': {'key': 'typeProperties.maxConcurrency', 'type': 'int'}, 'retry_policy': {'key': 'typeProperties.retryPolicy', 'type': 'RetryPolicy'}, + 'depends_on': {'key': 'typeProperties.dependsOn', 'type': '[DependencyReference]'}, } - def __init__(self, *, pipeline, frequency, interval: int, start_time, max_concurrency: int, additional_properties=None, description: str=None, end_time=None, delay=None, retry_policy=None, **kwargs) -> None: + def __init__(self, *, pipeline, frequency, interval: int, start_time, max_concurrency: int, additional_properties=None, description: str=None, end_time=None, delay=None, retry_policy=None, depends_on=None, **kwargs) -> None: super(TumblingWindowTrigger, self).__init__(additional_properties=additional_properties, description=description, **kwargs) self.pipeline = pipeline self.frequency = frequency @@ -100,4 +104,5 @@ def __init__(self, *, pipeline, frequency, interval: int, start_time, max_concur self.delay = delay self.max_concurrency = max_concurrency self.retry_policy = retry_policy + self.depends_on = depends_on self.type = 'TumblingWindowTrigger' From 07ede438e5aee2d1092cfb8763f4165c1dc11083 Mon Sep 17 00:00:00 2001 From: Azure SDK for Python bot Date: Thu, 6 Sep 2018 11:13:48 -0700 Subject: [PATCH 18/54] Generated from 8b87f503ecc9be21483bf4836e503a4a39655629 (#3276) update FactoryGitHubConfiguration discriminator value --- .../mgmt/datafactory/models/factory_git_hub_configuration.py | 2 +- .../datafactory/models/factory_git_hub_configuration_py3.py | 2 +- .../azure/mgmt/datafactory/models/factory_repo_configuration.py | 2 +- .../mgmt/datafactory/models/factory_repo_configuration_py3.py | 2 +- 4 files changed, 4 insertions(+), 4 deletions(-) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_git_hub_configuration.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_git_hub_configuration.py index f870e4c10229..2607d41d510b 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_git_hub_configuration.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_git_hub_configuration.py @@ -55,4 +55,4 @@ class FactoryGitHubConfiguration(FactoryRepoConfiguration): def __init__(self, **kwargs): super(FactoryGitHubConfiguration, self).__init__(**kwargs) self.host_name = kwargs.get('host_name', None) - self.type = 'FactoryGithubConfiguration' + self.type = 'FactoryGitHubConfiguration' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_git_hub_configuration_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_git_hub_configuration_py3.py index 7f9b2f210f83..8948a4e4420b 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_git_hub_configuration_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_git_hub_configuration_py3.py @@ -55,4 +55,4 @@ class FactoryGitHubConfiguration(FactoryRepoConfiguration): def __init__(self, *, account_name: str, repository_name: str, collaboration_branch: str, root_folder: str, last_commit_id: str=None, host_name: str=None, **kwargs) -> None: super(FactoryGitHubConfiguration, self).__init__(account_name=account_name, repository_name=repository_name, collaboration_branch=collaboration_branch, root_folder=root_folder, last_commit_id=last_commit_id, **kwargs) self.host_name = host_name - self.type = 'FactoryGithubConfiguration' + self.type = 'FactoryGitHubConfiguration' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_repo_configuration.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_repo_configuration.py index 2c7175f3a58c..753bd2bdb039 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_repo_configuration.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_repo_configuration.py @@ -52,7 +52,7 @@ class FactoryRepoConfiguration(Model): } _subtype_map = { - 'type': {'FactoryVSTSConfiguration': 'FactoryVSTSConfiguration', 'FactoryGithubConfiguration': 'FactoryGitHubConfiguration'} + 'type': {'FactoryVSTSConfiguration': 'FactoryVSTSConfiguration', 'FactoryGitHubConfiguration': 'FactoryGitHubConfiguration'} } def __init__(self, **kwargs): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_repo_configuration_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_repo_configuration_py3.py index c974f2d5e7da..c724cb64a18d 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_repo_configuration_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_repo_configuration_py3.py @@ -52,7 +52,7 @@ class FactoryRepoConfiguration(Model): } _subtype_map = { - 'type': {'FactoryVSTSConfiguration': 'FactoryVSTSConfiguration', 'FactoryGithubConfiguration': 'FactoryGitHubConfiguration'} + 'type': {'FactoryVSTSConfiguration': 'FactoryVSTSConfiguration', 'FactoryGitHubConfiguration': 'FactoryGitHubConfiguration'} } def __init__(self, *, account_name: str, repository_name: str, collaboration_branch: str, root_folder: str, last_commit_id: str=None, **kwargs) -> None: From b3a9d0c6f2cbd202297d1a51330249cbbcd237d8 Mon Sep 17 00:00:00 2001 From: azuresdkci Date: Thu, 6 Sep 2018 18:47:41 +0000 Subject: [PATCH 19/54] Packaging update of azure-mgmt-datafactory --- azure-mgmt-datafactory/README.rst | 8 ++++---- azure-mgmt-datafactory/sdk_packaging.toml | 6 ++++++ azure-mgmt-datafactory/setup.py | 6 ++++-- 3 files changed, 14 insertions(+), 6 deletions(-) create mode 100644 azure-mgmt-datafactory/sdk_packaging.toml diff --git a/azure-mgmt-datafactory/README.rst b/azure-mgmt-datafactory/README.rst index 1d331b91cec5..45de4e6b43dd 100644 --- a/azure-mgmt-datafactory/README.rst +++ b/azure-mgmt-datafactory/README.rst @@ -1,12 +1,12 @@ Microsoft Azure SDK for Python ============================== -This is the Microsoft Azure Data Factory Management Client Library. +This is the Microsoft Azure MyService Management Client Library. Azure Resource Manager (ARM) is the next generation of management APIs that replace the old Azure Service Management (ASM). -This package has been tested with Python 2.7, 3.4, 3.5 and 3.6. +This package has been tested with Python 2.7, 3.4, 3.5, 3.6 and 3.7. For the older Azure Service Management (ASM) libraries, see `azure-servicemanagement-legacy `__ library. @@ -36,8 +36,8 @@ If you see azure==0.11.0 (or any version below 1.0), uninstall it first: Usage ===== -For code examples, see `DataFactory Management -`__ +For code examples, see `MyService Management +`__ on docs.microsoft.com. diff --git a/azure-mgmt-datafactory/sdk_packaging.toml b/azure-mgmt-datafactory/sdk_packaging.toml new file mode 100644 index 000000000000..5c75d690daf1 --- /dev/null +++ b/azure-mgmt-datafactory/sdk_packaging.toml @@ -0,0 +1,6 @@ +[packaging] +package_name = "azure-mgmt-datafactory" +package_pprint_name = "MyService Management" +package_doc_id = "" +is_stable = false +is_arm = true diff --git a/azure-mgmt-datafactory/setup.py b/azure-mgmt-datafactory/setup.py index 94a4f2de250d..98b696f39617 100644 --- a/azure-mgmt-datafactory/setup.py +++ b/azure-mgmt-datafactory/setup.py @@ -19,7 +19,7 @@ # Change the PACKAGE_NAME only to change folder and different name PACKAGE_NAME = "azure-mgmt-datafactory" -PACKAGE_PPRINT_NAME = "Data Factory Management" +PACKAGE_PPRINT_NAME = "MyService Management" # a-b-c => a/b/c package_folder_path = PACKAGE_NAME.replace('-', '/') @@ -72,12 +72,14 @@ 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', 'Programming Language :: Python :: 3.6', + 'Programming Language :: Python :: 3.7', 'License :: OSI Approved :: MIT License', ], zip_safe=False, packages=find_packages(exclude=["tests"]), install_requires=[ - 'msrestazure>=0.4.20,<2.0.0', + 'msrest>=0.5.0', + 'msrestazure>=0.4.32,<2.0.0', 'azure-common~=1.1', ], cmdclass=cmdclass From 5108e0f62fc7cdedcb06a871fa54bfa6f6de3df2 Mon Sep 17 00:00:00 2001 From: Azure SDK for Python bot Date: Fri, 7 Sep 2018 13:38:49 -0700 Subject: [PATCH 20/54] Generated from e2194ea26e21490f9c88bf4c63868716b5b772d7 (#3290) Removing the example for ExposureControl API --- .../azure/mgmt/datafactory/models/__init__.py | 6 ++ .../models/git_hub_access_token_request.py | 44 ++++++++++++ .../git_hub_access_token_request_py3.py | 44 ++++++++++++ .../models/git_hub_access_token_response.py | 28 ++++++++ .../git_hub_access_token_response_py3.py | 28 ++++++++ .../operations/factories_operations.py | 70 +++++++++++++++++++ 6 files changed, 220 insertions(+) create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/git_hub_access_token_request.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/git_hub_access_token_request_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/git_hub_access_token_response.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/git_hub_access_token_response_py3.py diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py index e2a2058cf443..436179448b73 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py @@ -47,6 +47,8 @@ from .factory_vsts_configuration_py3 import FactoryVSTSConfiguration from .factory_git_hub_configuration_py3 import FactoryGitHubConfiguration from .factory_repo_update_py3 import FactoryRepoUpdate + from .git_hub_access_token_request_py3 import GitHubAccessTokenRequest + from .git_hub_access_token_response_py3 import GitHubAccessTokenResponse from .pipeline_reference_py3 import PipelineReference from .trigger_pipeline_reference_py3 import TriggerPipelineReference from .factory_update_parameters_py3 import FactoryUpdateParameters @@ -390,6 +392,8 @@ from .factory_vsts_configuration import FactoryVSTSConfiguration from .factory_git_hub_configuration import FactoryGitHubConfiguration from .factory_repo_update import FactoryRepoUpdate + from .git_hub_access_token_request import GitHubAccessTokenRequest + from .git_hub_access_token_response import GitHubAccessTokenResponse from .pipeline_reference import PipelineReference from .trigger_pipeline_reference import TriggerPipelineReference from .factory_update_parameters import FactoryUpdateParameters @@ -801,6 +805,8 @@ 'FactoryVSTSConfiguration', 'FactoryGitHubConfiguration', 'FactoryRepoUpdate', + 'GitHubAccessTokenRequest', + 'GitHubAccessTokenResponse', 'PipelineReference', 'TriggerPipelineReference', 'FactoryUpdateParameters', diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/git_hub_access_token_request.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/git_hub_access_token_request.py new file mode 100644 index 000000000000..cadecdf70f44 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/git_hub_access_token_request.py @@ -0,0 +1,44 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class GitHubAccessTokenRequest(Model): + """Get GitHub access token request definition. + + All required parameters must be populated in order to send to Azure. + + :param git_hub_access_code: Required. GitHub access code. + :type git_hub_access_code: str + :param git_hub_client_id: GitHub application client ID. + :type git_hub_client_id: str + :param git_hub_access_token_base_url: Required. GitHub access token base + URL. + :type git_hub_access_token_base_url: str + """ + + _validation = { + 'git_hub_access_code': {'required': True}, + 'git_hub_access_token_base_url': {'required': True}, + } + + _attribute_map = { + 'git_hub_access_code': {'key': 'gitHubAccessCode', 'type': 'str'}, + 'git_hub_client_id': {'key': 'gitHubClientId', 'type': 'str'}, + 'git_hub_access_token_base_url': {'key': 'gitHubAccessTokenBaseUrl', 'type': 'str'}, + } + + def __init__(self, **kwargs): + super(GitHubAccessTokenRequest, self).__init__(**kwargs) + self.git_hub_access_code = kwargs.get('git_hub_access_code', None) + self.git_hub_client_id = kwargs.get('git_hub_client_id', None) + self.git_hub_access_token_base_url = kwargs.get('git_hub_access_token_base_url', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/git_hub_access_token_request_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/git_hub_access_token_request_py3.py new file mode 100644 index 000000000000..7961e1bc33ed --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/git_hub_access_token_request_py3.py @@ -0,0 +1,44 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class GitHubAccessTokenRequest(Model): + """Get GitHub access token request definition. + + All required parameters must be populated in order to send to Azure. + + :param git_hub_access_code: Required. GitHub access code. + :type git_hub_access_code: str + :param git_hub_client_id: GitHub application client ID. + :type git_hub_client_id: str + :param git_hub_access_token_base_url: Required. GitHub access token base + URL. + :type git_hub_access_token_base_url: str + """ + + _validation = { + 'git_hub_access_code': {'required': True}, + 'git_hub_access_token_base_url': {'required': True}, + } + + _attribute_map = { + 'git_hub_access_code': {'key': 'gitHubAccessCode', 'type': 'str'}, + 'git_hub_client_id': {'key': 'gitHubClientId', 'type': 'str'}, + 'git_hub_access_token_base_url': {'key': 'gitHubAccessTokenBaseUrl', 'type': 'str'}, + } + + def __init__(self, *, git_hub_access_code: str, git_hub_access_token_base_url: str, git_hub_client_id: str=None, **kwargs) -> None: + super(GitHubAccessTokenRequest, self).__init__(**kwargs) + self.git_hub_access_code = git_hub_access_code + self.git_hub_client_id = git_hub_client_id + self.git_hub_access_token_base_url = git_hub_access_token_base_url diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/git_hub_access_token_response.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/git_hub_access_token_response.py new file mode 100644 index 000000000000..4a4afce8f0f0 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/git_hub_access_token_response.py @@ -0,0 +1,28 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class GitHubAccessTokenResponse(Model): + """Get GitHub access token response definition. + + :param git_hub_access_token: GitHub access token. + :type git_hub_access_token: str + """ + + _attribute_map = { + 'git_hub_access_token': {'key': 'gitHubAccessToken', 'type': 'str'}, + } + + def __init__(self, **kwargs): + super(GitHubAccessTokenResponse, self).__init__(**kwargs) + self.git_hub_access_token = kwargs.get('git_hub_access_token', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/git_hub_access_token_response_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/git_hub_access_token_response_py3.py new file mode 100644 index 000000000000..4f28ade6e914 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/git_hub_access_token_response_py3.py @@ -0,0 +1,28 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class GitHubAccessTokenResponse(Model): + """Get GitHub access token response definition. + + :param git_hub_access_token: GitHub access token. + :type git_hub_access_token: str + """ + + _attribute_map = { + 'git_hub_access_token': {'key': 'gitHubAccessToken', 'type': 'str'}, + } + + def __init__(self, *, git_hub_access_token: str=None, **kwargs) -> None: + super(GitHubAccessTokenResponse, self).__init__(**kwargs) + self.git_hub_access_token = git_hub_access_token diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/factories_operations.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/factories_operations.py index 587cb30a87d1..975421b1e222 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/factories_operations.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/factories_operations.py @@ -504,3 +504,73 @@ def delete( client_raw_response = ClientRawResponse(None, response) return client_raw_response delete.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}'} + + def get_git_hub_access_token( + self, resource_group_name, factory_name, git_hub_access_token_request, custom_headers=None, raw=False, **operation_config): + """Get GitHub Access Token. + + :param resource_group_name: The resource group name. + :type resource_group_name: str + :param factory_name: The factory name. + :type factory_name: str + :param git_hub_access_token_request: Get GitHub access token request + definition. + :type git_hub_access_token_request: + ~azure.mgmt.datafactory.models.GitHubAccessTokenRequest + :param dict custom_headers: headers that will be added to the request + :param bool raw: returns the direct response alongside the + deserialized response + :param operation_config: :ref:`Operation configuration + overrides`. + :return: GitHubAccessTokenResponse or ClientRawResponse if raw=true + :rtype: ~azure.mgmt.datafactory.models.GitHubAccessTokenResponse or + ~msrest.pipeline.ClientRawResponse + :raises: :class:`CloudError` + """ + # Construct URL + url = self.get_git_hub_access_token.metadata['url'] + path_format_arguments = { + 'subscriptionId': self._serialize.url("self.config.subscription_id", self.config.subscription_id, 'str'), + 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str', max_length=90, min_length=1, pattern=r'^[-\w\._\(\)]+$'), + 'factoryName': self._serialize.url("factory_name", factory_name, 'str', max_length=63, min_length=3, pattern=r'^[A-Za-z0-9]+(?:-[A-Za-z0-9]+)*$') + } + url = self._client.format_url(url, **path_format_arguments) + + # Construct parameters + query_parameters = {} + query_parameters['api-version'] = self._serialize.query("self.api_version", self.api_version, 'str') + + # Construct headers + header_parameters = {} + header_parameters['Accept'] = 'application/json' + header_parameters['Content-Type'] = 'application/json; charset=utf-8' + if self.config.generate_client_request_id: + header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) + if custom_headers: + header_parameters.update(custom_headers) + if self.config.accept_language is not None: + header_parameters['accept-language'] = self._serialize.header("self.config.accept_language", self.config.accept_language, 'str') + + # Construct body + body_content = self._serialize.body(git_hub_access_token_request, 'GitHubAccessTokenRequest') + + # Construct and send request + request = self._client.post(url, query_parameters, header_parameters, body_content) + response = self._client.send(request, stream=False, **operation_config) + + if response.status_code not in [200]: + exp = CloudError(response) + exp.request_id = response.headers.get('x-ms-request-id') + raise exp + + deserialized = None + + if response.status_code == 200: + deserialized = self._deserialize('GitHubAccessTokenResponse', response) + + if raw: + client_raw_response = ClientRawResponse(deserialized, response) + return client_raw_response + + return deserialized + get_git_hub_access_token.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/getGitHubAccessToken'} From 01e8731f7867312e3d876adc3144a7ae6687dbff Mon Sep 17 00:00:00 2001 From: Azure SDK for Python bot Date: Fri, 7 Sep 2018 15:39:24 -0700 Subject: [PATCH 21/54] [AutoPR datafactory/resource-manager] Azure Data Factory: Add swagger definition for specifying Spark environment variables in AzureDatabricksLinkedService (#3286) * Generated from 6ca83791a4b40b65ea144cc8d4e50924f081b2f7 Azure Data Factory: Add swagger definition for specifying Spark environment variables in AzureDatabricksLinkedService. * Generated from ffaa86d5306aa8d8b1547aa3e55da79e26e76753 (nit: Capitalize sentence case in description of newClusterSparkConf and newClusterSparkEnvVars) * Generated from ffaa86d5306aa8d8b1547aa3e55da79e26e76753 (nit: Capitalize sentence case in description of newClusterSparkConf and newClusterSparkEnvVars) --- .../models/azure_databricks_linked_service.py | 7 ++++++- .../models/azure_databricks_linked_service_py3.py | 9 +++++++-- 2 files changed, 13 insertions(+), 3 deletions(-) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_databricks_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_databricks_linked_service.py index 270c9ff302a2..c036b299fff0 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_databricks_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_databricks_linked_service.py @@ -56,9 +56,12 @@ class AzureDatabricksLinkedService(LinkedService): :param new_cluster_node_type: The node types of new cluster. Type: string (or Expression with resultType string). :type new_cluster_node_type: object - :param new_cluster_spark_conf: a set of optional, user-specified Spark + :param new_cluster_spark_conf: A set of optional, user-specified Spark configuration key-value pairs. :type new_cluster_spark_conf: dict[str, object] + :param new_cluster_spark_env_vars: A set of optional, user-specified Spark + environment variables key-value pairs. + :type new_cluster_spark_env_vars: dict[str, object] :param new_cluster_custom_tags: Additional tags for cluster resources. :type new_cluster_custom_tags: dict[str, object] :param encrypted_credential: The encrypted credential used for @@ -87,6 +90,7 @@ class AzureDatabricksLinkedService(LinkedService): 'new_cluster_num_of_worker': {'key': 'typeProperties.newClusterNumOfWorker', 'type': 'object'}, 'new_cluster_node_type': {'key': 'typeProperties.newClusterNodeType', 'type': 'object'}, 'new_cluster_spark_conf': {'key': 'typeProperties.newClusterSparkConf', 'type': '{object}'}, + 'new_cluster_spark_env_vars': {'key': 'typeProperties.newClusterSparkEnvVars', 'type': '{object}'}, 'new_cluster_custom_tags': {'key': 'typeProperties.newClusterCustomTags', 'type': '{object}'}, 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } @@ -100,6 +104,7 @@ def __init__(self, **kwargs): self.new_cluster_num_of_worker = kwargs.get('new_cluster_num_of_worker', None) self.new_cluster_node_type = kwargs.get('new_cluster_node_type', None) self.new_cluster_spark_conf = kwargs.get('new_cluster_spark_conf', None) + self.new_cluster_spark_env_vars = kwargs.get('new_cluster_spark_env_vars', None) self.new_cluster_custom_tags = kwargs.get('new_cluster_custom_tags', None) self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'AzureDatabricks' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_databricks_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_databricks_linked_service_py3.py index aa74ccc05987..8060311a4e0d 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_databricks_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_databricks_linked_service_py3.py @@ -56,9 +56,12 @@ class AzureDatabricksLinkedService(LinkedService): :param new_cluster_node_type: The node types of new cluster. Type: string (or Expression with resultType string). :type new_cluster_node_type: object - :param new_cluster_spark_conf: a set of optional, user-specified Spark + :param new_cluster_spark_conf: A set of optional, user-specified Spark configuration key-value pairs. :type new_cluster_spark_conf: dict[str, object] + :param new_cluster_spark_env_vars: A set of optional, user-specified Spark + environment variables key-value pairs. + :type new_cluster_spark_env_vars: dict[str, object] :param new_cluster_custom_tags: Additional tags for cluster resources. :type new_cluster_custom_tags: dict[str, object] :param encrypted_credential: The encrypted credential used for @@ -87,11 +90,12 @@ class AzureDatabricksLinkedService(LinkedService): 'new_cluster_num_of_worker': {'key': 'typeProperties.newClusterNumOfWorker', 'type': 'object'}, 'new_cluster_node_type': {'key': 'typeProperties.newClusterNodeType', 'type': 'object'}, 'new_cluster_spark_conf': {'key': 'typeProperties.newClusterSparkConf', 'type': '{object}'}, + 'new_cluster_spark_env_vars': {'key': 'typeProperties.newClusterSparkEnvVars', 'type': '{object}'}, 'new_cluster_custom_tags': {'key': 'typeProperties.newClusterCustomTags', 'type': '{object}'}, 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, *, domain, access_token, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, existing_cluster_id=None, new_cluster_version=None, new_cluster_num_of_worker=None, new_cluster_node_type=None, new_cluster_spark_conf=None, new_cluster_custom_tags=None, encrypted_credential=None, **kwargs) -> None: + def __init__(self, *, domain, access_token, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, existing_cluster_id=None, new_cluster_version=None, new_cluster_num_of_worker=None, new_cluster_node_type=None, new_cluster_spark_conf=None, new_cluster_spark_env_vars=None, new_cluster_custom_tags=None, encrypted_credential=None, **kwargs) -> None: super(AzureDatabricksLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) self.domain = domain self.access_token = access_token @@ -100,6 +104,7 @@ def __init__(self, *, domain, access_token, additional_properties=None, connect_ self.new_cluster_num_of_worker = new_cluster_num_of_worker self.new_cluster_node_type = new_cluster_node_type self.new_cluster_spark_conf = new_cluster_spark_conf + self.new_cluster_spark_env_vars = new_cluster_spark_env_vars self.new_cluster_custom_tags = new_cluster_custom_tags self.encrypted_credential = encrypted_credential self.type = 'AzureDatabricks' From eb0767e213ac7823bfc7af713bb5e1ac681fd67e Mon Sep 17 00:00:00 2001 From: Azure SDK for Python bot Date: Wed, 19 Sep 2018 12:55:14 -0700 Subject: [PATCH 22/54] Generated from bfcac8c748456051e8aecca89cd0bad13704eb99 (#3361) Fix formats --- .../azure/mgmt/datafactory/models/activity_policy.py | 5 +++++ .../azure/mgmt/datafactory/models/activity_policy_py3.py | 7 ++++++- .../datafactory/operations/pipeline_runs_operations.py | 7 ++++++- 3 files changed, 17 insertions(+), 2 deletions(-) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/activity_policy.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/activity_policy.py index 76cb83407fb8..4475cdbd9bea 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/activity_policy.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/activity_policy.py @@ -28,6 +28,9 @@ class ActivityPolicy(Model): :param retry_interval_in_seconds: Interval between each retry attempt (in seconds). The default is 30 sec. :type retry_interval_in_seconds: int + :param secure_input: When set to true, Input from activity is considered + as secure and will not be logged to monitoring. + :type secure_input: bool :param secure_output: When set to true, Output from activity is considered as secure and will not be logged to monitoring. :type secure_output: bool @@ -42,6 +45,7 @@ class ActivityPolicy(Model): 'timeout': {'key': 'timeout', 'type': 'object'}, 'retry': {'key': 'retry', 'type': 'object'}, 'retry_interval_in_seconds': {'key': 'retryIntervalInSeconds', 'type': 'int'}, + 'secure_input': {'key': 'secureInput', 'type': 'bool'}, 'secure_output': {'key': 'secureOutput', 'type': 'bool'}, } @@ -51,4 +55,5 @@ def __init__(self, **kwargs): self.timeout = kwargs.get('timeout', None) self.retry = kwargs.get('retry', None) self.retry_interval_in_seconds = kwargs.get('retry_interval_in_seconds', None) + self.secure_input = kwargs.get('secure_input', None) self.secure_output = kwargs.get('secure_output', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/activity_policy_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/activity_policy_py3.py index 0a99fdb4d92d..52d469679974 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/activity_policy_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/activity_policy_py3.py @@ -28,6 +28,9 @@ class ActivityPolicy(Model): :param retry_interval_in_seconds: Interval between each retry attempt (in seconds). The default is 30 sec. :type retry_interval_in_seconds: int + :param secure_input: When set to true, Input from activity is considered + as secure and will not be logged to monitoring. + :type secure_input: bool :param secure_output: When set to true, Output from activity is considered as secure and will not be logged to monitoring. :type secure_output: bool @@ -42,13 +45,15 @@ class ActivityPolicy(Model): 'timeout': {'key': 'timeout', 'type': 'object'}, 'retry': {'key': 'retry', 'type': 'object'}, 'retry_interval_in_seconds': {'key': 'retryIntervalInSeconds', 'type': 'int'}, + 'secure_input': {'key': 'secureInput', 'type': 'bool'}, 'secure_output': {'key': 'secureOutput', 'type': 'bool'}, } - def __init__(self, *, additional_properties=None, timeout=None, retry=None, retry_interval_in_seconds: int=None, secure_output: bool=None, **kwargs) -> None: + def __init__(self, *, additional_properties=None, timeout=None, retry=None, retry_interval_in_seconds: int=None, secure_input: bool=None, secure_output: bool=None, **kwargs) -> None: super(ActivityPolicy, self).__init__(**kwargs) self.additional_properties = additional_properties self.timeout = timeout self.retry = retry self.retry_interval_in_seconds = retry_interval_in_seconds + self.secure_input = secure_input self.secure_output = secure_output diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/pipeline_runs_operations.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/pipeline_runs_operations.py index 28799443c0d3..de8744612d20 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/pipeline_runs_operations.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/pipeline_runs_operations.py @@ -172,7 +172,7 @@ def get( get.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/pipelineruns/{runId}'} def cancel( - self, resource_group_name, factory_name, run_id, custom_headers=None, raw=False, **operation_config): + self, resource_group_name, factory_name, run_id, is_recursive=None, custom_headers=None, raw=False, **operation_config): """Cancel a pipeline run by its run ID. :param resource_group_name: The resource group name. @@ -181,6 +181,9 @@ def cancel( :type factory_name: str :param run_id: The pipeline run identifier. :type run_id: str + :param is_recursive: If true, cancel all the Child pipelines that are + triggered by the current pipeline. + :type is_recursive: bool :param dict custom_headers: headers that will be added to the request :param bool raw: returns the direct response alongside the deserialized response @@ -202,6 +205,8 @@ def cancel( # Construct parameters query_parameters = {} + if is_recursive is not None: + query_parameters['isRecursive'] = self._serialize.query("is_recursive", is_recursive, 'bool') query_parameters['api-version'] = self._serialize.query("self.api_version", self.api_version, 'str') # Construct headers From fa39544b90cd30d79ef597a4f16db467867c1a59 Mon Sep 17 00:00:00 2001 From: Azure SDK for Python bot Date: Thu, 27 Sep 2018 23:41:46 -0700 Subject: [PATCH 23/54] [AutoPR datafactory/resource-manager] [Datafactory] Adding script actions support for HDI on demand linked service (#3451) * Generated from 9dec6b08742c433d0429a8737a65fdc690863d6b Adding parameters * Packaging update of azure-mgmt-datafactory --- azure-mgmt-datafactory/MANIFEST.in | 1 - azure-mgmt-datafactory/azure/__init__.py | 2 +- azure-mgmt-datafactory/azure/mgmt/__init__.py | 2 +- .../azure/mgmt/datafactory/models/__init__.py | 3 ++ .../hd_insight_on_demand_linked_service.py | 6 +++ ...hd_insight_on_demand_linked_service_py3.py | 8 ++- .../mgmt/datafactory/models/script_action.py | 49 +++++++++++++++++ .../datafactory/models/script_action_py3.py | 49 +++++++++++++++++ azure-mgmt-datafactory/azure_bdist_wheel.py | 54 ------------------- azure-mgmt-datafactory/setup.cfg | 1 - azure-mgmt-datafactory/setup.py | 17 +++--- 11 files changed, 125 insertions(+), 67 deletions(-) create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/script_action.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/script_action_py3.py delete mode 100644 azure-mgmt-datafactory/azure_bdist_wheel.py diff --git a/azure-mgmt-datafactory/MANIFEST.in b/azure-mgmt-datafactory/MANIFEST.in index 9ecaeb15de50..bb37a2723dae 100644 --- a/azure-mgmt-datafactory/MANIFEST.in +++ b/azure-mgmt-datafactory/MANIFEST.in @@ -1,2 +1 @@ include *.rst -include azure_bdist_wheel.py \ No newline at end of file diff --git a/azure-mgmt-datafactory/azure/__init__.py b/azure-mgmt-datafactory/azure/__init__.py index de40ea7ca058..0260537a02bb 100644 --- a/azure-mgmt-datafactory/azure/__init__.py +++ b/azure-mgmt-datafactory/azure/__init__.py @@ -1 +1 @@ -__import__('pkg_resources').declare_namespace(__name__) +__path__ = __import__('pkgutil').extend_path(__path__, __name__) \ No newline at end of file diff --git a/azure-mgmt-datafactory/azure/mgmt/__init__.py b/azure-mgmt-datafactory/azure/mgmt/__init__.py index de40ea7ca058..0260537a02bb 100644 --- a/azure-mgmt-datafactory/azure/mgmt/__init__.py +++ b/azure-mgmt-datafactory/azure/mgmt/__init__.py @@ -1 +1 @@ -__import__('pkg_resources').declare_namespace(__name__) +__path__ = __import__('pkgutil').extend_path(__path__, __name__) \ No newline at end of file diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py index 436179448b73..e958df5cca8c 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py @@ -73,6 +73,7 @@ from .responsys_linked_service_py3 import ResponsysLinkedService from .azure_databricks_linked_service_py3 import AzureDatabricksLinkedService from .azure_data_lake_analytics_linked_service_py3 import AzureDataLakeAnalyticsLinkedService + from .script_action_py3 import ScriptAction from .hd_insight_on_demand_linked_service_py3 import HDInsightOnDemandLinkedService from .salesforce_marketing_cloud_linked_service_py3 import SalesforceMarketingCloudLinkedService from .netezza_linked_service_py3 import NetezzaLinkedService @@ -418,6 +419,7 @@ from .responsys_linked_service import ResponsysLinkedService from .azure_databricks_linked_service import AzureDatabricksLinkedService from .azure_data_lake_analytics_linked_service import AzureDataLakeAnalyticsLinkedService + from .script_action import ScriptAction from .hd_insight_on_demand_linked_service import HDInsightOnDemandLinkedService from .salesforce_marketing_cloud_linked_service import SalesforceMarketingCloudLinkedService from .netezza_linked_service import NetezzaLinkedService @@ -831,6 +833,7 @@ 'ResponsysLinkedService', 'AzureDatabricksLinkedService', 'AzureDataLakeAnalyticsLinkedService', + 'ScriptAction', 'HDInsightOnDemandLinkedService', 'SalesforceMarketingCloudLinkedService', 'NetezzaLinkedService', diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_on_demand_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_on_demand_linked_service.py index 8069bd2d2d26..bd84aabc5012 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_on_demand_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_on_demand_linked_service.py @@ -132,6 +132,10 @@ class HDInsightOnDemandLinkedService(LinkedService): :param zookeeper_node_size: Specifies the size of the Zoo Keeper node for the HDInsight cluster. :type zookeeper_node_size: object + :param script_actions: Custom script actions to run on HDI ondemand + cluster once it's up. Please refer to + https://docs.microsoft.com/en-us/azure/hdinsight/hdinsight-hadoop-customize-cluster-linux?toc=%2Fen-us%2Fazure%2Fhdinsight%2Fr-server%2FTOC.json&bc=%2Fen-us%2Fazure%2Fbread%2Ftoc.json#understanding-script-actions. + :type script_actions: list[~azure.mgmt.datafactory.models.ScriptAction] """ _validation = { @@ -182,6 +186,7 @@ class HDInsightOnDemandLinkedService(LinkedService): 'head_node_size': {'key': 'typeProperties.headNodeSize', 'type': 'object'}, 'data_node_size': {'key': 'typeProperties.dataNodeSize', 'type': 'object'}, 'zookeeper_node_size': {'key': 'typeProperties.zookeeperNodeSize', 'type': 'object'}, + 'script_actions': {'key': 'typeProperties.scriptActions', 'type': '[ScriptAction]'}, } def __init__(self, **kwargs): @@ -216,4 +221,5 @@ def __init__(self, **kwargs): self.head_node_size = kwargs.get('head_node_size', None) self.data_node_size = kwargs.get('data_node_size', None) self.zookeeper_node_size = kwargs.get('zookeeper_node_size', None) + self.script_actions = kwargs.get('script_actions', None) self.type = 'HDInsightOnDemand' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_on_demand_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_on_demand_linked_service_py3.py index 75f601f6a97b..5566a022bda2 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_on_demand_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_on_demand_linked_service_py3.py @@ -132,6 +132,10 @@ class HDInsightOnDemandLinkedService(LinkedService): :param zookeeper_node_size: Specifies the size of the Zoo Keeper node for the HDInsight cluster. :type zookeeper_node_size: object + :param script_actions: Custom script actions to run on HDI ondemand + cluster once it's up. Please refer to + https://docs.microsoft.com/en-us/azure/hdinsight/hdinsight-hadoop-customize-cluster-linux?toc=%2Fen-us%2Fazure%2Fhdinsight%2Fr-server%2FTOC.json&bc=%2Fen-us%2Fazure%2Fbread%2Ftoc.json#understanding-script-actions. + :type script_actions: list[~azure.mgmt.datafactory.models.ScriptAction] """ _validation = { @@ -182,9 +186,10 @@ class HDInsightOnDemandLinkedService(LinkedService): 'head_node_size': {'key': 'typeProperties.headNodeSize', 'type': 'object'}, 'data_node_size': {'key': 'typeProperties.dataNodeSize', 'type': 'object'}, 'zookeeper_node_size': {'key': 'typeProperties.zookeeperNodeSize', 'type': 'object'}, + 'script_actions': {'key': 'typeProperties.scriptActions', 'type': '[ScriptAction]'}, } - def __init__(self, *, cluster_size, time_to_live, version, linked_service_name, host_subscription_id, tenant, cluster_resource_group, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, service_principal_id=None, service_principal_key=None, cluster_name_prefix=None, cluster_user_name=None, cluster_password=None, cluster_ssh_user_name=None, cluster_ssh_password=None, additional_linked_service_names=None, hcatalog_linked_service_name=None, cluster_type=None, spark_version=None, core_configuration=None, h_base_configuration=None, hdfs_configuration=None, hive_configuration=None, map_reduce_configuration=None, oozie_configuration=None, storm_configuration=None, yarn_configuration=None, encrypted_credential=None, head_node_size=None, data_node_size=None, zookeeper_node_size=None, **kwargs) -> None: + def __init__(self, *, cluster_size, time_to_live, version, linked_service_name, host_subscription_id, tenant, cluster_resource_group, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, service_principal_id=None, service_principal_key=None, cluster_name_prefix=None, cluster_user_name=None, cluster_password=None, cluster_ssh_user_name=None, cluster_ssh_password=None, additional_linked_service_names=None, hcatalog_linked_service_name=None, cluster_type=None, spark_version=None, core_configuration=None, h_base_configuration=None, hdfs_configuration=None, hive_configuration=None, map_reduce_configuration=None, oozie_configuration=None, storm_configuration=None, yarn_configuration=None, encrypted_credential=None, head_node_size=None, data_node_size=None, zookeeper_node_size=None, script_actions=None, **kwargs) -> None: super(HDInsightOnDemandLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) self.cluster_size = cluster_size self.time_to_live = time_to_live @@ -216,4 +221,5 @@ def __init__(self, *, cluster_size, time_to_live, version, linked_service_name, self.head_node_size = head_node_size self.data_node_size = data_node_size self.zookeeper_node_size = zookeeper_node_size + self.script_actions = script_actions self.type = 'HDInsightOnDemand' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/script_action.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/script_action.py new file mode 100644 index 000000000000..50bc0131a5cf --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/script_action.py @@ -0,0 +1,49 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class ScriptAction(Model): + """Custom script action to run on HDI ondemand cluster once it's up. + + All required parameters must be populated in order to send to Azure. + + :param name: Required. The user provided name of the script action. + :type name: str + :param uri: Required. The URI for the script action. + :type uri: str + :param roles: Required. The node types on which the script action should + be executed. + :type roles: object + :param parameters: The parameters for the script action. + :type parameters: str + """ + + _validation = { + 'name': {'required': True}, + 'uri': {'required': True}, + 'roles': {'required': True}, + } + + _attribute_map = { + 'name': {'key': 'name', 'type': 'str'}, + 'uri': {'key': 'uri', 'type': 'str'}, + 'roles': {'key': 'roles', 'type': 'object'}, + 'parameters': {'key': 'parameters', 'type': 'str'}, + } + + def __init__(self, **kwargs): + super(ScriptAction, self).__init__(**kwargs) + self.name = kwargs.get('name', None) + self.uri = kwargs.get('uri', None) + self.roles = kwargs.get('roles', None) + self.parameters = kwargs.get('parameters', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/script_action_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/script_action_py3.py new file mode 100644 index 000000000000..c0e278073219 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/script_action_py3.py @@ -0,0 +1,49 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class ScriptAction(Model): + """Custom script action to run on HDI ondemand cluster once it's up. + + All required parameters must be populated in order to send to Azure. + + :param name: Required. The user provided name of the script action. + :type name: str + :param uri: Required. The URI for the script action. + :type uri: str + :param roles: Required. The node types on which the script action should + be executed. + :type roles: object + :param parameters: The parameters for the script action. + :type parameters: str + """ + + _validation = { + 'name': {'required': True}, + 'uri': {'required': True}, + 'roles': {'required': True}, + } + + _attribute_map = { + 'name': {'key': 'name', 'type': 'str'}, + 'uri': {'key': 'uri', 'type': 'str'}, + 'roles': {'key': 'roles', 'type': 'object'}, + 'parameters': {'key': 'parameters', 'type': 'str'}, + } + + def __init__(self, *, name: str, uri: str, roles, parameters: str=None, **kwargs) -> None: + super(ScriptAction, self).__init__(**kwargs) + self.name = name + self.uri = uri + self.roles = roles + self.parameters = parameters diff --git a/azure-mgmt-datafactory/azure_bdist_wheel.py b/azure-mgmt-datafactory/azure_bdist_wheel.py deleted file mode 100644 index 8a81d1b61775..000000000000 --- a/azure-mgmt-datafactory/azure_bdist_wheel.py +++ /dev/null @@ -1,54 +0,0 @@ -#------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for -# license information. -#-------------------------------------------------------------------------- - -from distutils import log as logger -import os.path - -from wheel.bdist_wheel import bdist_wheel -class azure_bdist_wheel(bdist_wheel): - """The purpose of this class is to build wheel a little differently than the sdist, - without requiring to build the wheel from the sdist (i.e. you can build the wheel - directly from source). - """ - - description = "Create an Azure wheel distribution" - - user_options = bdist_wheel.user_options + \ - [('azure-namespace-package=', None, - "Name of the deepest nspkg used")] - - def initialize_options(self): - bdist_wheel.initialize_options(self) - self.azure_namespace_package = None - - def finalize_options(self): - bdist_wheel.finalize_options(self) - if self.azure_namespace_package and not self.azure_namespace_package.endswith("-nspkg"): - raise ValueError("azure_namespace_package must finish by -nspkg") - - def run(self): - if not self.distribution.install_requires: - self.distribution.install_requires = [] - self.distribution.install_requires.append( - "{}>=2.0.0".format(self.azure_namespace_package)) - bdist_wheel.run(self) - - def write_record(self, bdist_dir, distinfo_dir): - if self.azure_namespace_package: - # Split and remove last part, assuming it's "nspkg" - subparts = self.azure_namespace_package.split('-')[0:-1] - folder_with_init = [os.path.join(*subparts[0:i+1]) for i in range(len(subparts))] - for azure_sub_package in folder_with_init: - init_file = os.path.join(bdist_dir, azure_sub_package, '__init__.py') - if os.path.isfile(init_file): - logger.info("manually remove {} while building the wheel".format(init_file)) - os.remove(init_file) - else: - raise ValueError("Unable to find {}. Are you sure of your namespace package?".format(init_file)) - bdist_wheel.write_record(self, bdist_dir, distinfo_dir) -cmdclass = { - 'bdist_wheel': azure_bdist_wheel, -} diff --git a/azure-mgmt-datafactory/setup.cfg b/azure-mgmt-datafactory/setup.cfg index 856f4164982c..3c6e79cf31da 100644 --- a/azure-mgmt-datafactory/setup.cfg +++ b/azure-mgmt-datafactory/setup.cfg @@ -1,3 +1,2 @@ [bdist_wheel] universal=1 -azure-namespace-package=azure-mgmt-nspkg \ No newline at end of file diff --git a/azure-mgmt-datafactory/setup.py b/azure-mgmt-datafactory/setup.py index 98b696f39617..8fb4b233c1b8 100644 --- a/azure-mgmt-datafactory/setup.py +++ b/azure-mgmt-datafactory/setup.py @@ -10,12 +10,6 @@ import os.path from io import open from setuptools import find_packages, setup -try: - from azure_bdist_wheel import cmdclass -except ImportError: - from distutils import log as logger - logger.warn("Wheel is not available, disabling bdist_wheel hook") - cmdclass = {} # Change the PACKAGE_NAME only to change folder and different name PACKAGE_NAME = "azure-mgmt-datafactory" @@ -76,11 +70,18 @@ 'License :: OSI Approved :: MIT License', ], zip_safe=False, - packages=find_packages(exclude=["tests"]), + packages=find_packages(exclude=[ + 'tests', + # Exclude packages that will be covered by PEP420 or nspkg + 'azure', + 'azure.mgmt', + ]), install_requires=[ 'msrest>=0.5.0', 'msrestazure>=0.4.32,<2.0.0', 'azure-common~=1.1', ], - cmdclass=cmdclass + extras_require={ + ":python_version<'3.0'": ['azure-mgmt-nspkg'], + } ) From 97119a08bd5a6c5bf202b6de92ec5ef88b0e9174 Mon Sep 17 00:00:00 2001 From: Azure SDK for Python bot Date: Mon, 1 Oct 2018 08:40:23 -0700 Subject: [PATCH 24/54] Generated from 68c58c2d8602adda9e80068e105f3d449474aa51 (#3465) Fix grammar and typo --- .../azure/mgmt/datafactory/models/__init__.py | 11 ++++ .../models/append_variable_activity.py | 60 +++++++++++++++++++ .../models/append_variable_activity_py3.py | 60 +++++++++++++++++++ .../datafactory/models/control_activity.py | 7 ++- .../models/control_activity_py3.py | 7 ++- .../data_factory_management_client_enums.py | 7 +++ .../datafactory/models/pipeline_resource.py | 5 ++ .../models/pipeline_resource_py3.py | 7 ++- .../models/set_variable_activity.py | 59 ++++++++++++++++++ .../models/set_variable_activity_py3.py | 59 ++++++++++++++++++ .../models/variable_specification.py | 39 ++++++++++++ .../models/variable_specification_py3.py | 39 ++++++++++++ 12 files changed, 353 insertions(+), 7 deletions(-) create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/append_variable_activity.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/append_variable_activity_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/set_variable_activity.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/set_variable_activity_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/variable_specification.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/variable_specification_py3.py diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py index e958df5cca8c..e758e9dae31f 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py @@ -39,6 +39,7 @@ from .activity_dependency_py3 import ActivityDependency from .user_property_py3 import UserProperty from .activity_py3 import Activity + from .variable_specification_py3 import VariableSpecification from .pipeline_folder_py3 import PipelineFolder from .pipeline_resource_py3 import PipelineResource from .trigger_py3 import Trigger @@ -325,6 +326,8 @@ from .copy_sink_py3 import CopySink from .copy_activity_py3 import CopyActivity from .execution_activity_py3 import ExecutionActivity + from .append_variable_activity_py3 import AppendVariableActivity + from .set_variable_activity_py3 import SetVariableActivity from .filter_activity_py3 import FilterActivity from .until_activity_py3 import UntilActivity from .wait_activity_py3 import WaitActivity @@ -385,6 +388,7 @@ from .activity_dependency import ActivityDependency from .user_property import UserProperty from .activity import Activity + from .variable_specification import VariableSpecification from .pipeline_folder import PipelineFolder from .pipeline_resource import PipelineResource from .trigger import Trigger @@ -671,6 +675,8 @@ from .copy_sink import CopySink from .copy_activity import CopyActivity from .execution_activity import ExecutionActivity + from .append_variable_activity import AppendVariableActivity + from .set_variable_activity import SetVariableActivity from .filter_activity import FilterActivity from .until_activity import UntilActivity from .wait_activity import WaitActivity @@ -713,6 +719,7 @@ IntegrationRuntimeAutoUpdate, ParameterType, DependencyCondition, + VariableType, TriggerRuntimeState, RunQueryFilterOperand, RunQueryFilterOperator, @@ -799,6 +806,7 @@ 'ActivityDependency', 'UserProperty', 'Activity', + 'VariableSpecification', 'PipelineFolder', 'PipelineResource', 'Trigger', @@ -1085,6 +1093,8 @@ 'CopySink', 'CopyActivity', 'ExecutionActivity', + 'AppendVariableActivity', + 'SetVariableActivity', 'FilterActivity', 'UntilActivity', 'WaitActivity', @@ -1126,6 +1136,7 @@ 'IntegrationRuntimeAutoUpdate', 'ParameterType', 'DependencyCondition', + 'VariableType', 'TriggerRuntimeState', 'RunQueryFilterOperand', 'RunQueryFilterOperator', diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/append_variable_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/append_variable_activity.py new file mode 100644 index 000000000000..36a25e959061 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/append_variable_activity.py @@ -0,0 +1,60 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .control_activity import ControlActivity + + +class AppendVariableActivity(ControlActivity): + """Append value for a Variable of type Array. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param name: Required. Activity name. + :type name: str + :param description: Activity description. + :type description: str + :param depends_on: Activity depends on condition. + :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param user_properties: Activity user properties. + :type user_properties: list[~azure.mgmt.datafactory.models.UserProperty] + :param type: Required. Constant filled by server. + :type type: str + :param variable_name: Name of the variable whose value needs to be + appended to. + :type variable_name: str + :param value: Value to be appended. Could be a static value or Expression + :type value: object + """ + + _validation = { + 'name': {'required': True}, + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'name': {'key': 'name', 'type': 'str'}, + 'description': {'key': 'description', 'type': 'str'}, + 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'variable_name': {'key': 'typeProperties.variableName', 'type': 'str'}, + 'value': {'key': 'typeProperties.value', 'type': 'object'}, + } + + def __init__(self, **kwargs): + super(AppendVariableActivity, self).__init__(**kwargs) + self.variable_name = kwargs.get('variable_name', None) + self.value = kwargs.get('value', None) + self.type = 'AppendVariable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/append_variable_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/append_variable_activity_py3.py new file mode 100644 index 000000000000..4526a6e4a45e --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/append_variable_activity_py3.py @@ -0,0 +1,60 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .control_activity_py3 import ControlActivity + + +class AppendVariableActivity(ControlActivity): + """Append value for a Variable of type Array. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param name: Required. Activity name. + :type name: str + :param description: Activity description. + :type description: str + :param depends_on: Activity depends on condition. + :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param user_properties: Activity user properties. + :type user_properties: list[~azure.mgmt.datafactory.models.UserProperty] + :param type: Required. Constant filled by server. + :type type: str + :param variable_name: Name of the variable whose value needs to be + appended to. + :type variable_name: str + :param value: Value to be appended. Could be a static value or Expression + :type value: object + """ + + _validation = { + 'name': {'required': True}, + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'name': {'key': 'name', 'type': 'str'}, + 'description': {'key': 'description', 'type': 'str'}, + 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'variable_name': {'key': 'typeProperties.variableName', 'type': 'str'}, + 'value': {'key': 'typeProperties.value', 'type': 'object'}, + } + + def __init__(self, *, name: str, additional_properties=None, description: str=None, depends_on=None, user_properties=None, variable_name: str=None, value=None, **kwargs) -> None: + super(AppendVariableActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, **kwargs) + self.variable_name = variable_name + self.value = value + self.type = 'AppendVariable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/control_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/control_activity.py index 87677dcfa6bf..16581581786b 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/control_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/control_activity.py @@ -16,8 +16,9 @@ class ControlActivity(Activity): """Base class for all control activities like IfCondition, ForEach , Until. You probably want to use the sub-classes and not this class directly. Known - sub-classes are: FilterActivity, UntilActivity, WaitActivity, - ForEachActivity, IfConditionActivity, ExecutePipelineActivity + sub-classes are: AppendVariableActivity, SetVariableActivity, + FilterActivity, UntilActivity, WaitActivity, ForEachActivity, + IfConditionActivity, ExecutePipelineActivity All required parameters must be populated in order to send to Azure. @@ -51,7 +52,7 @@ class ControlActivity(Activity): } _subtype_map = { - 'type': {'Filter': 'FilterActivity', 'Until': 'UntilActivity', 'Wait': 'WaitActivity', 'ForEach': 'ForEachActivity', 'IfCondition': 'IfConditionActivity', 'ExecutePipeline': 'ExecutePipelineActivity'} + 'type': {'AppendVariable': 'AppendVariableActivity', 'SetVariable': 'SetVariableActivity', 'Filter': 'FilterActivity', 'Until': 'UntilActivity', 'Wait': 'WaitActivity', 'ForEach': 'ForEachActivity', 'IfCondition': 'IfConditionActivity', 'ExecutePipeline': 'ExecutePipelineActivity'} } def __init__(self, **kwargs): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/control_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/control_activity_py3.py index 09004ce2cee2..739d8b9c311b 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/control_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/control_activity_py3.py @@ -16,8 +16,9 @@ class ControlActivity(Activity): """Base class for all control activities like IfCondition, ForEach , Until. You probably want to use the sub-classes and not this class directly. Known - sub-classes are: FilterActivity, UntilActivity, WaitActivity, - ForEachActivity, IfConditionActivity, ExecutePipelineActivity + sub-classes are: AppendVariableActivity, SetVariableActivity, + FilterActivity, UntilActivity, WaitActivity, ForEachActivity, + IfConditionActivity, ExecutePipelineActivity All required parameters must be populated in order to send to Azure. @@ -51,7 +52,7 @@ class ControlActivity(Activity): } _subtype_map = { - 'type': {'Filter': 'FilterActivity', 'Until': 'UntilActivity', 'Wait': 'WaitActivity', 'ForEach': 'ForEachActivity', 'IfCondition': 'IfConditionActivity', 'ExecutePipeline': 'ExecutePipelineActivity'} + 'type': {'AppendVariable': 'AppendVariableActivity', 'SetVariable': 'SetVariableActivity', 'Filter': 'FilterActivity', 'Until': 'UntilActivity', 'Wait': 'WaitActivity', 'ForEach': 'ForEachActivity', 'IfCondition': 'IfConditionActivity', 'ExecutePipeline': 'ExecutePipelineActivity'} } def __init__(self, *, name: str, additional_properties=None, description: str=None, depends_on=None, user_properties=None, **kwargs) -> None: diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_factory_management_client_enums.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_factory_management_client_enums.py index 086348af981d..2de647ff9d8b 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_factory_management_client_enums.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_factory_management_client_enums.py @@ -51,6 +51,13 @@ class DependencyCondition(str, Enum): completed = "Completed" +class VariableType(str, Enum): + + string = "String" + bool_enum = "Bool" + array = "Array" + + class TriggerRuntimeState(str, Enum): started = "Started" diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_resource.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_resource.py index 00d1ec8cf81a..a39deaccc87b 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_resource.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_resource.py @@ -36,6 +36,9 @@ class PipelineResource(SubResource): :param parameters: List of parameters for pipeline. :type parameters: dict[str, ~azure.mgmt.datafactory.models.ParameterSpecification] + :param variables: List of variables for pipeline. + :type variables: dict[str, + ~azure.mgmt.datafactory.models.VariableSpecification] :param concurrency: The max number of concurrent runs for the pipeline. :type concurrency: int :param annotations: List of tags that can be used for describing the @@ -63,6 +66,7 @@ class PipelineResource(SubResource): 'description': {'key': 'properties.description', 'type': 'str'}, 'activities': {'key': 'properties.activities', 'type': '[Activity]'}, 'parameters': {'key': 'properties.parameters', 'type': '{ParameterSpecification}'}, + 'variables': {'key': 'properties.variables', 'type': '{VariableSpecification}'}, 'concurrency': {'key': 'properties.concurrency', 'type': 'int'}, 'annotations': {'key': 'properties.annotations', 'type': '[object]'}, 'folder': {'key': 'properties.folder', 'type': 'PipelineFolder'}, @@ -74,6 +78,7 @@ def __init__(self, **kwargs): self.description = kwargs.get('description', None) self.activities = kwargs.get('activities', None) self.parameters = kwargs.get('parameters', None) + self.variables = kwargs.get('variables', None) self.concurrency = kwargs.get('concurrency', None) self.annotations = kwargs.get('annotations', None) self.folder = kwargs.get('folder', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_resource_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_resource_py3.py index 44eb05b05522..8299cdb73887 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_resource_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/pipeline_resource_py3.py @@ -36,6 +36,9 @@ class PipelineResource(SubResource): :param parameters: List of parameters for pipeline. :type parameters: dict[str, ~azure.mgmt.datafactory.models.ParameterSpecification] + :param variables: List of variables for pipeline. + :type variables: dict[str, + ~azure.mgmt.datafactory.models.VariableSpecification] :param concurrency: The max number of concurrent runs for the pipeline. :type concurrency: int :param annotations: List of tags that can be used for describing the @@ -63,17 +66,19 @@ class PipelineResource(SubResource): 'description': {'key': 'properties.description', 'type': 'str'}, 'activities': {'key': 'properties.activities', 'type': '[Activity]'}, 'parameters': {'key': 'properties.parameters', 'type': '{ParameterSpecification}'}, + 'variables': {'key': 'properties.variables', 'type': '{VariableSpecification}'}, 'concurrency': {'key': 'properties.concurrency', 'type': 'int'}, 'annotations': {'key': 'properties.annotations', 'type': '[object]'}, 'folder': {'key': 'properties.folder', 'type': 'PipelineFolder'}, } - def __init__(self, *, additional_properties=None, description: str=None, activities=None, parameters=None, concurrency: int=None, annotations=None, folder=None, **kwargs) -> None: + def __init__(self, *, additional_properties=None, description: str=None, activities=None, parameters=None, variables=None, concurrency: int=None, annotations=None, folder=None, **kwargs) -> None: super(PipelineResource, self).__init__(**kwargs) self.additional_properties = additional_properties self.description = description self.activities = activities self.parameters = parameters + self.variables = variables self.concurrency = concurrency self.annotations = annotations self.folder = folder diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/set_variable_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/set_variable_activity.py new file mode 100644 index 000000000000..e8dd1690862d --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/set_variable_activity.py @@ -0,0 +1,59 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .control_activity import ControlActivity + + +class SetVariableActivity(ControlActivity): + """Set value for a Variable. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param name: Required. Activity name. + :type name: str + :param description: Activity description. + :type description: str + :param depends_on: Activity depends on condition. + :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param user_properties: Activity user properties. + :type user_properties: list[~azure.mgmt.datafactory.models.UserProperty] + :param type: Required. Constant filled by server. + :type type: str + :param variable_name: Name of the variable whose value needs to be set. + :type variable_name: str + :param value: Value to be set. Could be a static value or Expression + :type value: object + """ + + _validation = { + 'name': {'required': True}, + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'name': {'key': 'name', 'type': 'str'}, + 'description': {'key': 'description', 'type': 'str'}, + 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'variable_name': {'key': 'typeProperties.variableName', 'type': 'str'}, + 'value': {'key': 'typeProperties.value', 'type': 'object'}, + } + + def __init__(self, **kwargs): + super(SetVariableActivity, self).__init__(**kwargs) + self.variable_name = kwargs.get('variable_name', None) + self.value = kwargs.get('value', None) + self.type = 'SetVariable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/set_variable_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/set_variable_activity_py3.py new file mode 100644 index 000000000000..e045abee3dfb --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/set_variable_activity_py3.py @@ -0,0 +1,59 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .control_activity_py3 import ControlActivity + + +class SetVariableActivity(ControlActivity): + """Set value for a Variable. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param name: Required. Activity name. + :type name: str + :param description: Activity description. + :type description: str + :param depends_on: Activity depends on condition. + :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param user_properties: Activity user properties. + :type user_properties: list[~azure.mgmt.datafactory.models.UserProperty] + :param type: Required. Constant filled by server. + :type type: str + :param variable_name: Name of the variable whose value needs to be set. + :type variable_name: str + :param value: Value to be set. Could be a static value or Expression + :type value: object + """ + + _validation = { + 'name': {'required': True}, + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'name': {'key': 'name', 'type': 'str'}, + 'description': {'key': 'description', 'type': 'str'}, + 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'variable_name': {'key': 'typeProperties.variableName', 'type': 'str'}, + 'value': {'key': 'typeProperties.value', 'type': 'object'}, + } + + def __init__(self, *, name: str, additional_properties=None, description: str=None, depends_on=None, user_properties=None, variable_name: str=None, value=None, **kwargs) -> None: + super(SetVariableActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, **kwargs) + self.variable_name = variable_name + self.value = value + self.type = 'SetVariable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/variable_specification.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/variable_specification.py new file mode 100644 index 000000000000..6d7fd808fa44 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/variable_specification.py @@ -0,0 +1,39 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class VariableSpecification(Model): + """Definition of a single variable for a Pipeline. + + All required parameters must be populated in order to send to Azure. + + :param type: Required. Variable type. Possible values include: 'String', + 'Bool', 'Array' + :type type: str or ~azure.mgmt.datafactory.models.VariableType + :param default_value: Default value of variable. + :type default_value: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'type': {'key': 'type', 'type': 'str'}, + 'default_value': {'key': 'defaultValue', 'type': 'object'}, + } + + def __init__(self, **kwargs): + super(VariableSpecification, self).__init__(**kwargs) + self.type = kwargs.get('type', None) + self.default_value = kwargs.get('default_value', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/variable_specification_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/variable_specification_py3.py new file mode 100644 index 000000000000..d60b3b4b1591 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/variable_specification_py3.py @@ -0,0 +1,39 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class VariableSpecification(Model): + """Definition of a single variable for a Pipeline. + + All required parameters must be populated in order to send to Azure. + + :param type: Required. Variable type. Possible values include: 'String', + 'Bool', 'Array' + :type type: str or ~azure.mgmt.datafactory.models.VariableType + :param default_value: Default value of variable. + :type default_value: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'type': {'key': 'type', 'type': 'str'}, + 'default_value': {'key': 'defaultValue', 'type': 'object'}, + } + + def __init__(self, *, type, default_value=None, **kwargs) -> None: + super(VariableSpecification, self).__init__(**kwargs) + self.type = type + self.default_value = default_value From 6fea123f7f4571a239fdaecedafeb88fc644058a Mon Sep 17 00:00:00 2001 From: Azure SDK for Python bot Date: Fri, 5 Oct 2018 11:35:35 -0700 Subject: [PATCH 25/54] [AutoPR datafactory/resource-manager] Add APIs for tumbling window rerun trigger (#3487) * Generated from 307ab1cc99d392c3ec5e58713c4f60185f8e3808 fix parameter name * Generated from 6648a0fbe9895dae878aaf0d47b8c5de5696ebf0 1. fix ? in description 2. mark nextLink readonly --- .../data_factory_management_client.py | 5 + .../azure/mgmt/datafactory/models/__init__.py | 115 +++-- .../data_factory_management_client_enums.py | 90 ++-- .../models/rerun_trigger_resource.py | 54 +++ .../models/rerun_trigger_resource_paged.py | 27 ++ .../models/rerun_trigger_resource_py3.py | 54 +++ .../models/rerun_tumbling_window_trigger.py | 74 +++ ...mbling_window_trigger_action_parameters.py | 47 ++ ...ng_window_trigger_action_parameters_py3.py | 47 ++ .../rerun_tumbling_window_trigger_py3.py | 74 +++ .../azure/mgmt/datafactory/models/trigger.py | 5 +- .../mgmt/datafactory/models/trigger_py3.py | 5 +- .../mgmt/datafactory/operations/__init__.py | 2 + .../operations/rerun_triggers_operations.py | 450 ++++++++++++++++++ 14 files changed, 948 insertions(+), 101 deletions(-) create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/rerun_trigger_resource.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/rerun_trigger_resource_paged.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/rerun_trigger_resource_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/rerun_tumbling_window_trigger.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/rerun_tumbling_window_trigger_action_parameters.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/rerun_tumbling_window_trigger_action_parameters_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/rerun_tumbling_window_trigger_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/operations/rerun_triggers_operations.py diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/data_factory_management_client.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/data_factory_management_client.py index ab57aa1c0bfa..864f9f3d0d34 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/data_factory_management_client.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/data_factory_management_client.py @@ -23,6 +23,7 @@ from .operations.pipeline_runs_operations import PipelineRunsOperations from .operations.activity_runs_operations import ActivityRunsOperations from .operations.triggers_operations import TriggersOperations +from .operations.rerun_triggers_operations import RerunTriggersOperations from .operations.trigger_runs_operations import TriggerRunsOperations from . import models @@ -85,6 +86,8 @@ class DataFactoryManagementClient(SDKClient): :vartype activity_runs: azure.mgmt.datafactory.operations.ActivityRunsOperations :ivar triggers: Triggers operations :vartype triggers: azure.mgmt.datafactory.operations.TriggersOperations + :ivar rerun_triggers: RerunTriggers operations + :vartype rerun_triggers: azure.mgmt.datafactory.operations.RerunTriggersOperations :ivar trigger_runs: TriggerRuns operations :vartype trigger_runs: azure.mgmt.datafactory.operations.TriggerRunsOperations @@ -127,5 +130,7 @@ def __init__( self._client, self.config, self._serialize, self._deserialize) self.triggers = TriggersOperations( self._client, self.config, self._serialize, self._deserialize) + self.rerun_triggers = RerunTriggersOperations( + self._client, self.config, self._serialize, self._deserialize) self.trigger_runs = TriggerRunsOperations( self._client, self.config, self._serialize, self._deserialize) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py index e758e9dae31f..cab29db469a7 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py @@ -64,6 +64,9 @@ from .activity_runs_query_response_py3 import ActivityRunsQueryResponse from .trigger_run_py3 import TriggerRun from .trigger_runs_query_response_py3 import TriggerRunsQueryResponse + from .rerun_tumbling_window_trigger_action_parameters_py3 import RerunTumblingWindowTriggerActionParameters + from .rerun_tumbling_window_trigger_py3 import RerunTumblingWindowTrigger + from .rerun_trigger_resource_py3 import RerunTriggerResource from .operation_display_py3 import OperationDisplay from .operation_log_specification_py3 import OperationLogSpecification from .operation_metric_availability_py3 import OperationMetricAvailability @@ -71,6 +74,20 @@ from .operation_metric_specification_py3 import OperationMetricSpecification from .operation_service_specification_py3 import OperationServiceSpecification from .operation_py3 import Operation + from .self_dependency_tumbling_window_trigger_reference_py3 import SelfDependencyTumblingWindowTriggerReference + from .trigger_reference_py3 import TriggerReference + from .tumbling_window_trigger_dependency_reference_py3 import TumblingWindowTriggerDependencyReference + from .trigger_dependency_reference_py3 import TriggerDependencyReference + from .dependency_reference_py3 import DependencyReference + from .retry_policy_py3 import RetryPolicy + from .tumbling_window_trigger_py3 import TumblingWindowTrigger + from .blob_events_trigger_py3 import BlobEventsTrigger + from .blob_trigger_py3 import BlobTrigger + from .recurrence_schedule_occurrence_py3 import RecurrenceScheduleOccurrence + from .recurrence_schedule_py3 import RecurrenceSchedule + from .schedule_trigger_recurrence_py3 import ScheduleTriggerRecurrence + from .schedule_trigger_py3 import ScheduleTrigger + from .multiple_pipeline_trigger_py3 import MultiplePipelineTrigger from .responsys_linked_service_py3 import ResponsysLinkedService from .azure_databricks_linked_service_py3 import AzureDatabricksLinkedService from .azure_data_lake_analytics_linked_service_py3 import AzureDataLakeAnalyticsLinkedService @@ -212,20 +229,6 @@ from .azure_table_dataset_py3 import AzureTableDataset from .azure_blob_dataset_py3 import AzureBlobDataset from .amazon_s3_dataset_py3 import AmazonS3Dataset - from .self_dependency_tumbling_window_trigger_reference_py3 import SelfDependencyTumblingWindowTriggerReference - from .trigger_reference_py3 import TriggerReference - from .tumbling_window_trigger_dependency_reference_py3 import TumblingWindowTriggerDependencyReference - from .trigger_dependency_reference_py3 import TriggerDependencyReference - from .dependency_reference_py3 import DependencyReference - from .retry_policy_py3 import RetryPolicy - from .tumbling_window_trigger_py3 import TumblingWindowTrigger - from .blob_events_trigger_py3 import BlobEventsTrigger - from .blob_trigger_py3 import BlobTrigger - from .recurrence_schedule_occurrence_py3 import RecurrenceScheduleOccurrence - from .recurrence_schedule_py3 import RecurrenceSchedule - from .schedule_trigger_recurrence_py3 import ScheduleTriggerRecurrence - from .schedule_trigger_py3 import ScheduleTrigger - from .multiple_pipeline_trigger_py3 import MultiplePipelineTrigger from .activity_policy_py3 import ActivityPolicy from .databricks_spark_python_activity_py3 import DatabricksSparkPythonActivity from .databricks_spark_jar_activity_py3 import DatabricksSparkJarActivity @@ -413,6 +416,9 @@ from .activity_runs_query_response import ActivityRunsQueryResponse from .trigger_run import TriggerRun from .trigger_runs_query_response import TriggerRunsQueryResponse + from .rerun_tumbling_window_trigger_action_parameters import RerunTumblingWindowTriggerActionParameters + from .rerun_tumbling_window_trigger import RerunTumblingWindowTrigger + from .rerun_trigger_resource import RerunTriggerResource from .operation_display import OperationDisplay from .operation_log_specification import OperationLogSpecification from .operation_metric_availability import OperationMetricAvailability @@ -420,6 +426,20 @@ from .operation_metric_specification import OperationMetricSpecification from .operation_service_specification import OperationServiceSpecification from .operation import Operation + from .self_dependency_tumbling_window_trigger_reference import SelfDependencyTumblingWindowTriggerReference + from .trigger_reference import TriggerReference + from .tumbling_window_trigger_dependency_reference import TumblingWindowTriggerDependencyReference + from .trigger_dependency_reference import TriggerDependencyReference + from .dependency_reference import DependencyReference + from .retry_policy import RetryPolicy + from .tumbling_window_trigger import TumblingWindowTrigger + from .blob_events_trigger import BlobEventsTrigger + from .blob_trigger import BlobTrigger + from .recurrence_schedule_occurrence import RecurrenceScheduleOccurrence + from .recurrence_schedule import RecurrenceSchedule + from .schedule_trigger_recurrence import ScheduleTriggerRecurrence + from .schedule_trigger import ScheduleTrigger + from .multiple_pipeline_trigger import MultiplePipelineTrigger from .responsys_linked_service import ResponsysLinkedService from .azure_databricks_linked_service import AzureDatabricksLinkedService from .azure_data_lake_analytics_linked_service import AzureDataLakeAnalyticsLinkedService @@ -561,20 +581,6 @@ from .azure_table_dataset import AzureTableDataset from .azure_blob_dataset import AzureBlobDataset from .amazon_s3_dataset import AmazonS3Dataset - from .self_dependency_tumbling_window_trigger_reference import SelfDependencyTumblingWindowTriggerReference - from .trigger_reference import TriggerReference - from .tumbling_window_trigger_dependency_reference import TumblingWindowTriggerDependencyReference - from .trigger_dependency_reference import TriggerDependencyReference - from .dependency_reference import DependencyReference - from .retry_policy import RetryPolicy - from .tumbling_window_trigger import TumblingWindowTrigger - from .blob_events_trigger import BlobEventsTrigger - from .blob_trigger import BlobTrigger - from .recurrence_schedule_occurrence import RecurrenceScheduleOccurrence - from .recurrence_schedule import RecurrenceSchedule - from .schedule_trigger_recurrence import ScheduleTriggerRecurrence - from .schedule_trigger import ScheduleTrigger - from .multiple_pipeline_trigger import MultiplePipelineTrigger from .activity_policy import ActivityPolicy from .databricks_spark_python_activity import DatabricksSparkPythonActivity from .databricks_spark_jar_activity import DatabricksSparkJarActivity @@ -714,6 +720,7 @@ from .dataset_resource_paged import DatasetResourcePaged from .pipeline_resource_paged import PipelineResourcePaged from .trigger_resource_paged import TriggerResourcePaged +from .rerun_trigger_resource_paged import RerunTriggerResourcePaged from .data_factory_management_client_enums import ( IntegrationRuntimeState, IntegrationRuntimeAutoUpdate, @@ -726,6 +733,11 @@ RunQueryOrderByField, RunQueryOrder, TriggerRunStatus, + TumblingWindowFrequency, + BlobEventTypes, + DayOfWeek, + DaysOfWeek, + RecurrenceFrequency, SparkServerType, SparkThriftTransportProtocol, SparkAuthenticationType, @@ -749,11 +761,6 @@ SybaseAuthenticationType, DatasetCompressionLevel, JsonFormatFilePattern, - TumblingWindowFrequency, - BlobEventTypes, - DayOfWeek, - DaysOfWeek, - RecurrenceFrequency, WebActivityMethod, CassandraSourceReadConsistencyLevels, StoredProcedureParameterType, @@ -831,6 +838,9 @@ 'ActivityRunsQueryResponse', 'TriggerRun', 'TriggerRunsQueryResponse', + 'RerunTumblingWindowTriggerActionParameters', + 'RerunTumblingWindowTrigger', + 'RerunTriggerResource', 'OperationDisplay', 'OperationLogSpecification', 'OperationMetricAvailability', @@ -838,6 +848,20 @@ 'OperationMetricSpecification', 'OperationServiceSpecification', 'Operation', + 'SelfDependencyTumblingWindowTriggerReference', + 'TriggerReference', + 'TumblingWindowTriggerDependencyReference', + 'TriggerDependencyReference', + 'DependencyReference', + 'RetryPolicy', + 'TumblingWindowTrigger', + 'BlobEventsTrigger', + 'BlobTrigger', + 'RecurrenceScheduleOccurrence', + 'RecurrenceSchedule', + 'ScheduleTriggerRecurrence', + 'ScheduleTrigger', + 'MultiplePipelineTrigger', 'ResponsysLinkedService', 'AzureDatabricksLinkedService', 'AzureDataLakeAnalyticsLinkedService', @@ -979,20 +1003,6 @@ 'AzureTableDataset', 'AzureBlobDataset', 'AmazonS3Dataset', - 'SelfDependencyTumblingWindowTriggerReference', - 'TriggerReference', - 'TumblingWindowTriggerDependencyReference', - 'TriggerDependencyReference', - 'DependencyReference', - 'RetryPolicy', - 'TumblingWindowTrigger', - 'BlobEventsTrigger', - 'BlobTrigger', - 'RecurrenceScheduleOccurrence', - 'RecurrenceSchedule', - 'ScheduleTriggerRecurrence', - 'ScheduleTrigger', - 'MultiplePipelineTrigger', 'ActivityPolicy', 'DatabricksSparkPythonActivity', 'DatabricksSparkJarActivity', @@ -1132,6 +1142,7 @@ 'DatasetResourcePaged', 'PipelineResourcePaged', 'TriggerResourcePaged', + 'RerunTriggerResourcePaged', 'IntegrationRuntimeState', 'IntegrationRuntimeAutoUpdate', 'ParameterType', @@ -1143,6 +1154,11 @@ 'RunQueryOrderByField', 'RunQueryOrder', 'TriggerRunStatus', + 'TumblingWindowFrequency', + 'BlobEventTypes', + 'DayOfWeek', + 'DaysOfWeek', + 'RecurrenceFrequency', 'SparkServerType', 'SparkThriftTransportProtocol', 'SparkAuthenticationType', @@ -1166,11 +1182,6 @@ 'SybaseAuthenticationType', 'DatasetCompressionLevel', 'JsonFormatFilePattern', - 'TumblingWindowFrequency', - 'BlobEventTypes', - 'DayOfWeek', - 'DaysOfWeek', - 'RecurrenceFrequency', 'WebActivityMethod', 'CassandraSourceReadConsistencyLevels', 'StoredProcedureParameterType', diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_factory_management_client_enums.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_factory_management_client_enums.py index 2de647ff9d8b..aa76fe05ddc0 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_factory_management_client_enums.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_factory_management_client_enums.py @@ -113,6 +113,51 @@ class TriggerRunStatus(str, Enum): inprogress = "Inprogress" +class TumblingWindowFrequency(str, Enum): + + minute = "Minute" + hour = "Hour" + + +class BlobEventTypes(str, Enum): + + microsoft_storage_blob_created = "Microsoft.Storage.BlobCreated" + microsoft_storage_blob_deleted = "Microsoft.Storage.BlobDeleted" + + +class DayOfWeek(str, Enum): + + sunday = "Sunday" + monday = "Monday" + tuesday = "Tuesday" + wednesday = "Wednesday" + thursday = "Thursday" + friday = "Friday" + saturday = "Saturday" + + +class DaysOfWeek(str, Enum): + + sunday = "Sunday" + monday = "Monday" + tuesday = "Tuesday" + wednesday = "Wednesday" + thursday = "Thursday" + friday = "Friday" + saturday = "Saturday" + + +class RecurrenceFrequency(str, Enum): + + not_specified = "NotSpecified" + minute = "Minute" + hour = "Hour" + day = "Day" + week = "Week" + month = "Month" + year = "Year" + + class SparkServerType(str, Enum): shark_server = "SharkServer" @@ -263,51 +308,6 @@ class JsonFormatFilePattern(str, Enum): array_of_objects = "arrayOfObjects" -class TumblingWindowFrequency(str, Enum): - - minute = "Minute" - hour = "Hour" - - -class BlobEventTypes(str, Enum): - - microsoft_storage_blob_created = "Microsoft.Storage.BlobCreated" - microsoft_storage_blob_deleted = "Microsoft.Storage.BlobDeleted" - - -class DayOfWeek(str, Enum): - - sunday = "Sunday" - monday = "Monday" - tuesday = "Tuesday" - wednesday = "Wednesday" - thursday = "Thursday" - friday = "Friday" - saturday = "Saturday" - - -class DaysOfWeek(str, Enum): - - sunday = "Sunday" - monday = "Monday" - tuesday = "Tuesday" - wednesday = "Wednesday" - thursday = "Thursday" - friday = "Friday" - saturday = "Saturday" - - -class RecurrenceFrequency(str, Enum): - - not_specified = "NotSpecified" - minute = "Minute" - hour = "Hour" - day = "Day" - week = "Week" - month = "Month" - year = "Year" - - class WebActivityMethod(str, Enum): get = "GET" diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/rerun_trigger_resource.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/rerun_trigger_resource.py new file mode 100644 index 000000000000..8de6a70ecc99 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/rerun_trigger_resource.py @@ -0,0 +1,54 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .sub_resource import SubResource + + +class RerunTriggerResource(SubResource): + """RerunTrigger resource type. + + Variables are only populated by the server, and will be ignored when + sending a request. + + All required parameters must be populated in order to send to Azure. + + :ivar id: The resource identifier. + :vartype id: str + :ivar name: The resource name. + :vartype name: str + :ivar type: The resource type. + :vartype type: str + :ivar etag: Etag identifies change in the resource. + :vartype etag: str + :param properties: Required. Properties of the rerun trigger. + :type properties: + ~azure.mgmt.datafactory.models.RerunTumblingWindowTrigger + """ + + _validation = { + 'id': {'readonly': True}, + 'name': {'readonly': True}, + 'type': {'readonly': True}, + 'etag': {'readonly': True}, + 'properties': {'required': True}, + } + + _attribute_map = { + 'id': {'key': 'id', 'type': 'str'}, + 'name': {'key': 'name', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'str'}, + 'etag': {'key': 'etag', 'type': 'str'}, + 'properties': {'key': 'properties', 'type': 'RerunTumblingWindowTrigger'}, + } + + def __init__(self, **kwargs): + super(RerunTriggerResource, self).__init__(**kwargs) + self.properties = kwargs.get('properties', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/rerun_trigger_resource_paged.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/rerun_trigger_resource_paged.py new file mode 100644 index 000000000000..23d971c1082e --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/rerun_trigger_resource_paged.py @@ -0,0 +1,27 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.paging import Paged + + +class RerunTriggerResourcePaged(Paged): + """ + A paging container for iterating over a list of :class:`RerunTriggerResource ` object + """ + + _attribute_map = { + 'next_link': {'key': 'nextLink', 'type': 'str'}, + 'current_page': {'key': 'value', 'type': '[RerunTriggerResource]'} + } + + def __init__(self, *args, **kwargs): + + super(RerunTriggerResourcePaged, self).__init__(*args, **kwargs) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/rerun_trigger_resource_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/rerun_trigger_resource_py3.py new file mode 100644 index 000000000000..19814ad0d76f --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/rerun_trigger_resource_py3.py @@ -0,0 +1,54 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .sub_resource_py3 import SubResource + + +class RerunTriggerResource(SubResource): + """RerunTrigger resource type. + + Variables are only populated by the server, and will be ignored when + sending a request. + + All required parameters must be populated in order to send to Azure. + + :ivar id: The resource identifier. + :vartype id: str + :ivar name: The resource name. + :vartype name: str + :ivar type: The resource type. + :vartype type: str + :ivar etag: Etag identifies change in the resource. + :vartype etag: str + :param properties: Required. Properties of the rerun trigger. + :type properties: + ~azure.mgmt.datafactory.models.RerunTumblingWindowTrigger + """ + + _validation = { + 'id': {'readonly': True}, + 'name': {'readonly': True}, + 'type': {'readonly': True}, + 'etag': {'readonly': True}, + 'properties': {'required': True}, + } + + _attribute_map = { + 'id': {'key': 'id', 'type': 'str'}, + 'name': {'key': 'name', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'str'}, + 'etag': {'key': 'etag', 'type': 'str'}, + 'properties': {'key': 'properties', 'type': 'RerunTumblingWindowTrigger'}, + } + + def __init__(self, *, properties, **kwargs) -> None: + super(RerunTriggerResource, self).__init__(**kwargs) + self.properties = properties diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/rerun_tumbling_window_trigger.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/rerun_tumbling_window_trigger.py new file mode 100644 index 000000000000..e66cf2feebbc --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/rerun_tumbling_window_trigger.py @@ -0,0 +1,74 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .trigger import Trigger + + +class RerunTumblingWindowTrigger(Trigger): + """Trigger that schedules pipeline reruns for all fixed time interval windows + from a requested start time to requested end time. + + Variables are only populated by the server, and will be ignored when + sending a request. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param description: Trigger description. + :type description: str + :ivar runtime_state: Indicates if trigger is running or not. Updated when + Start/Stop APIs are called on the Trigger. Possible values include: + 'Started', 'Stopped', 'Disabled' + :vartype runtime_state: str or + ~azure.mgmt.datafactory.models.TriggerRuntimeState + :param type: Required. Constant filled by server. + :type type: str + :param parent_trigger: The parent trigger reference. + :type parent_trigger: object + :param requested_start_time: Required. The start time for the time period + for which restatement is initiated. Only UTC time is currently supported. + :type requested_start_time: datetime + :param requested_end_time: Required. The end time for the time period for + which restatement is initiated. Only UTC time is currently supported. + :type requested_end_time: datetime + :param max_concurrency: Required. The max number of parallel time windows + (ready for execution) for which a rerun is triggered. + :type max_concurrency: int + """ + + _validation = { + 'runtime_state': {'readonly': True}, + 'type': {'required': True}, + 'requested_start_time': {'required': True}, + 'requested_end_time': {'required': True}, + 'max_concurrency': {'required': True, 'maximum': 50, 'minimum': 1}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'runtime_state': {'key': 'runtimeState', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'str'}, + 'parent_trigger': {'key': 'typeProperties.parentTrigger', 'type': 'object'}, + 'requested_start_time': {'key': 'typeProperties.requestedStartTime', 'type': 'iso-8601'}, + 'requested_end_time': {'key': 'typeProperties.requestedEndTime', 'type': 'iso-8601'}, + 'max_concurrency': {'key': 'typeProperties.maxConcurrency', 'type': 'int'}, + } + + def __init__(self, **kwargs): + super(RerunTumblingWindowTrigger, self).__init__(**kwargs) + self.parent_trigger = kwargs.get('parent_trigger', None) + self.requested_start_time = kwargs.get('requested_start_time', None) + self.requested_end_time = kwargs.get('requested_end_time', None) + self.max_concurrency = kwargs.get('max_concurrency', None) + self.type = 'RerunTumblingWindowTrigger' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/rerun_tumbling_window_trigger_action_parameters.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/rerun_tumbling_window_trigger_action_parameters.py new file mode 100644 index 000000000000..4b87f070b6be --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/rerun_tumbling_window_trigger_action_parameters.py @@ -0,0 +1,47 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class RerunTumblingWindowTriggerActionParameters(Model): + """Rerun tumbling window trigger Parameters. + + All required parameters must be populated in order to send to Azure. + + :param start_time: Required. The start time for the time period for which + restatement is initiated. Only UTC time is currently supported. + :type start_time: datetime + :param end_time: Required. The end time for the time period for which + restatement is initiated. Only UTC time is currently supported. + :type end_time: datetime + :param max_concurrency: Required. The max number of parallel time windows + (ready for execution) for which a rerun is triggered. + :type max_concurrency: int + """ + + _validation = { + 'start_time': {'required': True}, + 'end_time': {'required': True}, + 'max_concurrency': {'required': True, 'maximum': 50, 'minimum': 1}, + } + + _attribute_map = { + 'start_time': {'key': 'startTime', 'type': 'iso-8601'}, + 'end_time': {'key': 'endTime', 'type': 'iso-8601'}, + 'max_concurrency': {'key': 'maxConcurrency', 'type': 'int'}, + } + + def __init__(self, **kwargs): + super(RerunTumblingWindowTriggerActionParameters, self).__init__(**kwargs) + self.start_time = kwargs.get('start_time', None) + self.end_time = kwargs.get('end_time', None) + self.max_concurrency = kwargs.get('max_concurrency', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/rerun_tumbling_window_trigger_action_parameters_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/rerun_tumbling_window_trigger_action_parameters_py3.py new file mode 100644 index 000000000000..6fadecca588b --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/rerun_tumbling_window_trigger_action_parameters_py3.py @@ -0,0 +1,47 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class RerunTumblingWindowTriggerActionParameters(Model): + """Rerun tumbling window trigger Parameters. + + All required parameters must be populated in order to send to Azure. + + :param start_time: Required. The start time for the time period for which + restatement is initiated. Only UTC time is currently supported. + :type start_time: datetime + :param end_time: Required. The end time for the time period for which + restatement is initiated. Only UTC time is currently supported. + :type end_time: datetime + :param max_concurrency: Required. The max number of parallel time windows + (ready for execution) for which a rerun is triggered. + :type max_concurrency: int + """ + + _validation = { + 'start_time': {'required': True}, + 'end_time': {'required': True}, + 'max_concurrency': {'required': True, 'maximum': 50, 'minimum': 1}, + } + + _attribute_map = { + 'start_time': {'key': 'startTime', 'type': 'iso-8601'}, + 'end_time': {'key': 'endTime', 'type': 'iso-8601'}, + 'max_concurrency': {'key': 'maxConcurrency', 'type': 'int'}, + } + + def __init__(self, *, start_time, end_time, max_concurrency: int, **kwargs) -> None: + super(RerunTumblingWindowTriggerActionParameters, self).__init__(**kwargs) + self.start_time = start_time + self.end_time = end_time + self.max_concurrency = max_concurrency diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/rerun_tumbling_window_trigger_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/rerun_tumbling_window_trigger_py3.py new file mode 100644 index 000000000000..eafc3b5743a0 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/rerun_tumbling_window_trigger_py3.py @@ -0,0 +1,74 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .trigger_py3 import Trigger + + +class RerunTumblingWindowTrigger(Trigger): + """Trigger that schedules pipeline reruns for all fixed time interval windows + from a requested start time to requested end time. + + Variables are only populated by the server, and will be ignored when + sending a request. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param description: Trigger description. + :type description: str + :ivar runtime_state: Indicates if trigger is running or not. Updated when + Start/Stop APIs are called on the Trigger. Possible values include: + 'Started', 'Stopped', 'Disabled' + :vartype runtime_state: str or + ~azure.mgmt.datafactory.models.TriggerRuntimeState + :param type: Required. Constant filled by server. + :type type: str + :param parent_trigger: The parent trigger reference. + :type parent_trigger: object + :param requested_start_time: Required. The start time for the time period + for which restatement is initiated. Only UTC time is currently supported. + :type requested_start_time: datetime + :param requested_end_time: Required. The end time for the time period for + which restatement is initiated. Only UTC time is currently supported. + :type requested_end_time: datetime + :param max_concurrency: Required. The max number of parallel time windows + (ready for execution) for which a rerun is triggered. + :type max_concurrency: int + """ + + _validation = { + 'runtime_state': {'readonly': True}, + 'type': {'required': True}, + 'requested_start_time': {'required': True}, + 'requested_end_time': {'required': True}, + 'max_concurrency': {'required': True, 'maximum': 50, 'minimum': 1}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'runtime_state': {'key': 'runtimeState', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'str'}, + 'parent_trigger': {'key': 'typeProperties.parentTrigger', 'type': 'object'}, + 'requested_start_time': {'key': 'typeProperties.requestedStartTime', 'type': 'iso-8601'}, + 'requested_end_time': {'key': 'typeProperties.requestedEndTime', 'type': 'iso-8601'}, + 'max_concurrency': {'key': 'typeProperties.maxConcurrency', 'type': 'int'}, + } + + def __init__(self, *, requested_start_time, requested_end_time, max_concurrency: int, additional_properties=None, description: str=None, parent_trigger=None, **kwargs) -> None: + super(RerunTumblingWindowTrigger, self).__init__(additional_properties=additional_properties, description=description, **kwargs) + self.parent_trigger = parent_trigger + self.requested_start_time = requested_start_time + self.requested_end_time = requested_end_time + self.max_concurrency = max_concurrency + self.type = 'RerunTumblingWindowTrigger' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger.py index 0e7882159e95..398402178ae4 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger.py @@ -17,7 +17,8 @@ class Trigger(Model): pipeline run. You probably want to use the sub-classes and not this class directly. Known - sub-classes are: TumblingWindowTrigger, MultiplePipelineTrigger + sub-classes are: RerunTumblingWindowTrigger, TumblingWindowTrigger, + MultiplePipelineTrigger Variables are only populated by the server, and will be ignored when sending a request. @@ -51,7 +52,7 @@ class Trigger(Model): } _subtype_map = { - 'type': {'TumblingWindowTrigger': 'TumblingWindowTrigger', 'MultiplePipelineTrigger': 'MultiplePipelineTrigger'} + 'type': {'RerunTumblingWindowTrigger': 'RerunTumblingWindowTrigger', 'TumblingWindowTrigger': 'TumblingWindowTrigger', 'MultiplePipelineTrigger': 'MultiplePipelineTrigger'} } def __init__(self, **kwargs): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger_py3.py index 3e232b149f0b..09fb39534be1 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger_py3.py @@ -17,7 +17,8 @@ class Trigger(Model): pipeline run. You probably want to use the sub-classes and not this class directly. Known - sub-classes are: TumblingWindowTrigger, MultiplePipelineTrigger + sub-classes are: RerunTumblingWindowTrigger, TumblingWindowTrigger, + MultiplePipelineTrigger Variables are only populated by the server, and will be ignored when sending a request. @@ -51,7 +52,7 @@ class Trigger(Model): } _subtype_map = { - 'type': {'TumblingWindowTrigger': 'TumblingWindowTrigger', 'MultiplePipelineTrigger': 'MultiplePipelineTrigger'} + 'type': {'RerunTumblingWindowTrigger': 'RerunTumblingWindowTrigger', 'TumblingWindowTrigger': 'TumblingWindowTrigger', 'MultiplePipelineTrigger': 'MultiplePipelineTrigger'} } def __init__(self, *, additional_properties=None, description: str=None, **kwargs) -> None: diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/__init__.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/__init__.py index b6b9497ae922..a987d7ddfec7 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/__init__.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/__init__.py @@ -19,6 +19,7 @@ from .pipeline_runs_operations import PipelineRunsOperations from .activity_runs_operations import ActivityRunsOperations from .triggers_operations import TriggersOperations +from .rerun_triggers_operations import RerunTriggersOperations from .trigger_runs_operations import TriggerRunsOperations __all__ = [ @@ -32,5 +33,6 @@ 'PipelineRunsOperations', 'ActivityRunsOperations', 'TriggersOperations', + 'RerunTriggersOperations', 'TriggerRunsOperations', ] diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/rerun_triggers_operations.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/rerun_triggers_operations.py new file mode 100644 index 000000000000..58e0066a60dd --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/rerun_triggers_operations.py @@ -0,0 +1,450 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +import uuid +from msrest.pipeline import ClientRawResponse +from msrestazure.azure_exceptions import CloudError +from msrest.polling import LROPoller, NoPolling +from msrestazure.polling.arm_polling import ARMPolling + +from .. import models + + +class RerunTriggersOperations(object): + """RerunTriggersOperations operations. + + :param client: Client for service requests. + :param config: Configuration of service client. + :param serializer: An object model serializer. + :param deserializer: An object model deserializer. + :ivar api_version: The API version. Constant value: "2018-06-01". + """ + + models = models + + def __init__(self, client, config, serializer, deserializer): + + self._client = client + self._serialize = serializer + self._deserialize = deserializer + self.api_version = "2018-06-01" + + self.config = config + + def create( + self, resource_group_name, factory_name, trigger_name, rerun_trigger_name, rerun_tumbling_window_trigger_action_parameters, custom_headers=None, raw=False, **operation_config): + """Creates a rerun trigger. + + :param resource_group_name: The resource group name. + :type resource_group_name: str + :param factory_name: The factory name. + :type factory_name: str + :param trigger_name: The trigger name. + :type trigger_name: str + :param rerun_trigger_name: The rerun trigger name. + :type rerun_trigger_name: str + :param rerun_tumbling_window_trigger_action_parameters: Rerun tumbling + window trigger action parameters. + :type rerun_tumbling_window_trigger_action_parameters: + ~azure.mgmt.datafactory.models.RerunTumblingWindowTriggerActionParameters + :param dict custom_headers: headers that will be added to the request + :param bool raw: returns the direct response alongside the + deserialized response + :param operation_config: :ref:`Operation configuration + overrides`. + :return: TriggerResource or ClientRawResponse if raw=true + :rtype: ~azure.mgmt.datafactory.models.TriggerResource or + ~msrest.pipeline.ClientRawResponse + :raises: :class:`CloudError` + """ + # Construct URL + url = self.create.metadata['url'] + path_format_arguments = { + 'subscriptionId': self._serialize.url("self.config.subscription_id", self.config.subscription_id, 'str'), + 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str', max_length=90, min_length=1, pattern=r'^[-\w\._\(\)]+$'), + 'factoryName': self._serialize.url("factory_name", factory_name, 'str', max_length=63, min_length=3, pattern=r'^[A-Za-z0-9]+(?:-[A-Za-z0-9]+)*$'), + 'triggerName': self._serialize.url("trigger_name", trigger_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), + 'rerunTriggerName': self._serialize.url("rerun_trigger_name", rerun_trigger_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$') + } + url = self._client.format_url(url, **path_format_arguments) + + # Construct parameters + query_parameters = {} + query_parameters['api-version'] = self._serialize.query("self.api_version", self.api_version, 'str') + + # Construct headers + header_parameters = {} + header_parameters['Accept'] = 'application/json' + header_parameters['Content-Type'] = 'application/json; charset=utf-8' + if self.config.generate_client_request_id: + header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) + if custom_headers: + header_parameters.update(custom_headers) + if self.config.accept_language is not None: + header_parameters['accept-language'] = self._serialize.header("self.config.accept_language", self.config.accept_language, 'str') + + # Construct body + body_content = self._serialize.body(rerun_tumbling_window_trigger_action_parameters, 'RerunTumblingWindowTriggerActionParameters') + + # Construct and send request + request = self._client.put(url, query_parameters, header_parameters, body_content) + response = self._client.send(request, stream=False, **operation_config) + + if response.status_code not in [200]: + exp = CloudError(response) + exp.request_id = response.headers.get('x-ms-request-id') + raise exp + + deserialized = None + + if response.status_code == 200: + deserialized = self._deserialize('TriggerResource', response) + + if raw: + client_raw_response = ClientRawResponse(deserialized, response) + return client_raw_response + + return deserialized + create.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/triggers/{triggerName}/rerunTriggers/{rerunTriggerName}'} + + + def _start_initial( + self, resource_group_name, factory_name, trigger_name, rerun_trigger_name, custom_headers=None, raw=False, **operation_config): + # Construct URL + url = self.start.metadata['url'] + path_format_arguments = { + 'subscriptionId': self._serialize.url("self.config.subscription_id", self.config.subscription_id, 'str'), + 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str', max_length=90, min_length=1, pattern=r'^[-\w\._\(\)]+$'), + 'factoryName': self._serialize.url("factory_name", factory_name, 'str', max_length=63, min_length=3, pattern=r'^[A-Za-z0-9]+(?:-[A-Za-z0-9]+)*$'), + 'triggerName': self._serialize.url("trigger_name", trigger_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), + 'rerunTriggerName': self._serialize.url("rerun_trigger_name", rerun_trigger_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$') + } + url = self._client.format_url(url, **path_format_arguments) + + # Construct parameters + query_parameters = {} + query_parameters['api-version'] = self._serialize.query("self.api_version", self.api_version, 'str') + + # Construct headers + header_parameters = {} + if self.config.generate_client_request_id: + header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) + if custom_headers: + header_parameters.update(custom_headers) + if self.config.accept_language is not None: + header_parameters['accept-language'] = self._serialize.header("self.config.accept_language", self.config.accept_language, 'str') + + # Construct and send request + request = self._client.post(url, query_parameters, header_parameters) + response = self._client.send(request, stream=False, **operation_config) + + if response.status_code not in [200]: + exp = CloudError(response) + exp.request_id = response.headers.get('x-ms-request-id') + raise exp + + if raw: + client_raw_response = ClientRawResponse(None, response) + return client_raw_response + + def start( + self, resource_group_name, factory_name, trigger_name, rerun_trigger_name, custom_headers=None, raw=False, polling=True, **operation_config): + """Starts a trigger. + + :param resource_group_name: The resource group name. + :type resource_group_name: str + :param factory_name: The factory name. + :type factory_name: str + :param trigger_name: The trigger name. + :type trigger_name: str + :param rerun_trigger_name: The rerun trigger name. + :type rerun_trigger_name: str + :param dict custom_headers: headers that will be added to the request + :param bool raw: The poller return type is ClientRawResponse, the + direct response alongside the deserialized response + :param polling: True for ARMPolling, False for no polling, or a + polling object for personal polling strategy + :return: An instance of LROPoller that returns None or + ClientRawResponse if raw==True + :rtype: ~msrestazure.azure_operation.AzureOperationPoller[None] or + ~msrestazure.azure_operation.AzureOperationPoller[~msrest.pipeline.ClientRawResponse[None]] + :raises: :class:`CloudError` + """ + raw_result = self._start_initial( + resource_group_name=resource_group_name, + factory_name=factory_name, + trigger_name=trigger_name, + rerun_trigger_name=rerun_trigger_name, + custom_headers=custom_headers, + raw=True, + **operation_config + ) + + def get_long_running_output(response): + if raw: + client_raw_response = ClientRawResponse(None, response) + return client_raw_response + + lro_delay = operation_config.get( + 'long_running_operation_timeout', + self.config.long_running_operation_timeout) + if polling is True: polling_method = ARMPolling(lro_delay, **operation_config) + elif polling is False: polling_method = NoPolling() + else: polling_method = polling + return LROPoller(self._client, raw_result, get_long_running_output, polling_method) + start.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/triggers/{triggerName}/rerunTriggers/{rerunTriggerName}/start'} + + + def _stop_initial( + self, resource_group_name, factory_name, trigger_name, rerun_trigger_name, custom_headers=None, raw=False, **operation_config): + # Construct URL + url = self.stop.metadata['url'] + path_format_arguments = { + 'subscriptionId': self._serialize.url("self.config.subscription_id", self.config.subscription_id, 'str'), + 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str', max_length=90, min_length=1, pattern=r'^[-\w\._\(\)]+$'), + 'factoryName': self._serialize.url("factory_name", factory_name, 'str', max_length=63, min_length=3, pattern=r'^[A-Za-z0-9]+(?:-[A-Za-z0-9]+)*$'), + 'triggerName': self._serialize.url("trigger_name", trigger_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), + 'rerunTriggerName': self._serialize.url("rerun_trigger_name", rerun_trigger_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$') + } + url = self._client.format_url(url, **path_format_arguments) + + # Construct parameters + query_parameters = {} + query_parameters['api-version'] = self._serialize.query("self.api_version", self.api_version, 'str') + + # Construct headers + header_parameters = {} + if self.config.generate_client_request_id: + header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) + if custom_headers: + header_parameters.update(custom_headers) + if self.config.accept_language is not None: + header_parameters['accept-language'] = self._serialize.header("self.config.accept_language", self.config.accept_language, 'str') + + # Construct and send request + request = self._client.post(url, query_parameters, header_parameters) + response = self._client.send(request, stream=False, **operation_config) + + if response.status_code not in [200]: + exp = CloudError(response) + exp.request_id = response.headers.get('x-ms-request-id') + raise exp + + if raw: + client_raw_response = ClientRawResponse(None, response) + return client_raw_response + + def stop( + self, resource_group_name, factory_name, trigger_name, rerun_trigger_name, custom_headers=None, raw=False, polling=True, **operation_config): + """Stops a trigger. + + :param resource_group_name: The resource group name. + :type resource_group_name: str + :param factory_name: The factory name. + :type factory_name: str + :param trigger_name: The trigger name. + :type trigger_name: str + :param rerun_trigger_name: The rerun trigger name. + :type rerun_trigger_name: str + :param dict custom_headers: headers that will be added to the request + :param bool raw: The poller return type is ClientRawResponse, the + direct response alongside the deserialized response + :param polling: True for ARMPolling, False for no polling, or a + polling object for personal polling strategy + :return: An instance of LROPoller that returns None or + ClientRawResponse if raw==True + :rtype: ~msrestazure.azure_operation.AzureOperationPoller[None] or + ~msrestazure.azure_operation.AzureOperationPoller[~msrest.pipeline.ClientRawResponse[None]] + :raises: :class:`CloudError` + """ + raw_result = self._stop_initial( + resource_group_name=resource_group_name, + factory_name=factory_name, + trigger_name=trigger_name, + rerun_trigger_name=rerun_trigger_name, + custom_headers=custom_headers, + raw=True, + **operation_config + ) + + def get_long_running_output(response): + if raw: + client_raw_response = ClientRawResponse(None, response) + return client_raw_response + + lro_delay = operation_config.get( + 'long_running_operation_timeout', + self.config.long_running_operation_timeout) + if polling is True: polling_method = ARMPolling(lro_delay, **operation_config) + elif polling is False: polling_method = NoPolling() + else: polling_method = polling + return LROPoller(self._client, raw_result, get_long_running_output, polling_method) + stop.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/triggers/{triggerName}/rerunTriggers/{rerunTriggerName}/stop'} + + + def _cancel_initial( + self, resource_group_name, factory_name, trigger_name, rerun_trigger_name, custom_headers=None, raw=False, **operation_config): + # Construct URL + url = self.cancel.metadata['url'] + path_format_arguments = { + 'subscriptionId': self._serialize.url("self.config.subscription_id", self.config.subscription_id, 'str'), + 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str', max_length=90, min_length=1, pattern=r'^[-\w\._\(\)]+$'), + 'factoryName': self._serialize.url("factory_name", factory_name, 'str', max_length=63, min_length=3, pattern=r'^[A-Za-z0-9]+(?:-[A-Za-z0-9]+)*$'), + 'triggerName': self._serialize.url("trigger_name", trigger_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), + 'rerunTriggerName': self._serialize.url("rerun_trigger_name", rerun_trigger_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$') + } + url = self._client.format_url(url, **path_format_arguments) + + # Construct parameters + query_parameters = {} + query_parameters['api-version'] = self._serialize.query("self.api_version", self.api_version, 'str') + + # Construct headers + header_parameters = {} + if self.config.generate_client_request_id: + header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) + if custom_headers: + header_parameters.update(custom_headers) + if self.config.accept_language is not None: + header_parameters['accept-language'] = self._serialize.header("self.config.accept_language", self.config.accept_language, 'str') + + # Construct and send request + request = self._client.post(url, query_parameters, header_parameters) + response = self._client.send(request, stream=False, **operation_config) + + if response.status_code not in [200]: + exp = CloudError(response) + exp.request_id = response.headers.get('x-ms-request-id') + raise exp + + if raw: + client_raw_response = ClientRawResponse(None, response) + return client_raw_response + + def cancel( + self, resource_group_name, factory_name, trigger_name, rerun_trigger_name, custom_headers=None, raw=False, polling=True, **operation_config): + """Cancels a trigger. + + :param resource_group_name: The resource group name. + :type resource_group_name: str + :param factory_name: The factory name. + :type factory_name: str + :param trigger_name: The trigger name. + :type trigger_name: str + :param rerun_trigger_name: The rerun trigger name. + :type rerun_trigger_name: str + :param dict custom_headers: headers that will be added to the request + :param bool raw: The poller return type is ClientRawResponse, the + direct response alongside the deserialized response + :param polling: True for ARMPolling, False for no polling, or a + polling object for personal polling strategy + :return: An instance of LROPoller that returns None or + ClientRawResponse if raw==True + :rtype: ~msrestazure.azure_operation.AzureOperationPoller[None] or + ~msrestazure.azure_operation.AzureOperationPoller[~msrest.pipeline.ClientRawResponse[None]] + :raises: :class:`CloudError` + """ + raw_result = self._cancel_initial( + resource_group_name=resource_group_name, + factory_name=factory_name, + trigger_name=trigger_name, + rerun_trigger_name=rerun_trigger_name, + custom_headers=custom_headers, + raw=True, + **operation_config + ) + + def get_long_running_output(response): + if raw: + client_raw_response = ClientRawResponse(None, response) + return client_raw_response + + lro_delay = operation_config.get( + 'long_running_operation_timeout', + self.config.long_running_operation_timeout) + if polling is True: polling_method = ARMPolling(lro_delay, **operation_config) + elif polling is False: polling_method = NoPolling() + else: polling_method = polling + return LROPoller(self._client, raw_result, get_long_running_output, polling_method) + cancel.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/triggers/{triggerName}/rerunTriggers/{rerunTriggerName}/cancel'} + + def list_by_trigger( + self, resource_group_name, factory_name, trigger_name, custom_headers=None, raw=False, **operation_config): + """Lists rerun triggers by an original trigger name. + + :param resource_group_name: The resource group name. + :type resource_group_name: str + :param factory_name: The factory name. + :type factory_name: str + :param trigger_name: The trigger name. + :type trigger_name: str + :param dict custom_headers: headers that will be added to the request + :param bool raw: returns the direct response alongside the + deserialized response + :param operation_config: :ref:`Operation configuration + overrides`. + :return: An iterator like instance of RerunTriggerResource + :rtype: + ~azure.mgmt.datafactory.models.RerunTriggerResourcePaged[~azure.mgmt.datafactory.models.RerunTriggerResource] + :raises: :class:`CloudError` + """ + def internal_paging(next_link=None, raw=False): + + if not next_link: + # Construct URL + url = self.list_by_trigger.metadata['url'] + path_format_arguments = { + 'subscriptionId': self._serialize.url("self.config.subscription_id", self.config.subscription_id, 'str'), + 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str', max_length=90, min_length=1, pattern=r'^[-\w\._\(\)]+$'), + 'factoryName': self._serialize.url("factory_name", factory_name, 'str', max_length=63, min_length=3, pattern=r'^[A-Za-z0-9]+(?:-[A-Za-z0-9]+)*$'), + 'triggerName': self._serialize.url("trigger_name", trigger_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$') + } + url = self._client.format_url(url, **path_format_arguments) + + # Construct parameters + query_parameters = {} + query_parameters['api-version'] = self._serialize.query("self.api_version", self.api_version, 'str') + + else: + url = next_link + query_parameters = {} + + # Construct headers + header_parameters = {} + header_parameters['Accept'] = 'application/json' + if self.config.generate_client_request_id: + header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) + if custom_headers: + header_parameters.update(custom_headers) + if self.config.accept_language is not None: + header_parameters['accept-language'] = self._serialize.header("self.config.accept_language", self.config.accept_language, 'str') + + # Construct and send request + request = self._client.get(url, query_parameters, header_parameters) + response = self._client.send(request, stream=False, **operation_config) + + if response.status_code not in [200]: + exp = CloudError(response) + exp.request_id = response.headers.get('x-ms-request-id') + raise exp + + return response + + # Deserialize response + deserialized = models.RerunTriggerResourcePaged(internal_paging, self._deserialize.dependencies) + + if raw: + header_dict = {} + client_raw_response = models.RerunTriggerResourcePaged(internal_paging, self._deserialize.dependencies, header_dict) + return client_raw_response + + return deserialized + list_by_trigger.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/triggers/{triggerName}/rerunTriggers'} From ca8b5b3f8cc1ca0fbb44d5a1e037c88597c91324 Mon Sep 17 00:00:00 2001 From: Azure SDK for Python bot Date: Thu, 18 Oct 2018 11:25:09 -0700 Subject: [PATCH 26/54] [AutoPR datafactory/resource-manager] Swagger changes for connectors: Google AdWords, Oracle Service Cloud, OData, Rest Service, Dynamics AX and support for wildcardPath. (#3637) * Generated from b0b2f8cc41a04777a10938d375bbebcef6389bf3 Zhenquan Xu: swagger for connector Dynamics AX. * Packaging update of azure-mgmt-datafactory --- azure-mgmt-datafactory/MANIFEST.in | 3 + .../azure/mgmt/datafactory/models/__init__.py | 42 +++++++ .../datafactory/models/amazon_s3_dataset.py | 5 + .../models/amazon_s3_dataset_py3.py | 7 +- .../datafactory/models/azure_blob_dataset.py | 6 + .../models/azure_blob_dataset_py3.py | 8 +- .../models/azure_data_lake_store_dataset.py | 11 +- .../azure_data_lake_store_dataset_py3.py | 13 +- .../mgmt/datafactory/models/copy_source.py | 15 +-- .../datafactory/models/copy_source_py3.py | 15 +-- .../data_factory_management_client_enums.py | 25 +++- .../azure/mgmt/datafactory/models/dataset.py | 19 +-- .../mgmt/datafactory/models/dataset_py3.py | 19 +-- .../models/dynamics_ax_linked_service.py | 90 +++++++++++++ .../models/dynamics_ax_linked_service_py3.py | 90 +++++++++++++ .../models/dynamics_ax_resource_dataset.py | 68 ++++++++++ .../dynamics_ax_resource_dataset_py3.py | 68 ++++++++++ .../datafactory/models/dynamics_ax_source.py | 52 ++++++++ .../models/dynamics_ax_source_py3.py | 52 ++++++++ .../datafactory/models/file_share_dataset.py | 6 + .../models/file_share_dataset_py3.py | 8 +- .../models/google_ad_words_linked_service.py | 119 ++++++++++++++++++ .../google_ad_words_linked_service_py3.py | 119 ++++++++++++++++++ .../models/google_ad_words_object_dataset.py | 62 +++++++++ .../google_ad_words_object_dataset_py3.py | 62 +++++++++ .../models/google_ad_words_source.py | 52 ++++++++ .../models/google_ad_words_source_py3.py | 52 ++++++++ .../mgmt/datafactory/models/linked_service.py | 17 +-- .../datafactory/models/linked_service_py3.py | 17 +-- .../models/odata_linked_service.py | 52 +++++++- .../models/odata_linked_service_py3.py | 54 +++++++- .../oracle_service_cloud_linked_service.py | 95 ++++++++++++++ ...oracle_service_cloud_linked_service_py3.py | 95 ++++++++++++++ .../oracle_service_cloud_object_dataset.py | 62 +++++++++ ...oracle_service_cloud_object_dataset_py3.py | 62 +++++++++ .../models/oracle_service_cloud_source.py | 52 ++++++++ .../models/oracle_service_cloud_source_py3.py | 52 ++++++++ .../models/rest_service_dataset.py | 88 +++++++++++++ .../models/rest_service_dataset_py3.py | 88 +++++++++++++ .../models/rest_service_linked_service.py | 107 ++++++++++++++++ .../models/rest_service_linked_service_py3.py | 107 ++++++++++++++++ .../datafactory/models/rest_service_source.py | 62 +++++++++ .../models/rest_service_source_py3.py | 62 +++++++++ .../models/stored_procedure_parameter.py | 8 +- .../models/stored_procedure_parameter_py3.py | 10 +- 45 files changed, 2099 insertions(+), 79 deletions(-) create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_ax_linked_service.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_ax_linked_service_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_ax_resource_dataset.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_ax_resource_dataset_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_ax_source.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_ax_source_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_ad_words_linked_service.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_ad_words_linked_service_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_ad_words_object_dataset.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_ad_words_object_dataset_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_ad_words_source.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_ad_words_source_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_service_cloud_linked_service.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_service_cloud_linked_service_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_service_cloud_object_dataset.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_service_cloud_object_dataset_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_service_cloud_source.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_service_cloud_source_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/rest_service_dataset.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/rest_service_dataset_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/rest_service_linked_service.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/rest_service_linked_service_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/rest_service_source.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/rest_service_source_py3.py diff --git a/azure-mgmt-datafactory/MANIFEST.in b/azure-mgmt-datafactory/MANIFEST.in index bb37a2723dae..6ceb27f7a96e 100644 --- a/azure-mgmt-datafactory/MANIFEST.in +++ b/azure-mgmt-datafactory/MANIFEST.in @@ -1 +1,4 @@ include *.rst +include azure/__init__.py +include azure/mgmt/__init__.py + diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py index cab29db469a7..ffdfed6c3c53 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py @@ -116,6 +116,7 @@ from .hbase_linked_service_py3 import HBaseLinkedService from .greenplum_linked_service_py3 import GreenplumLinkedService from .google_big_query_linked_service_py3 import GoogleBigQueryLinkedService + from .google_ad_words_linked_service_py3 import GoogleAdWordsLinkedService from .eloqua_linked_service_py3 import EloquaLinkedService from .drill_linked_service_py3 import DrillLinkedService from .couchbase_linked_service_py3 import CouchbaseLinkedService @@ -126,6 +127,7 @@ from .sap_bw_linked_service_py3 import SapBWLinkedService from .sftp_server_linked_service_py3 import SftpServerLinkedService from .ftp_server_linked_service_py3 import FtpServerLinkedService + from .rest_service_linked_service_py3 import RestServiceLinkedService from .http_linked_service_py3 import HttpLinkedService from .azure_search_linked_service_py3 import AzureSearchLinkedService from .custom_data_source_linked_service_py3 import CustomDataSourceLinkedService @@ -152,9 +154,11 @@ from .postgre_sql_linked_service_py3 import PostgreSqlLinkedService from .my_sql_linked_service_py3 import MySqlLinkedService from .azure_my_sql_linked_service_py3 import AzureMySqlLinkedService + from .oracle_service_cloud_linked_service_py3 import OracleServiceCloudLinkedService from .oracle_linked_service_py3 import OracleLinkedService from .file_server_linked_service_py3 import FileServerLinkedService from .hd_insight_linked_service_py3 import HDInsightLinkedService + from .dynamics_ax_linked_service_py3 import DynamicsAXLinkedService from .dynamics_linked_service_py3 import DynamicsLinkedService from .cosmos_db_linked_service_py3 import CosmosDbLinkedService from .azure_key_vault_linked_service_py3 import AzureKeyVaultLinkedService @@ -189,12 +193,14 @@ from .hbase_object_dataset_py3 import HBaseObjectDataset from .greenplum_table_dataset_py3 import GreenplumTableDataset from .google_big_query_object_dataset_py3 import GoogleBigQueryObjectDataset + from .google_ad_words_object_dataset_py3 import GoogleAdWordsObjectDataset from .eloqua_object_dataset_py3 import EloquaObjectDataset from .drill_table_dataset_py3 import DrillTableDataset from .couchbase_table_dataset_py3 import CouchbaseTableDataset from .concur_object_dataset_py3 import ConcurObjectDataset from .azure_postgre_sql_table_dataset_py3 import AzurePostgreSqlTableDataset from .amazon_mws_object_dataset_py3 import AmazonMWSObjectDataset + from .rest_service_dataset_py3 import RestServiceDataset from .dataset_zip_deflate_compression_py3 import DatasetZipDeflateCompression from .dataset_deflate_compression_py3 import DatasetDeflateCompression from .dataset_gzip_compression_py3 import DatasetGZipCompression @@ -215,11 +221,13 @@ from .salesforce_object_dataset_py3 import SalesforceObjectDataset from .relational_table_dataset_py3 import RelationalTableDataset from .azure_my_sql_table_dataset_py3 import AzureMySqlTableDataset + from .oracle_service_cloud_object_dataset_py3 import OracleServiceCloudObjectDataset from .oracle_table_dataset_py3 import OracleTableDataset from .odata_resource_dataset_py3 import ODataResourceDataset from .mongo_db_collection_dataset_py3 import MongoDbCollectionDataset from .file_share_dataset_py3 import FileShareDataset from .azure_data_lake_store_dataset_py3 import AzureDataLakeStoreDataset + from .dynamics_ax_resource_dataset_py3 import DynamicsAXResourceDataset from .dynamics_entity_dataset_py3 import DynamicsEntityDataset from .document_db_collection_dataset_py3 import DocumentDbCollectionDataset from .custom_dataset_py3 import CustomDataset @@ -266,17 +274,20 @@ from .hbase_source_py3 import HBaseSource from .greenplum_source_py3 import GreenplumSource from .google_big_query_source_py3 import GoogleBigQuerySource + from .google_ad_words_source_py3 import GoogleAdWordsSource from .eloqua_source_py3 import EloquaSource from .drill_source_py3 import DrillSource from .couchbase_source_py3 import CouchbaseSource from .concur_source_py3 import ConcurSource from .azure_postgre_sql_source_py3 import AzurePostgreSqlSource from .amazon_mws_source_py3 import AmazonMWSSource + from .rest_service_source_py3 import RestServiceSource from .http_source_py3 import HttpSource from .azure_data_lake_store_source_py3 import AzureDataLakeStoreSource from .mongo_db_source_py3 import MongoDbSource from .cassandra_source_py3 import CassandraSource from .web_source_py3 import WebSource + from .oracle_service_cloud_source_py3 import OracleServiceCloudSource from .oracle_source_py3 import OracleSource from .azure_my_sql_source_py3 import AzureMySqlSource from .distcp_settings_py3 import DistcpSettings @@ -289,6 +300,7 @@ from .sap_cloud_for_customer_source_py3 import SapCloudForCustomerSource from .salesforce_source_py3 import SalesforceSource from .relational_source_py3 import RelationalSource + from .dynamics_ax_source_py3 import DynamicsAXSource from .dynamics_source_py3 import DynamicsSource from .document_db_collection_source_py3 import DocumentDbCollectionSource from .blob_source_py3 import BlobSource @@ -468,6 +480,7 @@ from .hbase_linked_service import HBaseLinkedService from .greenplum_linked_service import GreenplumLinkedService from .google_big_query_linked_service import GoogleBigQueryLinkedService + from .google_ad_words_linked_service import GoogleAdWordsLinkedService from .eloqua_linked_service import EloquaLinkedService from .drill_linked_service import DrillLinkedService from .couchbase_linked_service import CouchbaseLinkedService @@ -478,6 +491,7 @@ from .sap_bw_linked_service import SapBWLinkedService from .sftp_server_linked_service import SftpServerLinkedService from .ftp_server_linked_service import FtpServerLinkedService + from .rest_service_linked_service import RestServiceLinkedService from .http_linked_service import HttpLinkedService from .azure_search_linked_service import AzureSearchLinkedService from .custom_data_source_linked_service import CustomDataSourceLinkedService @@ -504,9 +518,11 @@ from .postgre_sql_linked_service import PostgreSqlLinkedService from .my_sql_linked_service import MySqlLinkedService from .azure_my_sql_linked_service import AzureMySqlLinkedService + from .oracle_service_cloud_linked_service import OracleServiceCloudLinkedService from .oracle_linked_service import OracleLinkedService from .file_server_linked_service import FileServerLinkedService from .hd_insight_linked_service import HDInsightLinkedService + from .dynamics_ax_linked_service import DynamicsAXLinkedService from .dynamics_linked_service import DynamicsLinkedService from .cosmos_db_linked_service import CosmosDbLinkedService from .azure_key_vault_linked_service import AzureKeyVaultLinkedService @@ -541,12 +557,14 @@ from .hbase_object_dataset import HBaseObjectDataset from .greenplum_table_dataset import GreenplumTableDataset from .google_big_query_object_dataset import GoogleBigQueryObjectDataset + from .google_ad_words_object_dataset import GoogleAdWordsObjectDataset from .eloqua_object_dataset import EloquaObjectDataset from .drill_table_dataset import DrillTableDataset from .couchbase_table_dataset import CouchbaseTableDataset from .concur_object_dataset import ConcurObjectDataset from .azure_postgre_sql_table_dataset import AzurePostgreSqlTableDataset from .amazon_mws_object_dataset import AmazonMWSObjectDataset + from .rest_service_dataset import RestServiceDataset from .dataset_zip_deflate_compression import DatasetZipDeflateCompression from .dataset_deflate_compression import DatasetDeflateCompression from .dataset_gzip_compression import DatasetGZipCompression @@ -567,11 +585,13 @@ from .salesforce_object_dataset import SalesforceObjectDataset from .relational_table_dataset import RelationalTableDataset from .azure_my_sql_table_dataset import AzureMySqlTableDataset + from .oracle_service_cloud_object_dataset import OracleServiceCloudObjectDataset from .oracle_table_dataset import OracleTableDataset from .odata_resource_dataset import ODataResourceDataset from .mongo_db_collection_dataset import MongoDbCollectionDataset from .file_share_dataset import FileShareDataset from .azure_data_lake_store_dataset import AzureDataLakeStoreDataset + from .dynamics_ax_resource_dataset import DynamicsAXResourceDataset from .dynamics_entity_dataset import DynamicsEntityDataset from .document_db_collection_dataset import DocumentDbCollectionDataset from .custom_dataset import CustomDataset @@ -618,17 +638,20 @@ from .hbase_source import HBaseSource from .greenplum_source import GreenplumSource from .google_big_query_source import GoogleBigQuerySource + from .google_ad_words_source import GoogleAdWordsSource from .eloqua_source import EloquaSource from .drill_source import DrillSource from .couchbase_source import CouchbaseSource from .concur_source import ConcurSource from .azure_postgre_sql_source import AzurePostgreSqlSource from .amazon_mws_source import AmazonMWSSource + from .rest_service_source import RestServiceSource from .http_source import HttpSource from .azure_data_lake_store_source import AzureDataLakeStoreSource from .mongo_db_source import MongoDbSource from .cassandra_source import CassandraSource from .web_source import WebSource + from .oracle_service_cloud_source import OracleServiceCloudSource from .oracle_source import OracleSource from .azure_my_sql_source import AzureMySqlSource from .distcp_settings import DistcpSettings @@ -641,6 +664,7 @@ from .sap_cloud_for_customer_source import SapCloudForCustomerSource from .salesforce_source import SalesforceSource from .relational_source import RelationalSource + from .dynamics_ax_source import DynamicsAXSource from .dynamics_source import DynamicsSource from .document_db_collection_source import DocumentDbCollectionSource from .blob_source import BlobSource @@ -750,12 +774,15 @@ HiveAuthenticationType, HBaseAuthenticationType, GoogleBigQueryAuthenticationType, + GoogleAdWordsAuthenticationType, SapHanaAuthenticationType, SftpAuthenticationType, FtpAuthenticationType, + RestServiceAuthenticationType, HttpAuthenticationType, MongoDbAuthenticationType, ODataAuthenticationType, + ODataAadServicePrincipalCredentialType, TeradataAuthenticationType, Db2AuthenticationType, SybaseAuthenticationType, @@ -890,6 +917,7 @@ 'HBaseLinkedService', 'GreenplumLinkedService', 'GoogleBigQueryLinkedService', + 'GoogleAdWordsLinkedService', 'EloquaLinkedService', 'DrillLinkedService', 'CouchbaseLinkedService', @@ -900,6 +928,7 @@ 'SapBWLinkedService', 'SftpServerLinkedService', 'FtpServerLinkedService', + 'RestServiceLinkedService', 'HttpLinkedService', 'AzureSearchLinkedService', 'CustomDataSourceLinkedService', @@ -926,9 +955,11 @@ 'PostgreSqlLinkedService', 'MySqlLinkedService', 'AzureMySqlLinkedService', + 'OracleServiceCloudLinkedService', 'OracleLinkedService', 'FileServerLinkedService', 'HDInsightLinkedService', + 'DynamicsAXLinkedService', 'DynamicsLinkedService', 'CosmosDbLinkedService', 'AzureKeyVaultLinkedService', @@ -963,12 +994,14 @@ 'HBaseObjectDataset', 'GreenplumTableDataset', 'GoogleBigQueryObjectDataset', + 'GoogleAdWordsObjectDataset', 'EloquaObjectDataset', 'DrillTableDataset', 'CouchbaseTableDataset', 'ConcurObjectDataset', 'AzurePostgreSqlTableDataset', 'AmazonMWSObjectDataset', + 'RestServiceDataset', 'DatasetZipDeflateCompression', 'DatasetDeflateCompression', 'DatasetGZipCompression', @@ -989,11 +1022,13 @@ 'SalesforceObjectDataset', 'RelationalTableDataset', 'AzureMySqlTableDataset', + 'OracleServiceCloudObjectDataset', 'OracleTableDataset', 'ODataResourceDataset', 'MongoDbCollectionDataset', 'FileShareDataset', 'AzureDataLakeStoreDataset', + 'DynamicsAXResourceDataset', 'DynamicsEntityDataset', 'DocumentDbCollectionDataset', 'CustomDataset', @@ -1040,17 +1075,20 @@ 'HBaseSource', 'GreenplumSource', 'GoogleBigQuerySource', + 'GoogleAdWordsSource', 'EloquaSource', 'DrillSource', 'CouchbaseSource', 'ConcurSource', 'AzurePostgreSqlSource', 'AmazonMWSSource', + 'RestServiceSource', 'HttpSource', 'AzureDataLakeStoreSource', 'MongoDbSource', 'CassandraSource', 'WebSource', + 'OracleServiceCloudSource', 'OracleSource', 'AzureMySqlSource', 'DistcpSettings', @@ -1063,6 +1101,7 @@ 'SapCloudForCustomerSource', 'SalesforceSource', 'RelationalSource', + 'DynamicsAXSource', 'DynamicsSource', 'DocumentDbCollectionSource', 'BlobSource', @@ -1171,12 +1210,15 @@ 'HiveAuthenticationType', 'HBaseAuthenticationType', 'GoogleBigQueryAuthenticationType', + 'GoogleAdWordsAuthenticationType', 'SapHanaAuthenticationType', 'SftpAuthenticationType', 'FtpAuthenticationType', + 'RestServiceAuthenticationType', 'HttpAuthenticationType', 'MongoDbAuthenticationType', 'ODataAuthenticationType', + 'ODataAadServicePrincipalCredentialType', 'TeradataAuthenticationType', 'Db2AuthenticationType', 'SybaseAuthenticationType', diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_s3_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_s3_dataset.py index 87540eb0eb11..403e0e7811c4 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_s3_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_s3_dataset.py @@ -48,6 +48,9 @@ class AmazonS3Dataset(Dataset): :param prefix: The prefix filter for the S3 object name. Type: string (or Expression with resultType string). :type prefix: object + :param wildcard_path: The path of the S3 object with wildcard supported. + Type: string (or Expression with resultType string). + :type wildcard_path: object :param version: The version for the S3 object. Type: string (or Expression with resultType string). :type version: object @@ -76,6 +79,7 @@ class AmazonS3Dataset(Dataset): 'bucket_name': {'key': 'typeProperties.bucketName', 'type': 'object'}, 'key': {'key': 'typeProperties.key', 'type': 'object'}, 'prefix': {'key': 'typeProperties.prefix', 'type': 'object'}, + 'wildcard_path': {'key': 'typeProperties.wildcardPath', 'type': 'object'}, 'version': {'key': 'typeProperties.version', 'type': 'object'}, 'format': {'key': 'typeProperties.format', 'type': 'DatasetStorageFormat'}, 'compression': {'key': 'typeProperties.compression', 'type': 'DatasetCompression'}, @@ -86,6 +90,7 @@ def __init__(self, **kwargs): self.bucket_name = kwargs.get('bucket_name', None) self.key = kwargs.get('key', None) self.prefix = kwargs.get('prefix', None) + self.wildcard_path = kwargs.get('wildcard_path', None) self.version = kwargs.get('version', None) self.format = kwargs.get('format', None) self.compression = kwargs.get('compression', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_s3_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_s3_dataset_py3.py index c4b602da1303..09a3130c8c70 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_s3_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_s3_dataset_py3.py @@ -48,6 +48,9 @@ class AmazonS3Dataset(Dataset): :param prefix: The prefix filter for the S3 object name. Type: string (or Expression with resultType string). :type prefix: object + :param wildcard_path: The path of the S3 object with wildcard supported. + Type: string (or Expression with resultType string). + :type wildcard_path: object :param version: The version for the S3 object. Type: string (or Expression with resultType string). :type version: object @@ -76,16 +79,18 @@ class AmazonS3Dataset(Dataset): 'bucket_name': {'key': 'typeProperties.bucketName', 'type': 'object'}, 'key': {'key': 'typeProperties.key', 'type': 'object'}, 'prefix': {'key': 'typeProperties.prefix', 'type': 'object'}, + 'wildcard_path': {'key': 'typeProperties.wildcardPath', 'type': 'object'}, 'version': {'key': 'typeProperties.version', 'type': 'object'}, 'format': {'key': 'typeProperties.format', 'type': 'DatasetStorageFormat'}, 'compression': {'key': 'typeProperties.compression', 'type': 'DatasetCompression'}, } - def __init__(self, *, linked_service_name, bucket_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, key=None, prefix=None, version=None, format=None, compression=None, **kwargs) -> None: + def __init__(self, *, linked_service_name, bucket_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, key=None, prefix=None, wildcard_path=None, version=None, format=None, compression=None, **kwargs) -> None: super(AmazonS3Dataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.bucket_name = bucket_name self.key = key self.prefix = prefix + self.wildcard_path = wildcard_path self.version = version self.format = format self.compression = compression diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_dataset.py index dfa31040d658..f3070d3d6b46 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_dataset.py @@ -48,6 +48,10 @@ class AzureBlobDataset(Dataset): :param file_name: The name of the Azure Blob. Type: string (or Expression with resultType string). :type file_name: object + :param wildcard_path: The whole path include file name of the Azure Blob + storage with wildcard supported. Type: string (or Expression with + resultType string). + :type wildcard_path: object :param format: The format of the Azure Blob storage. :type format: ~azure.mgmt.datafactory.models.DatasetStorageFormat :param compression: The data compression method used for the blob storage. @@ -71,6 +75,7 @@ class AzureBlobDataset(Dataset): 'folder_path': {'key': 'typeProperties.folderPath', 'type': 'object'}, 'table_root_location': {'key': 'typeProperties.tableRootLocation', 'type': 'object'}, 'file_name': {'key': 'typeProperties.fileName', 'type': 'object'}, + 'wildcard_path': {'key': 'typeProperties.wildcardPath', 'type': 'object'}, 'format': {'key': 'typeProperties.format', 'type': 'DatasetStorageFormat'}, 'compression': {'key': 'typeProperties.compression', 'type': 'DatasetCompression'}, } @@ -80,6 +85,7 @@ def __init__(self, **kwargs): self.folder_path = kwargs.get('folder_path', None) self.table_root_location = kwargs.get('table_root_location', None) self.file_name = kwargs.get('file_name', None) + self.wildcard_path = kwargs.get('wildcard_path', None) self.format = kwargs.get('format', None) self.compression = kwargs.get('compression', None) self.type = 'AzureBlob' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_dataset_py3.py index adef2dbd6def..f84716fd9548 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_dataset_py3.py @@ -48,6 +48,10 @@ class AzureBlobDataset(Dataset): :param file_name: The name of the Azure Blob. Type: string (or Expression with resultType string). :type file_name: object + :param wildcard_path: The whole path include file name of the Azure Blob + storage with wildcard supported. Type: string (or Expression with + resultType string). + :type wildcard_path: object :param format: The format of the Azure Blob storage. :type format: ~azure.mgmt.datafactory.models.DatasetStorageFormat :param compression: The data compression method used for the blob storage. @@ -71,15 +75,17 @@ class AzureBlobDataset(Dataset): 'folder_path': {'key': 'typeProperties.folderPath', 'type': 'object'}, 'table_root_location': {'key': 'typeProperties.tableRootLocation', 'type': 'object'}, 'file_name': {'key': 'typeProperties.fileName', 'type': 'object'}, + 'wildcard_path': {'key': 'typeProperties.wildcardPath', 'type': 'object'}, 'format': {'key': 'typeProperties.format', 'type': 'DatasetStorageFormat'}, 'compression': {'key': 'typeProperties.compression', 'type': 'DatasetCompression'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, folder_path=None, table_root_location=None, file_name=None, format=None, compression=None, **kwargs) -> None: + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, folder_path=None, table_root_location=None, file_name=None, wildcard_path=None, format=None, compression=None, **kwargs) -> None: super(AzureBlobDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.folder_path = folder_path self.table_root_location = table_root_location self.file_name = file_name + self.wildcard_path = wildcard_path self.format = format self.compression = compression self.type = 'AzureBlob' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_dataset.py index 0d3d9edf45d9..5dccec2d1bfe 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_dataset.py @@ -39,14 +39,18 @@ class AzureDataLakeStoreDataset(Dataset): :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str - :param folder_path: Required. Path to the folder in the Azure Data Lake - Store. Type: string (or Expression with resultType string). + :param folder_path: Path to the folder in the Azure Data Lake Store. Type: + string (or Expression with resultType string). :type folder_path: object :param file_name: The name of the file in the Azure Data Lake Store. Type: string (or Expression with resultType string). :type file_name: object :param format: The format of the Data Lake Store. :type format: ~azure.mgmt.datafactory.models.DatasetStorageFormat + :param wildcard_path: The whole path include file name in the Azure Data + Lake Store with wildcard supported. Type: string (or Expression with + resultType string). + :type wildcard_path: object :param compression: The data compression method used for the item(s) in the Azure Data Lake Store. :type compression: ~azure.mgmt.datafactory.models.DatasetCompression @@ -55,7 +59,6 @@ class AzureDataLakeStoreDataset(Dataset): _validation = { 'linked_service_name': {'required': True}, 'type': {'required': True}, - 'folder_path': {'required': True}, } _attribute_map = { @@ -70,6 +73,7 @@ class AzureDataLakeStoreDataset(Dataset): 'folder_path': {'key': 'typeProperties.folderPath', 'type': 'object'}, 'file_name': {'key': 'typeProperties.fileName', 'type': 'object'}, 'format': {'key': 'typeProperties.format', 'type': 'DatasetStorageFormat'}, + 'wildcard_path': {'key': 'typeProperties.wildcardPath', 'type': 'object'}, 'compression': {'key': 'typeProperties.compression', 'type': 'DatasetCompression'}, } @@ -78,5 +82,6 @@ def __init__(self, **kwargs): self.folder_path = kwargs.get('folder_path', None) self.file_name = kwargs.get('file_name', None) self.format = kwargs.get('format', None) + self.wildcard_path = kwargs.get('wildcard_path', None) self.compression = kwargs.get('compression', None) self.type = 'AzureDataLakeStoreFile' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_dataset_py3.py index 8d108c944dcb..98b5432f9cf3 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_dataset_py3.py @@ -39,14 +39,18 @@ class AzureDataLakeStoreDataset(Dataset): :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str - :param folder_path: Required. Path to the folder in the Azure Data Lake - Store. Type: string (or Expression with resultType string). + :param folder_path: Path to the folder in the Azure Data Lake Store. Type: + string (or Expression with resultType string). :type folder_path: object :param file_name: The name of the file in the Azure Data Lake Store. Type: string (or Expression with resultType string). :type file_name: object :param format: The format of the Data Lake Store. :type format: ~azure.mgmt.datafactory.models.DatasetStorageFormat + :param wildcard_path: The whole path include file name in the Azure Data + Lake Store with wildcard supported. Type: string (or Expression with + resultType string). + :type wildcard_path: object :param compression: The data compression method used for the item(s) in the Azure Data Lake Store. :type compression: ~azure.mgmt.datafactory.models.DatasetCompression @@ -55,7 +59,6 @@ class AzureDataLakeStoreDataset(Dataset): _validation = { 'linked_service_name': {'required': True}, 'type': {'required': True}, - 'folder_path': {'required': True}, } _attribute_map = { @@ -70,13 +73,15 @@ class AzureDataLakeStoreDataset(Dataset): 'folder_path': {'key': 'typeProperties.folderPath', 'type': 'object'}, 'file_name': {'key': 'typeProperties.fileName', 'type': 'object'}, 'format': {'key': 'typeProperties.format', 'type': 'DatasetStorageFormat'}, + 'wildcard_path': {'key': 'typeProperties.wildcardPath', 'type': 'object'}, 'compression': {'key': 'typeProperties.compression', 'type': 'DatasetCompression'}, } - def __init__(self, *, linked_service_name, folder_path, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, file_name=None, format=None, compression=None, **kwargs) -> None: + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, folder_path=None, file_name=None, format=None, wildcard_path=None, compression=None, **kwargs) -> None: super(AzureDataLakeStoreDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.folder_path = folder_path self.file_name = file_name self.format = format + self.wildcard_path = wildcard_path self.compression = compression self.type = 'AzureDataLakeStoreFile' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_source.py index 9a11107fc8e8..8ad510c03d95 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_source.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_source.py @@ -22,12 +22,13 @@ class CopySource(Model): QuickBooksSource, PrestoSource, PhoenixSource, PaypalSource, MarketoSource, MariaDBSource, MagentoSource, JiraSource, ImpalaSource, HubspotSource, HiveSource, HBaseSource, GreenplumSource, GoogleBigQuerySource, - EloquaSource, DrillSource, CouchbaseSource, ConcurSource, - AzurePostgreSqlSource, AmazonMWSSource, HttpSource, - AzureDataLakeStoreSource, MongoDbSource, CassandraSource, WebSource, - OracleSource, AzureMySqlSource, HdfsSource, FileSystemSource, SqlDWSource, - SqlSource, SapEccSource, SapCloudForCustomerSource, SalesforceSource, - RelationalSource, DynamicsSource, DocumentDbCollectionSource, BlobSource, + GoogleAdWordsSource, EloquaSource, DrillSource, CouchbaseSource, + ConcurSource, AzurePostgreSqlSource, AmazonMWSSource, RestServiceSource, + HttpSource, AzureDataLakeStoreSource, MongoDbSource, CassandraSource, + WebSource, OracleServiceCloudSource, OracleSource, AzureMySqlSource, + HdfsSource, FileSystemSource, SqlDWSource, SqlSource, SapEccSource, + SapCloudForCustomerSource, SalesforceSource, RelationalSource, + DynamicsAXSource, DynamicsSource, DocumentDbCollectionSource, BlobSource, AzureTableSource All required parameters must be populated in order to send to Azure. @@ -58,7 +59,7 @@ class CopySource(Model): } _subtype_map = { - 'type': {'AmazonRedshiftSource': 'AmazonRedshiftSource', 'ResponsysSource': 'ResponsysSource', 'SalesforceMarketingCloudSource': 'SalesforceMarketingCloudSource', 'VerticaSource': 'VerticaSource', 'NetezzaSource': 'NetezzaSource', 'ZohoSource': 'ZohoSource', 'XeroSource': 'XeroSource', 'SquareSource': 'SquareSource', 'SparkSource': 'SparkSource', 'ShopifySource': 'ShopifySource', 'ServiceNowSource': 'ServiceNowSource', 'QuickBooksSource': 'QuickBooksSource', 'PrestoSource': 'PrestoSource', 'PhoenixSource': 'PhoenixSource', 'PaypalSource': 'PaypalSource', 'MarketoSource': 'MarketoSource', 'MariaDBSource': 'MariaDBSource', 'MagentoSource': 'MagentoSource', 'JiraSource': 'JiraSource', 'ImpalaSource': 'ImpalaSource', 'HubspotSource': 'HubspotSource', 'HiveSource': 'HiveSource', 'HBaseSource': 'HBaseSource', 'GreenplumSource': 'GreenplumSource', 'GoogleBigQuerySource': 'GoogleBigQuerySource', 'EloquaSource': 'EloquaSource', 'DrillSource': 'DrillSource', 'CouchbaseSource': 'CouchbaseSource', 'ConcurSource': 'ConcurSource', 'AzurePostgreSqlSource': 'AzurePostgreSqlSource', 'AmazonMWSSource': 'AmazonMWSSource', 'HttpSource': 'HttpSource', 'AzureDataLakeStoreSource': 'AzureDataLakeStoreSource', 'MongoDbSource': 'MongoDbSource', 'CassandraSource': 'CassandraSource', 'WebSource': 'WebSource', 'OracleSource': 'OracleSource', 'AzureMySqlSource': 'AzureMySqlSource', 'HdfsSource': 'HdfsSource', 'FileSystemSource': 'FileSystemSource', 'SqlDWSource': 'SqlDWSource', 'SqlSource': 'SqlSource', 'SapEccSource': 'SapEccSource', 'SapCloudForCustomerSource': 'SapCloudForCustomerSource', 'SalesforceSource': 'SalesforceSource', 'RelationalSource': 'RelationalSource', 'DynamicsSource': 'DynamicsSource', 'DocumentDbCollectionSource': 'DocumentDbCollectionSource', 'BlobSource': 'BlobSource', 'AzureTableSource': 'AzureTableSource'} + 'type': {'AmazonRedshiftSource': 'AmazonRedshiftSource', 'ResponsysSource': 'ResponsysSource', 'SalesforceMarketingCloudSource': 'SalesforceMarketingCloudSource', 'VerticaSource': 'VerticaSource', 'NetezzaSource': 'NetezzaSource', 'ZohoSource': 'ZohoSource', 'XeroSource': 'XeroSource', 'SquareSource': 'SquareSource', 'SparkSource': 'SparkSource', 'ShopifySource': 'ShopifySource', 'ServiceNowSource': 'ServiceNowSource', 'QuickBooksSource': 'QuickBooksSource', 'PrestoSource': 'PrestoSource', 'PhoenixSource': 'PhoenixSource', 'PaypalSource': 'PaypalSource', 'MarketoSource': 'MarketoSource', 'MariaDBSource': 'MariaDBSource', 'MagentoSource': 'MagentoSource', 'JiraSource': 'JiraSource', 'ImpalaSource': 'ImpalaSource', 'HubspotSource': 'HubspotSource', 'HiveSource': 'HiveSource', 'HBaseSource': 'HBaseSource', 'GreenplumSource': 'GreenplumSource', 'GoogleBigQuerySource': 'GoogleBigQuerySource', 'GoogleAdWordsSource': 'GoogleAdWordsSource', 'EloquaSource': 'EloquaSource', 'DrillSource': 'DrillSource', 'CouchbaseSource': 'CouchbaseSource', 'ConcurSource': 'ConcurSource', 'AzurePostgreSqlSource': 'AzurePostgreSqlSource', 'AmazonMWSSource': 'AmazonMWSSource', 'RestServiceSource': 'RestServiceSource', 'HttpSource': 'HttpSource', 'AzureDataLakeStoreSource': 'AzureDataLakeStoreSource', 'MongoDbSource': 'MongoDbSource', 'CassandraSource': 'CassandraSource', 'WebSource': 'WebSource', 'OracleServiceCloudSource': 'OracleServiceCloudSource', 'OracleSource': 'OracleSource', 'AzureMySqlSource': 'AzureMySqlSource', 'HdfsSource': 'HdfsSource', 'FileSystemSource': 'FileSystemSource', 'SqlDWSource': 'SqlDWSource', 'SqlSource': 'SqlSource', 'SapEccSource': 'SapEccSource', 'SapCloudForCustomerSource': 'SapCloudForCustomerSource', 'SalesforceSource': 'SalesforceSource', 'RelationalSource': 'RelationalSource', 'DynamicsAXSource': 'DynamicsAXSource', 'DynamicsSource': 'DynamicsSource', 'DocumentDbCollectionSource': 'DocumentDbCollectionSource', 'BlobSource': 'BlobSource', 'AzureTableSource': 'AzureTableSource'} } def __init__(self, **kwargs): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_source_py3.py index 7c1a96b2897a..b2d6beed13ce 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_source_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_source_py3.py @@ -22,12 +22,13 @@ class CopySource(Model): QuickBooksSource, PrestoSource, PhoenixSource, PaypalSource, MarketoSource, MariaDBSource, MagentoSource, JiraSource, ImpalaSource, HubspotSource, HiveSource, HBaseSource, GreenplumSource, GoogleBigQuerySource, - EloquaSource, DrillSource, CouchbaseSource, ConcurSource, - AzurePostgreSqlSource, AmazonMWSSource, HttpSource, - AzureDataLakeStoreSource, MongoDbSource, CassandraSource, WebSource, - OracleSource, AzureMySqlSource, HdfsSource, FileSystemSource, SqlDWSource, - SqlSource, SapEccSource, SapCloudForCustomerSource, SalesforceSource, - RelationalSource, DynamicsSource, DocumentDbCollectionSource, BlobSource, + GoogleAdWordsSource, EloquaSource, DrillSource, CouchbaseSource, + ConcurSource, AzurePostgreSqlSource, AmazonMWSSource, RestServiceSource, + HttpSource, AzureDataLakeStoreSource, MongoDbSource, CassandraSource, + WebSource, OracleServiceCloudSource, OracleSource, AzureMySqlSource, + HdfsSource, FileSystemSource, SqlDWSource, SqlSource, SapEccSource, + SapCloudForCustomerSource, SalesforceSource, RelationalSource, + DynamicsAXSource, DynamicsSource, DocumentDbCollectionSource, BlobSource, AzureTableSource All required parameters must be populated in order to send to Azure. @@ -58,7 +59,7 @@ class CopySource(Model): } _subtype_map = { - 'type': {'AmazonRedshiftSource': 'AmazonRedshiftSource', 'ResponsysSource': 'ResponsysSource', 'SalesforceMarketingCloudSource': 'SalesforceMarketingCloudSource', 'VerticaSource': 'VerticaSource', 'NetezzaSource': 'NetezzaSource', 'ZohoSource': 'ZohoSource', 'XeroSource': 'XeroSource', 'SquareSource': 'SquareSource', 'SparkSource': 'SparkSource', 'ShopifySource': 'ShopifySource', 'ServiceNowSource': 'ServiceNowSource', 'QuickBooksSource': 'QuickBooksSource', 'PrestoSource': 'PrestoSource', 'PhoenixSource': 'PhoenixSource', 'PaypalSource': 'PaypalSource', 'MarketoSource': 'MarketoSource', 'MariaDBSource': 'MariaDBSource', 'MagentoSource': 'MagentoSource', 'JiraSource': 'JiraSource', 'ImpalaSource': 'ImpalaSource', 'HubspotSource': 'HubspotSource', 'HiveSource': 'HiveSource', 'HBaseSource': 'HBaseSource', 'GreenplumSource': 'GreenplumSource', 'GoogleBigQuerySource': 'GoogleBigQuerySource', 'EloquaSource': 'EloquaSource', 'DrillSource': 'DrillSource', 'CouchbaseSource': 'CouchbaseSource', 'ConcurSource': 'ConcurSource', 'AzurePostgreSqlSource': 'AzurePostgreSqlSource', 'AmazonMWSSource': 'AmazonMWSSource', 'HttpSource': 'HttpSource', 'AzureDataLakeStoreSource': 'AzureDataLakeStoreSource', 'MongoDbSource': 'MongoDbSource', 'CassandraSource': 'CassandraSource', 'WebSource': 'WebSource', 'OracleSource': 'OracleSource', 'AzureMySqlSource': 'AzureMySqlSource', 'HdfsSource': 'HdfsSource', 'FileSystemSource': 'FileSystemSource', 'SqlDWSource': 'SqlDWSource', 'SqlSource': 'SqlSource', 'SapEccSource': 'SapEccSource', 'SapCloudForCustomerSource': 'SapCloudForCustomerSource', 'SalesforceSource': 'SalesforceSource', 'RelationalSource': 'RelationalSource', 'DynamicsSource': 'DynamicsSource', 'DocumentDbCollectionSource': 'DocumentDbCollectionSource', 'BlobSource': 'BlobSource', 'AzureTableSource': 'AzureTableSource'} + 'type': {'AmazonRedshiftSource': 'AmazonRedshiftSource', 'ResponsysSource': 'ResponsysSource', 'SalesforceMarketingCloudSource': 'SalesforceMarketingCloudSource', 'VerticaSource': 'VerticaSource', 'NetezzaSource': 'NetezzaSource', 'ZohoSource': 'ZohoSource', 'XeroSource': 'XeroSource', 'SquareSource': 'SquareSource', 'SparkSource': 'SparkSource', 'ShopifySource': 'ShopifySource', 'ServiceNowSource': 'ServiceNowSource', 'QuickBooksSource': 'QuickBooksSource', 'PrestoSource': 'PrestoSource', 'PhoenixSource': 'PhoenixSource', 'PaypalSource': 'PaypalSource', 'MarketoSource': 'MarketoSource', 'MariaDBSource': 'MariaDBSource', 'MagentoSource': 'MagentoSource', 'JiraSource': 'JiraSource', 'ImpalaSource': 'ImpalaSource', 'HubspotSource': 'HubspotSource', 'HiveSource': 'HiveSource', 'HBaseSource': 'HBaseSource', 'GreenplumSource': 'GreenplumSource', 'GoogleBigQuerySource': 'GoogleBigQuerySource', 'GoogleAdWordsSource': 'GoogleAdWordsSource', 'EloquaSource': 'EloquaSource', 'DrillSource': 'DrillSource', 'CouchbaseSource': 'CouchbaseSource', 'ConcurSource': 'ConcurSource', 'AzurePostgreSqlSource': 'AzurePostgreSqlSource', 'AmazonMWSSource': 'AmazonMWSSource', 'RestServiceSource': 'RestServiceSource', 'HttpSource': 'HttpSource', 'AzureDataLakeStoreSource': 'AzureDataLakeStoreSource', 'MongoDbSource': 'MongoDbSource', 'CassandraSource': 'CassandraSource', 'WebSource': 'WebSource', 'OracleServiceCloudSource': 'OracleServiceCloudSource', 'OracleSource': 'OracleSource', 'AzureMySqlSource': 'AzureMySqlSource', 'HdfsSource': 'HdfsSource', 'FileSystemSource': 'FileSystemSource', 'SqlDWSource': 'SqlDWSource', 'SqlSource': 'SqlSource', 'SapEccSource': 'SapEccSource', 'SapCloudForCustomerSource': 'SapCloudForCustomerSource', 'SalesforceSource': 'SalesforceSource', 'RelationalSource': 'RelationalSource', 'DynamicsAXSource': 'DynamicsAXSource', 'DynamicsSource': 'DynamicsSource', 'DocumentDbCollectionSource': 'DocumentDbCollectionSource', 'BlobSource': 'BlobSource', 'AzureTableSource': 'AzureTableSource'} } def __init__(self, *, additional_properties=None, source_retry_count=None, source_retry_wait=None, **kwargs) -> None: diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_factory_management_client_enums.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_factory_management_client_enums.py index aa76fe05ddc0..2b837286cc12 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_factory_management_client_enums.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_factory_management_client_enums.py @@ -240,6 +240,12 @@ class GoogleBigQueryAuthenticationType(str, Enum): user_authentication = "UserAuthentication" +class GoogleAdWordsAuthenticationType(str, Enum): + + service_authentication = "ServiceAuthentication" + user_authentication = "UserAuthentication" + + class SapHanaAuthenticationType(str, Enum): basic = "Basic" @@ -258,6 +264,14 @@ class FtpAuthenticationType(str, Enum): anonymous = "Anonymous" +class RestServiceAuthenticationType(str, Enum): + + anonymous = "Anonymous" + basic = "Basic" + aad_service_princial = "AadServicePrincial" + managed_service_identity = "ManagedServiceIdentity" + + class HttpAuthenticationType(str, Enum): basic = "Basic" @@ -275,8 +289,17 @@ class MongoDbAuthenticationType(str, Enum): class ODataAuthenticationType(str, Enum): - basic = "Basic" anonymous = "Anonymous" + basic = "Basic" + windows = "Windows" + aad_service_principal = "AadServicePrincipal" + managed_service_identity = "ManagedServiceIdentity" + + +class ODataAadServicePrincipalCredentialType(str, Enum): + + service_principal_key = "ServicePrincipalKey" + service_principal_cert = "ServicePrincipalCert" class TeradataAuthenticationType(str, Enum): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset.py index 626015fd5d11..2cefbb683276 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset.py @@ -26,14 +26,15 @@ class Dataset(Model): MariaDBTableDataset, MagentoObjectDataset, JiraObjectDataset, ImpalaObjectDataset, HubspotObjectDataset, HiveObjectDataset, HBaseObjectDataset, GreenplumTableDataset, GoogleBigQueryObjectDataset, - EloquaObjectDataset, DrillTableDataset, CouchbaseTableDataset, - ConcurObjectDataset, AzurePostgreSqlTableDataset, AmazonMWSObjectDataset, - HttpDataset, AzureSearchIndexDataset, WebTableDataset, - SqlServerTableDataset, SapEccResourceDataset, - SapCloudForCustomerResourceDataset, SalesforceObjectDataset, - RelationalTableDataset, AzureMySqlTableDataset, OracleTableDataset, - ODataResourceDataset, MongoDbCollectionDataset, FileShareDataset, - AzureDataLakeStoreDataset, DynamicsEntityDataset, + GoogleAdWordsObjectDataset, EloquaObjectDataset, DrillTableDataset, + CouchbaseTableDataset, ConcurObjectDataset, AzurePostgreSqlTableDataset, + AmazonMWSObjectDataset, RestServiceDataset, HttpDataset, + AzureSearchIndexDataset, WebTableDataset, SqlServerTableDataset, + SapEccResourceDataset, SapCloudForCustomerResourceDataset, + SalesforceObjectDataset, RelationalTableDataset, AzureMySqlTableDataset, + OracleServiceCloudObjectDataset, OracleTableDataset, ODataResourceDataset, + MongoDbCollectionDataset, FileShareDataset, AzureDataLakeStoreDataset, + DynamicsAXResourceDataset, DynamicsEntityDataset, DocumentDbCollectionDataset, CustomDataset, CassandraTableDataset, AzureSqlDWTableDataset, AzureSqlTableDataset, AzureTableDataset, AzureBlobDataset, AmazonS3Dataset @@ -81,7 +82,7 @@ class Dataset(Model): } _subtype_map = { - 'type': {'ResponsysObject': 'ResponsysObjectDataset', 'SalesforceMarketingCloudObject': 'SalesforceMarketingCloudObjectDataset', 'VerticaTable': 'VerticaTableDataset', 'NetezzaTable': 'NetezzaTableDataset', 'ZohoObject': 'ZohoObjectDataset', 'XeroObject': 'XeroObjectDataset', 'SquareObject': 'SquareObjectDataset', 'SparkObject': 'SparkObjectDataset', 'ShopifyObject': 'ShopifyObjectDataset', 'ServiceNowObject': 'ServiceNowObjectDataset', 'QuickBooksObject': 'QuickBooksObjectDataset', 'PrestoObject': 'PrestoObjectDataset', 'PhoenixObject': 'PhoenixObjectDataset', 'PaypalObject': 'PaypalObjectDataset', 'MarketoObject': 'MarketoObjectDataset', 'MariaDBTable': 'MariaDBTableDataset', 'MagentoObject': 'MagentoObjectDataset', 'JiraObject': 'JiraObjectDataset', 'ImpalaObject': 'ImpalaObjectDataset', 'HubspotObject': 'HubspotObjectDataset', 'HiveObject': 'HiveObjectDataset', 'HBaseObject': 'HBaseObjectDataset', 'GreenplumTable': 'GreenplumTableDataset', 'GoogleBigQueryObject': 'GoogleBigQueryObjectDataset', 'EloquaObject': 'EloquaObjectDataset', 'DrillTable': 'DrillTableDataset', 'CouchbaseTable': 'CouchbaseTableDataset', 'ConcurObject': 'ConcurObjectDataset', 'AzurePostgreSqlTable': 'AzurePostgreSqlTableDataset', 'AmazonMWSObject': 'AmazonMWSObjectDataset', 'HttpFile': 'HttpDataset', 'AzureSearchIndex': 'AzureSearchIndexDataset', 'WebTable': 'WebTableDataset', 'SqlServerTable': 'SqlServerTableDataset', 'SapEccResource': 'SapEccResourceDataset', 'SapCloudForCustomerResource': 'SapCloudForCustomerResourceDataset', 'SalesforceObject': 'SalesforceObjectDataset', 'RelationalTable': 'RelationalTableDataset', 'AzureMySqlTable': 'AzureMySqlTableDataset', 'OracleTable': 'OracleTableDataset', 'ODataResource': 'ODataResourceDataset', 'MongoDbCollection': 'MongoDbCollectionDataset', 'FileShare': 'FileShareDataset', 'AzureDataLakeStoreFile': 'AzureDataLakeStoreDataset', 'DynamicsEntity': 'DynamicsEntityDataset', 'DocumentDbCollection': 'DocumentDbCollectionDataset', 'CustomDataset': 'CustomDataset', 'CassandraTable': 'CassandraTableDataset', 'AzureSqlDWTable': 'AzureSqlDWTableDataset', 'AzureSqlTable': 'AzureSqlTableDataset', 'AzureTable': 'AzureTableDataset', 'AzureBlob': 'AzureBlobDataset', 'AmazonS3Object': 'AmazonS3Dataset'} + 'type': {'ResponsysObject': 'ResponsysObjectDataset', 'SalesforceMarketingCloudObject': 'SalesforceMarketingCloudObjectDataset', 'VerticaTable': 'VerticaTableDataset', 'NetezzaTable': 'NetezzaTableDataset', 'ZohoObject': 'ZohoObjectDataset', 'XeroObject': 'XeroObjectDataset', 'SquareObject': 'SquareObjectDataset', 'SparkObject': 'SparkObjectDataset', 'ShopifyObject': 'ShopifyObjectDataset', 'ServiceNowObject': 'ServiceNowObjectDataset', 'QuickBooksObject': 'QuickBooksObjectDataset', 'PrestoObject': 'PrestoObjectDataset', 'PhoenixObject': 'PhoenixObjectDataset', 'PaypalObject': 'PaypalObjectDataset', 'MarketoObject': 'MarketoObjectDataset', 'MariaDBTable': 'MariaDBTableDataset', 'MagentoObject': 'MagentoObjectDataset', 'JiraObject': 'JiraObjectDataset', 'ImpalaObject': 'ImpalaObjectDataset', 'HubspotObject': 'HubspotObjectDataset', 'HiveObject': 'HiveObjectDataset', 'HBaseObject': 'HBaseObjectDataset', 'GreenplumTable': 'GreenplumTableDataset', 'GoogleBigQueryObject': 'GoogleBigQueryObjectDataset', 'GoogleAdWordsObject': 'GoogleAdWordsObjectDataset', 'EloquaObject': 'EloquaObjectDataset', 'DrillTable': 'DrillTableDataset', 'CouchbaseTable': 'CouchbaseTableDataset', 'ConcurObject': 'ConcurObjectDataset', 'AzurePostgreSqlTable': 'AzurePostgreSqlTableDataset', 'AmazonMWSObject': 'AmazonMWSObjectDataset', 'RestResource': 'RestServiceDataset', 'HttpFile': 'HttpDataset', 'AzureSearchIndex': 'AzureSearchIndexDataset', 'WebTable': 'WebTableDataset', 'SqlServerTable': 'SqlServerTableDataset', 'SapEccResource': 'SapEccResourceDataset', 'SapCloudForCustomerResource': 'SapCloudForCustomerResourceDataset', 'SalesforceObject': 'SalesforceObjectDataset', 'RelationalTable': 'RelationalTableDataset', 'AzureMySqlTable': 'AzureMySqlTableDataset', 'OracleServiceCloudObject': 'OracleServiceCloudObjectDataset', 'OracleTable': 'OracleTableDataset', 'ODataResource': 'ODataResourceDataset', 'MongoDbCollection': 'MongoDbCollectionDataset', 'FileShare': 'FileShareDataset', 'AzureDataLakeStoreFile': 'AzureDataLakeStoreDataset', 'DynamicsAXResource': 'DynamicsAXResourceDataset', 'DynamicsEntity': 'DynamicsEntityDataset', 'DocumentDbCollection': 'DocumentDbCollectionDataset', 'CustomDataset': 'CustomDataset', 'CassandraTable': 'CassandraTableDataset', 'AzureSqlDWTable': 'AzureSqlDWTableDataset', 'AzureSqlTable': 'AzureSqlTableDataset', 'AzureTable': 'AzureTableDataset', 'AzureBlob': 'AzureBlobDataset', 'AmazonS3Object': 'AmazonS3Dataset'} } def __init__(self, **kwargs): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_py3.py index 3d67a1328cd4..5551144b5fa6 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_py3.py @@ -26,14 +26,15 @@ class Dataset(Model): MariaDBTableDataset, MagentoObjectDataset, JiraObjectDataset, ImpalaObjectDataset, HubspotObjectDataset, HiveObjectDataset, HBaseObjectDataset, GreenplumTableDataset, GoogleBigQueryObjectDataset, - EloquaObjectDataset, DrillTableDataset, CouchbaseTableDataset, - ConcurObjectDataset, AzurePostgreSqlTableDataset, AmazonMWSObjectDataset, - HttpDataset, AzureSearchIndexDataset, WebTableDataset, - SqlServerTableDataset, SapEccResourceDataset, - SapCloudForCustomerResourceDataset, SalesforceObjectDataset, - RelationalTableDataset, AzureMySqlTableDataset, OracleTableDataset, - ODataResourceDataset, MongoDbCollectionDataset, FileShareDataset, - AzureDataLakeStoreDataset, DynamicsEntityDataset, + GoogleAdWordsObjectDataset, EloquaObjectDataset, DrillTableDataset, + CouchbaseTableDataset, ConcurObjectDataset, AzurePostgreSqlTableDataset, + AmazonMWSObjectDataset, RestServiceDataset, HttpDataset, + AzureSearchIndexDataset, WebTableDataset, SqlServerTableDataset, + SapEccResourceDataset, SapCloudForCustomerResourceDataset, + SalesforceObjectDataset, RelationalTableDataset, AzureMySqlTableDataset, + OracleServiceCloudObjectDataset, OracleTableDataset, ODataResourceDataset, + MongoDbCollectionDataset, FileShareDataset, AzureDataLakeStoreDataset, + DynamicsAXResourceDataset, DynamicsEntityDataset, DocumentDbCollectionDataset, CustomDataset, CassandraTableDataset, AzureSqlDWTableDataset, AzureSqlTableDataset, AzureTableDataset, AzureBlobDataset, AmazonS3Dataset @@ -81,7 +82,7 @@ class Dataset(Model): } _subtype_map = { - 'type': {'ResponsysObject': 'ResponsysObjectDataset', 'SalesforceMarketingCloudObject': 'SalesforceMarketingCloudObjectDataset', 'VerticaTable': 'VerticaTableDataset', 'NetezzaTable': 'NetezzaTableDataset', 'ZohoObject': 'ZohoObjectDataset', 'XeroObject': 'XeroObjectDataset', 'SquareObject': 'SquareObjectDataset', 'SparkObject': 'SparkObjectDataset', 'ShopifyObject': 'ShopifyObjectDataset', 'ServiceNowObject': 'ServiceNowObjectDataset', 'QuickBooksObject': 'QuickBooksObjectDataset', 'PrestoObject': 'PrestoObjectDataset', 'PhoenixObject': 'PhoenixObjectDataset', 'PaypalObject': 'PaypalObjectDataset', 'MarketoObject': 'MarketoObjectDataset', 'MariaDBTable': 'MariaDBTableDataset', 'MagentoObject': 'MagentoObjectDataset', 'JiraObject': 'JiraObjectDataset', 'ImpalaObject': 'ImpalaObjectDataset', 'HubspotObject': 'HubspotObjectDataset', 'HiveObject': 'HiveObjectDataset', 'HBaseObject': 'HBaseObjectDataset', 'GreenplumTable': 'GreenplumTableDataset', 'GoogleBigQueryObject': 'GoogleBigQueryObjectDataset', 'EloquaObject': 'EloquaObjectDataset', 'DrillTable': 'DrillTableDataset', 'CouchbaseTable': 'CouchbaseTableDataset', 'ConcurObject': 'ConcurObjectDataset', 'AzurePostgreSqlTable': 'AzurePostgreSqlTableDataset', 'AmazonMWSObject': 'AmazonMWSObjectDataset', 'HttpFile': 'HttpDataset', 'AzureSearchIndex': 'AzureSearchIndexDataset', 'WebTable': 'WebTableDataset', 'SqlServerTable': 'SqlServerTableDataset', 'SapEccResource': 'SapEccResourceDataset', 'SapCloudForCustomerResource': 'SapCloudForCustomerResourceDataset', 'SalesforceObject': 'SalesforceObjectDataset', 'RelationalTable': 'RelationalTableDataset', 'AzureMySqlTable': 'AzureMySqlTableDataset', 'OracleTable': 'OracleTableDataset', 'ODataResource': 'ODataResourceDataset', 'MongoDbCollection': 'MongoDbCollectionDataset', 'FileShare': 'FileShareDataset', 'AzureDataLakeStoreFile': 'AzureDataLakeStoreDataset', 'DynamicsEntity': 'DynamicsEntityDataset', 'DocumentDbCollection': 'DocumentDbCollectionDataset', 'CustomDataset': 'CustomDataset', 'CassandraTable': 'CassandraTableDataset', 'AzureSqlDWTable': 'AzureSqlDWTableDataset', 'AzureSqlTable': 'AzureSqlTableDataset', 'AzureTable': 'AzureTableDataset', 'AzureBlob': 'AzureBlobDataset', 'AmazonS3Object': 'AmazonS3Dataset'} + 'type': {'ResponsysObject': 'ResponsysObjectDataset', 'SalesforceMarketingCloudObject': 'SalesforceMarketingCloudObjectDataset', 'VerticaTable': 'VerticaTableDataset', 'NetezzaTable': 'NetezzaTableDataset', 'ZohoObject': 'ZohoObjectDataset', 'XeroObject': 'XeroObjectDataset', 'SquareObject': 'SquareObjectDataset', 'SparkObject': 'SparkObjectDataset', 'ShopifyObject': 'ShopifyObjectDataset', 'ServiceNowObject': 'ServiceNowObjectDataset', 'QuickBooksObject': 'QuickBooksObjectDataset', 'PrestoObject': 'PrestoObjectDataset', 'PhoenixObject': 'PhoenixObjectDataset', 'PaypalObject': 'PaypalObjectDataset', 'MarketoObject': 'MarketoObjectDataset', 'MariaDBTable': 'MariaDBTableDataset', 'MagentoObject': 'MagentoObjectDataset', 'JiraObject': 'JiraObjectDataset', 'ImpalaObject': 'ImpalaObjectDataset', 'HubspotObject': 'HubspotObjectDataset', 'HiveObject': 'HiveObjectDataset', 'HBaseObject': 'HBaseObjectDataset', 'GreenplumTable': 'GreenplumTableDataset', 'GoogleBigQueryObject': 'GoogleBigQueryObjectDataset', 'GoogleAdWordsObject': 'GoogleAdWordsObjectDataset', 'EloquaObject': 'EloquaObjectDataset', 'DrillTable': 'DrillTableDataset', 'CouchbaseTable': 'CouchbaseTableDataset', 'ConcurObject': 'ConcurObjectDataset', 'AzurePostgreSqlTable': 'AzurePostgreSqlTableDataset', 'AmazonMWSObject': 'AmazonMWSObjectDataset', 'RestResource': 'RestServiceDataset', 'HttpFile': 'HttpDataset', 'AzureSearchIndex': 'AzureSearchIndexDataset', 'WebTable': 'WebTableDataset', 'SqlServerTable': 'SqlServerTableDataset', 'SapEccResource': 'SapEccResourceDataset', 'SapCloudForCustomerResource': 'SapCloudForCustomerResourceDataset', 'SalesforceObject': 'SalesforceObjectDataset', 'RelationalTable': 'RelationalTableDataset', 'AzureMySqlTable': 'AzureMySqlTableDataset', 'OracleServiceCloudObject': 'OracleServiceCloudObjectDataset', 'OracleTable': 'OracleTableDataset', 'ODataResource': 'ODataResourceDataset', 'MongoDbCollection': 'MongoDbCollectionDataset', 'FileShare': 'FileShareDataset', 'AzureDataLakeStoreFile': 'AzureDataLakeStoreDataset', 'DynamicsAXResource': 'DynamicsAXResourceDataset', 'DynamicsEntity': 'DynamicsEntityDataset', 'DocumentDbCollection': 'DocumentDbCollectionDataset', 'CustomDataset': 'CustomDataset', 'CassandraTable': 'CassandraTableDataset', 'AzureSqlDWTable': 'AzureSqlDWTableDataset', 'AzureSqlTable': 'AzureSqlTableDataset', 'AzureTable': 'AzureTableDataset', 'AzureBlob': 'AzureBlobDataset', 'AmazonS3Object': 'AmazonS3Dataset'} } def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_ax_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_ax_linked_service.py new file mode 100644 index 000000000000..f28a5071f400 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_ax_linked_service.py @@ -0,0 +1,90 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class DynamicsAXLinkedService(LinkedService): + """Dynamics AX linked service. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param url: Required. The URL of Dynamics AX OData API. Type: string (or + Expression with resultType string). + :type url: object + :param tenant: Specify the tenant information (domain name or tenant ID) + under which your application resides. Type: string (or Expression with + resultType string). + :type tenant: object + :param service_principal_id: Specify the application id of your + application registered in Azure Active Directory. Type: string (or + Expression with resultType string). + :type service_principal_id: object + :param aad_resource_id: Specify the resource you are requesting + authorization to use Directory. Type: string (or Expression with + resultType string). + :type aad_resource_id: object + :param service_principal_key: Specify the secret of your application + registered in Azure Active Directory. Type: string (or Expression with + resultType string). + :type service_principal_key: ~azure.mgmt.datafactory.models.SecretBase + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Either encryptedCredential or username/password must + be provided. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'url': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'url': {'key': 'typeProperties.url', 'type': 'object'}, + 'tenant': {'key': 'typeProperties.tenant', 'type': 'object'}, + 'service_principal_id': {'key': 'typeProperties.servicePrincipalId', 'type': 'object'}, + 'aad_resource_id': {'key': 'typeProperties.aadResourceId', 'type': 'object'}, + 'service_principal_key': {'key': 'typeProperties.servicePrincipalKey', 'type': 'SecretBase'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, **kwargs): + super(DynamicsAXLinkedService, self).__init__(**kwargs) + self.url = kwargs.get('url', None) + self.tenant = kwargs.get('tenant', None) + self.service_principal_id = kwargs.get('service_principal_id', None) + self.aad_resource_id = kwargs.get('aad_resource_id', None) + self.service_principal_key = kwargs.get('service_principal_key', None) + self.encrypted_credential = kwargs.get('encrypted_credential', None) + self.type = 'DynamicsAX' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_ax_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_ax_linked_service_py3.py new file mode 100644 index 000000000000..3875023c438e --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_ax_linked_service_py3.py @@ -0,0 +1,90 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service_py3 import LinkedService + + +class DynamicsAXLinkedService(LinkedService): + """Dynamics AX linked service. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param url: Required. The URL of Dynamics AX OData API. Type: string (or + Expression with resultType string). + :type url: object + :param tenant: Specify the tenant information (domain name or tenant ID) + under which your application resides. Type: string (or Expression with + resultType string). + :type tenant: object + :param service_principal_id: Specify the application id of your + application registered in Azure Active Directory. Type: string (or + Expression with resultType string). + :type service_principal_id: object + :param aad_resource_id: Specify the resource you are requesting + authorization to use Directory. Type: string (or Expression with + resultType string). + :type aad_resource_id: object + :param service_principal_key: Specify the secret of your application + registered in Azure Active Directory. Type: string (or Expression with + resultType string). + :type service_principal_key: ~azure.mgmt.datafactory.models.SecretBase + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Either encryptedCredential or username/password must + be provided. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'url': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'url': {'key': 'typeProperties.url', 'type': 'object'}, + 'tenant': {'key': 'typeProperties.tenant', 'type': 'object'}, + 'service_principal_id': {'key': 'typeProperties.servicePrincipalId', 'type': 'object'}, + 'aad_resource_id': {'key': 'typeProperties.aadResourceId', 'type': 'object'}, + 'service_principal_key': {'key': 'typeProperties.servicePrincipalKey', 'type': 'SecretBase'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, *, url, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, tenant=None, service_principal_id=None, aad_resource_id=None, service_principal_key=None, encrypted_credential=None, **kwargs) -> None: + super(DynamicsAXLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.url = url + self.tenant = tenant + self.service_principal_id = service_principal_id + self.aad_resource_id = aad_resource_id + self.service_principal_key = service_principal_key + self.encrypted_credential = encrypted_credential + self.type = 'DynamicsAX' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_ax_resource_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_ax_resource_dataset.py new file mode 100644 index 000000000000..2603fb97543c --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_ax_resource_dataset.py @@ -0,0 +1,68 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class DynamicsAXResourceDataset(Dataset): + """The path of the Dynamics AX OData entity. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Required. Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder + :param type: Required. Constant filled by server. + :type type: str + :param path: Required. The path of the Dynamics AX OData entity. Type: + string (or Expression with resultType string). + :type path: str + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + 'path': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, + 'type': {'key': 'type', 'type': 'str'}, + 'path': {'key': 'typeProperties.path', 'type': 'str'}, + } + + def __init__(self, **kwargs): + super(DynamicsAXResourceDataset, self).__init__(**kwargs) + self.path = kwargs.get('path', None) + self.type = 'DynamicsAXResource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_ax_resource_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_ax_resource_dataset_py3.py new file mode 100644 index 000000000000..9cd8602941e7 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_ax_resource_dataset_py3.py @@ -0,0 +1,68 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset_py3 import Dataset + + +class DynamicsAXResourceDataset(Dataset): + """The path of the Dynamics AX OData entity. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Required. Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder + :param type: Required. Constant filled by server. + :type type: str + :param path: Required. The path of the Dynamics AX OData entity. Type: + string (or Expression with resultType string). + :type path: str + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + 'path': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, + 'type': {'key': 'type', 'type': 'str'}, + 'path': {'key': 'typeProperties.path', 'type': 'str'}, + } + + def __init__(self, *, linked_service_name, path: str, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + super(DynamicsAXResourceDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + self.path = path + self.type = 'DynamicsAXResource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_ax_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_ax_source.py new file mode 100644 index 000000000000..60ef05c0bd0c --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_ax_source.py @@ -0,0 +1,52 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class DynamicsAXSource(CopySource): + """A copy activity Dynamics AX source. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Required. Constant filled by server. + :type type: str + :param query: A query to retrieve data from source. Type: string (or + Expression with resultType string). + :type query: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'query': {'key': 'query', 'type': 'object'}, + } + + def __init__(self, **kwargs): + super(DynamicsAXSource, self).__init__(**kwargs) + self.query = kwargs.get('query', None) + self.type = 'DynamicsAXSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_ax_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_ax_source_py3.py new file mode 100644 index 000000000000..0aa064facbf2 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_ax_source_py3.py @@ -0,0 +1,52 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source_py3 import CopySource + + +class DynamicsAXSource(CopySource): + """A copy activity Dynamics AX source. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Required. Constant filled by server. + :type type: str + :param query: A query to retrieve data from source. Type: string (or + Expression with resultType string). + :type query: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'query': {'key': 'query', 'type': 'object'}, + } + + def __init__(self, *, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None, **kwargs) -> None: + super(DynamicsAXSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, **kwargs) + self.query = query + self.type = 'DynamicsAXSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_share_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_share_dataset.py index d5cbf4ac6822..f28535bb54d3 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_share_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_share_dataset.py @@ -45,6 +45,10 @@ class FileShareDataset(Dataset): :param file_name: The name of the on-premises file system. Type: string (or Expression with resultType string). :type file_name: object + :param wildcard_path: The whole path include file name of the on-premises + file system with wildcard supported. Type: string (or Expression with + resultType string). + :type wildcard_path: object :param format: The format of the files. :type format: ~azure.mgmt.datafactory.models.DatasetStorageFormat :param file_filter: Specify a filter to be used to select a subset of @@ -71,6 +75,7 @@ class FileShareDataset(Dataset): 'type': {'key': 'type', 'type': 'str'}, 'folder_path': {'key': 'typeProperties.folderPath', 'type': 'object'}, 'file_name': {'key': 'typeProperties.fileName', 'type': 'object'}, + 'wildcard_path': {'key': 'typeProperties.wildcardPath', 'type': 'object'}, 'format': {'key': 'typeProperties.format', 'type': 'DatasetStorageFormat'}, 'file_filter': {'key': 'typeProperties.fileFilter', 'type': 'object'}, 'compression': {'key': 'typeProperties.compression', 'type': 'DatasetCompression'}, @@ -80,6 +85,7 @@ def __init__(self, **kwargs): super(FileShareDataset, self).__init__(**kwargs) self.folder_path = kwargs.get('folder_path', None) self.file_name = kwargs.get('file_name', None) + self.wildcard_path = kwargs.get('wildcard_path', None) self.format = kwargs.get('format', None) self.file_filter = kwargs.get('file_filter', None) self.compression = kwargs.get('compression', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_share_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_share_dataset_py3.py index fe980181543e..ae96956c9290 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_share_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_share_dataset_py3.py @@ -45,6 +45,10 @@ class FileShareDataset(Dataset): :param file_name: The name of the on-premises file system. Type: string (or Expression with resultType string). :type file_name: object + :param wildcard_path: The whole path include file name of the on-premises + file system with wildcard supported. Type: string (or Expression with + resultType string). + :type wildcard_path: object :param format: The format of the files. :type format: ~azure.mgmt.datafactory.models.DatasetStorageFormat :param file_filter: Specify a filter to be used to select a subset of @@ -71,15 +75,17 @@ class FileShareDataset(Dataset): 'type': {'key': 'type', 'type': 'str'}, 'folder_path': {'key': 'typeProperties.folderPath', 'type': 'object'}, 'file_name': {'key': 'typeProperties.fileName', 'type': 'object'}, + 'wildcard_path': {'key': 'typeProperties.wildcardPath', 'type': 'object'}, 'format': {'key': 'typeProperties.format', 'type': 'DatasetStorageFormat'}, 'file_filter': {'key': 'typeProperties.fileFilter', 'type': 'object'}, 'compression': {'key': 'typeProperties.compression', 'type': 'DatasetCompression'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, folder_path=None, file_name=None, format=None, file_filter=None, compression=None, **kwargs) -> None: + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, folder_path=None, file_name=None, wildcard_path=None, format=None, file_filter=None, compression=None, **kwargs) -> None: super(FileShareDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.folder_path = folder_path self.file_name = file_name + self.wildcard_path = wildcard_path self.format = format self.file_filter = file_filter self.compression = compression diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_ad_words_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_ad_words_linked_service.py new file mode 100644 index 000000000000..9f36df58d576 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_ad_words_linked_service.py @@ -0,0 +1,119 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class GoogleAdWordsLinkedService(LinkedService): + """Google Awords service linked service. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param client_customer_id: Required. The Client customer ID of the AdWords + account that you want to fetch report data for. + :type client_customer_id: object + :param developer_token: Required. The developer token associated with the + manager account that you use to grant access to the AdWords API. + :type developer_token: ~azure.mgmt.datafactory.models.SecretBase + :param authentication_type: Required. The OAuth 2.0 authentication + mechanism used for authentication. ServiceAuthentication can only be used + on self-hosted IR. Possible values include: 'ServiceAuthentication', + 'UserAuthentication' + :type authentication_type: str or + ~azure.mgmt.datafactory.models.GoogleAdWordsAuthenticationType + :param refresh_token: The refresh token obtained from Google for + authorizing access to AdWords for UserAuthentication. + :type refresh_token: ~azure.mgmt.datafactory.models.SecretBase + :param client_id: The client id of the google application used to acquire + the refresh token. + :type client_id: ~azure.mgmt.datafactory.models.SecretBase + :param client_secret: The client secret of the google application used to + acquire the refresh token. + :type client_secret: ~azure.mgmt.datafactory.models.SecretBase + :param email: The service account email ID that is used for + ServiceAuthentication and can only be used on self-hosted IR. + :type email: object + :param key_file_path: The full path to the .p12 key file that is used to + authenticate the service account email address and can only be used on + self-hosted IR. + :type key_file_path: object + :param trusted_cert_path: The full path of the .pem file containing + trusted CA certificates for verifying the server when connecting over SSL. + This property can only be set when using SSL on self-hosted IR. The + default value is the cacerts.pem file installed with the IR. + :type trusted_cert_path: object + :param use_system_trust_store: Specifies whether to use a CA certificate + from the system trust store or from a specified PEM file. The default + value is false. + :type use_system_trust_store: object + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'client_customer_id': {'required': True}, + 'developer_token': {'required': True}, + 'authentication_type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'client_customer_id': {'key': 'typeProperties.clientCustomerID', 'type': 'object'}, + 'developer_token': {'key': 'typeProperties.developerToken', 'type': 'SecretBase'}, + 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, + 'refresh_token': {'key': 'typeProperties.refreshToken', 'type': 'SecretBase'}, + 'client_id': {'key': 'typeProperties.clientId', 'type': 'SecretBase'}, + 'client_secret': {'key': 'typeProperties.clientSecret', 'type': 'SecretBase'}, + 'email': {'key': 'typeProperties.email', 'type': 'object'}, + 'key_file_path': {'key': 'typeProperties.keyFilePath', 'type': 'object'}, + 'trusted_cert_path': {'key': 'typeProperties.trustedCertPath', 'type': 'object'}, + 'use_system_trust_store': {'key': 'typeProperties.useSystemTrustStore', 'type': 'object'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, **kwargs): + super(GoogleAdWordsLinkedService, self).__init__(**kwargs) + self.client_customer_id = kwargs.get('client_customer_id', None) + self.developer_token = kwargs.get('developer_token', None) + self.authentication_type = kwargs.get('authentication_type', None) + self.refresh_token = kwargs.get('refresh_token', None) + self.client_id = kwargs.get('client_id', None) + self.client_secret = kwargs.get('client_secret', None) + self.email = kwargs.get('email', None) + self.key_file_path = kwargs.get('key_file_path', None) + self.trusted_cert_path = kwargs.get('trusted_cert_path', None) + self.use_system_trust_store = kwargs.get('use_system_trust_store', None) + self.encrypted_credential = kwargs.get('encrypted_credential', None) + self.type = 'GoogleAdWords' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_ad_words_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_ad_words_linked_service_py3.py new file mode 100644 index 000000000000..ff5ffe6df381 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_ad_words_linked_service_py3.py @@ -0,0 +1,119 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service_py3 import LinkedService + + +class GoogleAdWordsLinkedService(LinkedService): + """Google Awords service linked service. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param client_customer_id: Required. The Client customer ID of the AdWords + account that you want to fetch report data for. + :type client_customer_id: object + :param developer_token: Required. The developer token associated with the + manager account that you use to grant access to the AdWords API. + :type developer_token: ~azure.mgmt.datafactory.models.SecretBase + :param authentication_type: Required. The OAuth 2.0 authentication + mechanism used for authentication. ServiceAuthentication can only be used + on self-hosted IR. Possible values include: 'ServiceAuthentication', + 'UserAuthentication' + :type authentication_type: str or + ~azure.mgmt.datafactory.models.GoogleAdWordsAuthenticationType + :param refresh_token: The refresh token obtained from Google for + authorizing access to AdWords for UserAuthentication. + :type refresh_token: ~azure.mgmt.datafactory.models.SecretBase + :param client_id: The client id of the google application used to acquire + the refresh token. + :type client_id: ~azure.mgmt.datafactory.models.SecretBase + :param client_secret: The client secret of the google application used to + acquire the refresh token. + :type client_secret: ~azure.mgmt.datafactory.models.SecretBase + :param email: The service account email ID that is used for + ServiceAuthentication and can only be used on self-hosted IR. + :type email: object + :param key_file_path: The full path to the .p12 key file that is used to + authenticate the service account email address and can only be used on + self-hosted IR. + :type key_file_path: object + :param trusted_cert_path: The full path of the .pem file containing + trusted CA certificates for verifying the server when connecting over SSL. + This property can only be set when using SSL on self-hosted IR. The + default value is the cacerts.pem file installed with the IR. + :type trusted_cert_path: object + :param use_system_trust_store: Specifies whether to use a CA certificate + from the system trust store or from a specified PEM file. The default + value is false. + :type use_system_trust_store: object + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'client_customer_id': {'required': True}, + 'developer_token': {'required': True}, + 'authentication_type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'client_customer_id': {'key': 'typeProperties.clientCustomerID', 'type': 'object'}, + 'developer_token': {'key': 'typeProperties.developerToken', 'type': 'SecretBase'}, + 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, + 'refresh_token': {'key': 'typeProperties.refreshToken', 'type': 'SecretBase'}, + 'client_id': {'key': 'typeProperties.clientId', 'type': 'SecretBase'}, + 'client_secret': {'key': 'typeProperties.clientSecret', 'type': 'SecretBase'}, + 'email': {'key': 'typeProperties.email', 'type': 'object'}, + 'key_file_path': {'key': 'typeProperties.keyFilePath', 'type': 'object'}, + 'trusted_cert_path': {'key': 'typeProperties.trustedCertPath', 'type': 'object'}, + 'use_system_trust_store': {'key': 'typeProperties.useSystemTrustStore', 'type': 'object'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, *, client_customer_id, developer_token, authentication_type, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, refresh_token=None, client_id=None, client_secret=None, email=None, key_file_path=None, trusted_cert_path=None, use_system_trust_store=None, encrypted_credential=None, **kwargs) -> None: + super(GoogleAdWordsLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.client_customer_id = client_customer_id + self.developer_token = developer_token + self.authentication_type = authentication_type + self.refresh_token = refresh_token + self.client_id = client_id + self.client_secret = client_secret + self.email = email + self.key_file_path = key_file_path + self.trusted_cert_path = trusted_cert_path + self.use_system_trust_store = use_system_trust_store + self.encrypted_credential = encrypted_credential + self.type = 'GoogleAdWords' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_ad_words_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_ad_words_object_dataset.py new file mode 100644 index 000000000000..0daede4e239a --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_ad_words_object_dataset.py @@ -0,0 +1,62 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class GoogleAdWordsObjectDataset(Dataset): + """Google AdWords service dataset. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Required. Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder + :param type: Required. Constant filled by server. + :type type: str + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, **kwargs): + super(GoogleAdWordsObjectDataset, self).__init__(**kwargs) + self.type = 'GoogleAdWordsObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_ad_words_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_ad_words_object_dataset_py3.py new file mode 100644 index 000000000000..7c3fb4379041 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_ad_words_object_dataset_py3.py @@ -0,0 +1,62 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset_py3 import Dataset + + +class GoogleAdWordsObjectDataset(Dataset): + """Google AdWords service dataset. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Required. Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder + :param type: Required. Constant filled by server. + :type type: str + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + super(GoogleAdWordsObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + self.type = 'GoogleAdWordsObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_ad_words_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_ad_words_source.py new file mode 100644 index 000000000000..ccd21de4a364 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_ad_words_source.py @@ -0,0 +1,52 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class GoogleAdWordsSource(CopySource): + """A copy activity Google AdWords service source. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Required. Constant filled by server. + :type type: str + :param query: A query to retrieve data from source. Type: string (or + Expression with resultType string). + :type query: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'query': {'key': 'query', 'type': 'object'}, + } + + def __init__(self, **kwargs): + super(GoogleAdWordsSource, self).__init__(**kwargs) + self.query = kwargs.get('query', None) + self.type = 'GoogleAdWordsSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_ad_words_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_ad_words_source_py3.py new file mode 100644 index 000000000000..6e421dddc45f --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_ad_words_source_py3.py @@ -0,0 +1,52 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source_py3 import CopySource + + +class GoogleAdWordsSource(CopySource): + """A copy activity Google AdWords service source. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Required. Constant filled by server. + :type type: str + :param query: A query to retrieve data from source. Type: string (or + Expression with resultType string). + :type query: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'query': {'key': 'query', 'type': 'object'}, + } + + def __init__(self, *, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None, **kwargs) -> None: + super(GoogleAdWordsSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, **kwargs) + self.query = query + self.type = 'GoogleAdWordsSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service.py index 3dec3c0dcc24..c74f1320eb95 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service.py @@ -28,19 +28,20 @@ class LinkedService(Model): MariaDBLinkedService, MagentoLinkedService, JiraLinkedService, ImpalaLinkedService, HubspotLinkedService, HiveLinkedService, HBaseLinkedService, GreenplumLinkedService, GoogleBigQueryLinkedService, - EloquaLinkedService, DrillLinkedService, CouchbaseLinkedService, - ConcurLinkedService, AzurePostgreSqlLinkedService, AmazonMWSLinkedService, - SapHanaLinkedService, SapBWLinkedService, SftpServerLinkedService, - FtpServerLinkedService, HttpLinkedService, AzureSearchLinkedService, - CustomDataSourceLinkedService, AmazonRedshiftLinkedService, - AmazonS3LinkedService, SapEccLinkedService, + GoogleAdWordsLinkedService, EloquaLinkedService, DrillLinkedService, + CouchbaseLinkedService, ConcurLinkedService, AzurePostgreSqlLinkedService, + AmazonMWSLinkedService, SapHanaLinkedService, SapBWLinkedService, + SftpServerLinkedService, FtpServerLinkedService, RestServiceLinkedService, + HttpLinkedService, AzureSearchLinkedService, CustomDataSourceLinkedService, + AmazonRedshiftLinkedService, AmazonS3LinkedService, SapEccLinkedService, SapCloudForCustomerLinkedService, SalesforceLinkedService, AzureDataLakeStoreLinkedService, MongoDbLinkedService, CassandraLinkedService, WebLinkedService, ODataLinkedService, HdfsLinkedService, OdbcLinkedService, AzureMLLinkedService, TeradataLinkedService, Db2LinkedService, SybaseLinkedService, PostgreSqlLinkedService, MySqlLinkedService, AzureMySqlLinkedService, - OracleLinkedService, FileServerLinkedService, HDInsightLinkedService, + OracleServiceCloudLinkedService, OracleLinkedService, + FileServerLinkedService, HDInsightLinkedService, DynamicsAXLinkedService, DynamicsLinkedService, CosmosDbLinkedService, AzureKeyVaultLinkedService, AzureBatchLinkedService, AzureSqlDatabaseLinkedService, SqlServerLinkedService, AzureSqlDWLinkedService, @@ -81,7 +82,7 @@ class LinkedService(Model): } _subtype_map = { - 'type': {'Responsys': 'ResponsysLinkedService', 'AzureDatabricks': 'AzureDatabricksLinkedService', 'AzureDataLakeAnalytics': 'AzureDataLakeAnalyticsLinkedService', 'HDInsightOnDemand': 'HDInsightOnDemandLinkedService', 'SalesforceMarketingCloud': 'SalesforceMarketingCloudLinkedService', 'Netezza': 'NetezzaLinkedService', 'Vertica': 'VerticaLinkedService', 'Zoho': 'ZohoLinkedService', 'Xero': 'XeroLinkedService', 'Square': 'SquareLinkedService', 'Spark': 'SparkLinkedService', 'Shopify': 'ShopifyLinkedService', 'ServiceNow': 'ServiceNowLinkedService', 'QuickBooks': 'QuickBooksLinkedService', 'Presto': 'PrestoLinkedService', 'Phoenix': 'PhoenixLinkedService', 'Paypal': 'PaypalLinkedService', 'Marketo': 'MarketoLinkedService', 'MariaDB': 'MariaDBLinkedService', 'Magento': 'MagentoLinkedService', 'Jira': 'JiraLinkedService', 'Impala': 'ImpalaLinkedService', 'Hubspot': 'HubspotLinkedService', 'Hive': 'HiveLinkedService', 'HBase': 'HBaseLinkedService', 'Greenplum': 'GreenplumLinkedService', 'GoogleBigQuery': 'GoogleBigQueryLinkedService', 'Eloqua': 'EloquaLinkedService', 'Drill': 'DrillLinkedService', 'Couchbase': 'CouchbaseLinkedService', 'Concur': 'ConcurLinkedService', 'AzurePostgreSql': 'AzurePostgreSqlLinkedService', 'AmazonMWS': 'AmazonMWSLinkedService', 'SapHana': 'SapHanaLinkedService', 'SapBW': 'SapBWLinkedService', 'Sftp': 'SftpServerLinkedService', 'FtpServer': 'FtpServerLinkedService', 'HttpServer': 'HttpLinkedService', 'AzureSearch': 'AzureSearchLinkedService', 'CustomDataSource': 'CustomDataSourceLinkedService', 'AmazonRedshift': 'AmazonRedshiftLinkedService', 'AmazonS3': 'AmazonS3LinkedService', 'SapEcc': 'SapEccLinkedService', 'SapCloudForCustomer': 'SapCloudForCustomerLinkedService', 'Salesforce': 'SalesforceLinkedService', 'AzureDataLakeStore': 'AzureDataLakeStoreLinkedService', 'MongoDb': 'MongoDbLinkedService', 'Cassandra': 'CassandraLinkedService', 'Web': 'WebLinkedService', 'OData': 'ODataLinkedService', 'Hdfs': 'HdfsLinkedService', 'Odbc': 'OdbcLinkedService', 'AzureML': 'AzureMLLinkedService', 'Teradata': 'TeradataLinkedService', 'Db2': 'Db2LinkedService', 'Sybase': 'SybaseLinkedService', 'PostgreSql': 'PostgreSqlLinkedService', 'MySql': 'MySqlLinkedService', 'AzureMySql': 'AzureMySqlLinkedService', 'Oracle': 'OracleLinkedService', 'FileServer': 'FileServerLinkedService', 'HDInsight': 'HDInsightLinkedService', 'Dynamics': 'DynamicsLinkedService', 'CosmosDb': 'CosmosDbLinkedService', 'AzureKeyVault': 'AzureKeyVaultLinkedService', 'AzureBatch': 'AzureBatchLinkedService', 'AzureSqlDatabase': 'AzureSqlDatabaseLinkedService', 'SqlServer': 'SqlServerLinkedService', 'AzureSqlDW': 'AzureSqlDWLinkedService', 'AzureTableStorage': 'AzureTableStorageLinkedService', 'AzureBlobStorage': 'AzureBlobStorageLinkedService', 'AzureStorage': 'AzureStorageLinkedService'} + 'type': {'Responsys': 'ResponsysLinkedService', 'AzureDatabricks': 'AzureDatabricksLinkedService', 'AzureDataLakeAnalytics': 'AzureDataLakeAnalyticsLinkedService', 'HDInsightOnDemand': 'HDInsightOnDemandLinkedService', 'SalesforceMarketingCloud': 'SalesforceMarketingCloudLinkedService', 'Netezza': 'NetezzaLinkedService', 'Vertica': 'VerticaLinkedService', 'Zoho': 'ZohoLinkedService', 'Xero': 'XeroLinkedService', 'Square': 'SquareLinkedService', 'Spark': 'SparkLinkedService', 'Shopify': 'ShopifyLinkedService', 'ServiceNow': 'ServiceNowLinkedService', 'QuickBooks': 'QuickBooksLinkedService', 'Presto': 'PrestoLinkedService', 'Phoenix': 'PhoenixLinkedService', 'Paypal': 'PaypalLinkedService', 'Marketo': 'MarketoLinkedService', 'MariaDB': 'MariaDBLinkedService', 'Magento': 'MagentoLinkedService', 'Jira': 'JiraLinkedService', 'Impala': 'ImpalaLinkedService', 'Hubspot': 'HubspotLinkedService', 'Hive': 'HiveLinkedService', 'HBase': 'HBaseLinkedService', 'Greenplum': 'GreenplumLinkedService', 'GoogleBigQuery': 'GoogleBigQueryLinkedService', 'GoogleAdWords': 'GoogleAdWordsLinkedService', 'Eloqua': 'EloquaLinkedService', 'Drill': 'DrillLinkedService', 'Couchbase': 'CouchbaseLinkedService', 'Concur': 'ConcurLinkedService', 'AzurePostgreSql': 'AzurePostgreSqlLinkedService', 'AmazonMWS': 'AmazonMWSLinkedService', 'SapHana': 'SapHanaLinkedService', 'SapBW': 'SapBWLinkedService', 'Sftp': 'SftpServerLinkedService', 'FtpServer': 'FtpServerLinkedService', 'RestService': 'RestServiceLinkedService', 'HttpServer': 'HttpLinkedService', 'AzureSearch': 'AzureSearchLinkedService', 'CustomDataSource': 'CustomDataSourceLinkedService', 'AmazonRedshift': 'AmazonRedshiftLinkedService', 'AmazonS3': 'AmazonS3LinkedService', 'SapEcc': 'SapEccLinkedService', 'SapCloudForCustomer': 'SapCloudForCustomerLinkedService', 'Salesforce': 'SalesforceLinkedService', 'AzureDataLakeStore': 'AzureDataLakeStoreLinkedService', 'MongoDb': 'MongoDbLinkedService', 'Cassandra': 'CassandraLinkedService', 'Web': 'WebLinkedService', 'OData': 'ODataLinkedService', 'Hdfs': 'HdfsLinkedService', 'Odbc': 'OdbcLinkedService', 'AzureML': 'AzureMLLinkedService', 'Teradata': 'TeradataLinkedService', 'Db2': 'Db2LinkedService', 'Sybase': 'SybaseLinkedService', 'PostgreSql': 'PostgreSqlLinkedService', 'MySql': 'MySqlLinkedService', 'AzureMySql': 'AzureMySqlLinkedService', 'OracleServiceCloud': 'OracleServiceCloudLinkedService', 'Oracle': 'OracleLinkedService', 'FileServer': 'FileServerLinkedService', 'HDInsight': 'HDInsightLinkedService', 'DynamicsAX': 'DynamicsAXLinkedService', 'Dynamics': 'DynamicsLinkedService', 'CosmosDb': 'CosmosDbLinkedService', 'AzureKeyVault': 'AzureKeyVaultLinkedService', 'AzureBatch': 'AzureBatchLinkedService', 'AzureSqlDatabase': 'AzureSqlDatabaseLinkedService', 'SqlServer': 'SqlServerLinkedService', 'AzureSqlDW': 'AzureSqlDWLinkedService', 'AzureTableStorage': 'AzureTableStorageLinkedService', 'AzureBlobStorage': 'AzureBlobStorageLinkedService', 'AzureStorage': 'AzureStorageLinkedService'} } def __init__(self, **kwargs): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service_py3.py index 885af520bcda..ba49c525e830 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service_py3.py @@ -28,19 +28,20 @@ class LinkedService(Model): MariaDBLinkedService, MagentoLinkedService, JiraLinkedService, ImpalaLinkedService, HubspotLinkedService, HiveLinkedService, HBaseLinkedService, GreenplumLinkedService, GoogleBigQueryLinkedService, - EloquaLinkedService, DrillLinkedService, CouchbaseLinkedService, - ConcurLinkedService, AzurePostgreSqlLinkedService, AmazonMWSLinkedService, - SapHanaLinkedService, SapBWLinkedService, SftpServerLinkedService, - FtpServerLinkedService, HttpLinkedService, AzureSearchLinkedService, - CustomDataSourceLinkedService, AmazonRedshiftLinkedService, - AmazonS3LinkedService, SapEccLinkedService, + GoogleAdWordsLinkedService, EloquaLinkedService, DrillLinkedService, + CouchbaseLinkedService, ConcurLinkedService, AzurePostgreSqlLinkedService, + AmazonMWSLinkedService, SapHanaLinkedService, SapBWLinkedService, + SftpServerLinkedService, FtpServerLinkedService, RestServiceLinkedService, + HttpLinkedService, AzureSearchLinkedService, CustomDataSourceLinkedService, + AmazonRedshiftLinkedService, AmazonS3LinkedService, SapEccLinkedService, SapCloudForCustomerLinkedService, SalesforceLinkedService, AzureDataLakeStoreLinkedService, MongoDbLinkedService, CassandraLinkedService, WebLinkedService, ODataLinkedService, HdfsLinkedService, OdbcLinkedService, AzureMLLinkedService, TeradataLinkedService, Db2LinkedService, SybaseLinkedService, PostgreSqlLinkedService, MySqlLinkedService, AzureMySqlLinkedService, - OracleLinkedService, FileServerLinkedService, HDInsightLinkedService, + OracleServiceCloudLinkedService, OracleLinkedService, + FileServerLinkedService, HDInsightLinkedService, DynamicsAXLinkedService, DynamicsLinkedService, CosmosDbLinkedService, AzureKeyVaultLinkedService, AzureBatchLinkedService, AzureSqlDatabaseLinkedService, SqlServerLinkedService, AzureSqlDWLinkedService, @@ -81,7 +82,7 @@ class LinkedService(Model): } _subtype_map = { - 'type': {'Responsys': 'ResponsysLinkedService', 'AzureDatabricks': 'AzureDatabricksLinkedService', 'AzureDataLakeAnalytics': 'AzureDataLakeAnalyticsLinkedService', 'HDInsightOnDemand': 'HDInsightOnDemandLinkedService', 'SalesforceMarketingCloud': 'SalesforceMarketingCloudLinkedService', 'Netezza': 'NetezzaLinkedService', 'Vertica': 'VerticaLinkedService', 'Zoho': 'ZohoLinkedService', 'Xero': 'XeroLinkedService', 'Square': 'SquareLinkedService', 'Spark': 'SparkLinkedService', 'Shopify': 'ShopifyLinkedService', 'ServiceNow': 'ServiceNowLinkedService', 'QuickBooks': 'QuickBooksLinkedService', 'Presto': 'PrestoLinkedService', 'Phoenix': 'PhoenixLinkedService', 'Paypal': 'PaypalLinkedService', 'Marketo': 'MarketoLinkedService', 'MariaDB': 'MariaDBLinkedService', 'Magento': 'MagentoLinkedService', 'Jira': 'JiraLinkedService', 'Impala': 'ImpalaLinkedService', 'Hubspot': 'HubspotLinkedService', 'Hive': 'HiveLinkedService', 'HBase': 'HBaseLinkedService', 'Greenplum': 'GreenplumLinkedService', 'GoogleBigQuery': 'GoogleBigQueryLinkedService', 'Eloqua': 'EloquaLinkedService', 'Drill': 'DrillLinkedService', 'Couchbase': 'CouchbaseLinkedService', 'Concur': 'ConcurLinkedService', 'AzurePostgreSql': 'AzurePostgreSqlLinkedService', 'AmazonMWS': 'AmazonMWSLinkedService', 'SapHana': 'SapHanaLinkedService', 'SapBW': 'SapBWLinkedService', 'Sftp': 'SftpServerLinkedService', 'FtpServer': 'FtpServerLinkedService', 'HttpServer': 'HttpLinkedService', 'AzureSearch': 'AzureSearchLinkedService', 'CustomDataSource': 'CustomDataSourceLinkedService', 'AmazonRedshift': 'AmazonRedshiftLinkedService', 'AmazonS3': 'AmazonS3LinkedService', 'SapEcc': 'SapEccLinkedService', 'SapCloudForCustomer': 'SapCloudForCustomerLinkedService', 'Salesforce': 'SalesforceLinkedService', 'AzureDataLakeStore': 'AzureDataLakeStoreLinkedService', 'MongoDb': 'MongoDbLinkedService', 'Cassandra': 'CassandraLinkedService', 'Web': 'WebLinkedService', 'OData': 'ODataLinkedService', 'Hdfs': 'HdfsLinkedService', 'Odbc': 'OdbcLinkedService', 'AzureML': 'AzureMLLinkedService', 'Teradata': 'TeradataLinkedService', 'Db2': 'Db2LinkedService', 'Sybase': 'SybaseLinkedService', 'PostgreSql': 'PostgreSqlLinkedService', 'MySql': 'MySqlLinkedService', 'AzureMySql': 'AzureMySqlLinkedService', 'Oracle': 'OracleLinkedService', 'FileServer': 'FileServerLinkedService', 'HDInsight': 'HDInsightLinkedService', 'Dynamics': 'DynamicsLinkedService', 'CosmosDb': 'CosmosDbLinkedService', 'AzureKeyVault': 'AzureKeyVaultLinkedService', 'AzureBatch': 'AzureBatchLinkedService', 'AzureSqlDatabase': 'AzureSqlDatabaseLinkedService', 'SqlServer': 'SqlServerLinkedService', 'AzureSqlDW': 'AzureSqlDWLinkedService', 'AzureTableStorage': 'AzureTableStorageLinkedService', 'AzureBlobStorage': 'AzureBlobStorageLinkedService', 'AzureStorage': 'AzureStorageLinkedService'} + 'type': {'Responsys': 'ResponsysLinkedService', 'AzureDatabricks': 'AzureDatabricksLinkedService', 'AzureDataLakeAnalytics': 'AzureDataLakeAnalyticsLinkedService', 'HDInsightOnDemand': 'HDInsightOnDemandLinkedService', 'SalesforceMarketingCloud': 'SalesforceMarketingCloudLinkedService', 'Netezza': 'NetezzaLinkedService', 'Vertica': 'VerticaLinkedService', 'Zoho': 'ZohoLinkedService', 'Xero': 'XeroLinkedService', 'Square': 'SquareLinkedService', 'Spark': 'SparkLinkedService', 'Shopify': 'ShopifyLinkedService', 'ServiceNow': 'ServiceNowLinkedService', 'QuickBooks': 'QuickBooksLinkedService', 'Presto': 'PrestoLinkedService', 'Phoenix': 'PhoenixLinkedService', 'Paypal': 'PaypalLinkedService', 'Marketo': 'MarketoLinkedService', 'MariaDB': 'MariaDBLinkedService', 'Magento': 'MagentoLinkedService', 'Jira': 'JiraLinkedService', 'Impala': 'ImpalaLinkedService', 'Hubspot': 'HubspotLinkedService', 'Hive': 'HiveLinkedService', 'HBase': 'HBaseLinkedService', 'Greenplum': 'GreenplumLinkedService', 'GoogleBigQuery': 'GoogleBigQueryLinkedService', 'GoogleAdWords': 'GoogleAdWordsLinkedService', 'Eloqua': 'EloquaLinkedService', 'Drill': 'DrillLinkedService', 'Couchbase': 'CouchbaseLinkedService', 'Concur': 'ConcurLinkedService', 'AzurePostgreSql': 'AzurePostgreSqlLinkedService', 'AmazonMWS': 'AmazonMWSLinkedService', 'SapHana': 'SapHanaLinkedService', 'SapBW': 'SapBWLinkedService', 'Sftp': 'SftpServerLinkedService', 'FtpServer': 'FtpServerLinkedService', 'RestService': 'RestServiceLinkedService', 'HttpServer': 'HttpLinkedService', 'AzureSearch': 'AzureSearchLinkedService', 'CustomDataSource': 'CustomDataSourceLinkedService', 'AmazonRedshift': 'AmazonRedshiftLinkedService', 'AmazonS3': 'AmazonS3LinkedService', 'SapEcc': 'SapEccLinkedService', 'SapCloudForCustomer': 'SapCloudForCustomerLinkedService', 'Salesforce': 'SalesforceLinkedService', 'AzureDataLakeStore': 'AzureDataLakeStoreLinkedService', 'MongoDb': 'MongoDbLinkedService', 'Cassandra': 'CassandraLinkedService', 'Web': 'WebLinkedService', 'OData': 'ODataLinkedService', 'Hdfs': 'HdfsLinkedService', 'Odbc': 'OdbcLinkedService', 'AzureML': 'AzureMLLinkedService', 'Teradata': 'TeradataLinkedService', 'Db2': 'Db2LinkedService', 'Sybase': 'SybaseLinkedService', 'PostgreSql': 'PostgreSqlLinkedService', 'MySql': 'MySqlLinkedService', 'AzureMySql': 'AzureMySqlLinkedService', 'OracleServiceCloud': 'OracleServiceCloudLinkedService', 'Oracle': 'OracleLinkedService', 'FileServer': 'FileServerLinkedService', 'HDInsight': 'HDInsightLinkedService', 'DynamicsAX': 'DynamicsAXLinkedService', 'Dynamics': 'DynamicsLinkedService', 'CosmosDb': 'CosmosDbLinkedService', 'AzureKeyVault': 'AzureKeyVaultLinkedService', 'AzureBatch': 'AzureBatchLinkedService', 'AzureSqlDatabase': 'AzureSqlDatabaseLinkedService', 'SqlServer': 'SqlServerLinkedService', 'AzureSqlDW': 'AzureSqlDWLinkedService', 'AzureTableStorage': 'AzureTableStorageLinkedService', 'AzureBlobStorage': 'AzureBlobStorageLinkedService', 'AzureStorage': 'AzureStorageLinkedService'} } def __init__(self, *, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, **kwargs) -> None: diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odata_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odata_linked_service.py index 9a7edca9ddb1..2c2fbab04c62 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odata_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odata_linked_service.py @@ -36,8 +36,9 @@ class ODataLinkedService(LinkedService): :param url: Required. The URL of the OData service endpoint. Type: string (or Expression with resultType string). :type url: object - :param authentication_type: Type of authentication used to connect to the - OData service. Possible values include: 'Basic', 'Anonymous' + :param authentication_type: Required. Type of authentication used to + connect to the OData service. Possible values include: 'Anonymous', + 'Basic', 'Windows', 'AadServicePrincipal', 'ManagedServiceIdentity' :type authentication_type: str or ~azure.mgmt.datafactory.models.ODataAuthenticationType :param user_name: User name of the OData service. Type: string (or @@ -45,6 +46,38 @@ class ODataLinkedService(LinkedService): :type user_name: object :param password: Password of the OData service. :type password: ~azure.mgmt.datafactory.models.SecretBase + :param tenant: Specify the tenant information (domain name or tenant ID) + under which your application resides. Type: string (or Expression with + resultType string). + :type tenant: object + :param service_principal_id: Specify the application id of your + application registered in Azure Active Directory. Type: string (or + Expression with resultType string). + :type service_principal_id: object + :param aad_resource_id: Specify the resource you are requesting + authorization to use Directory. Type: string (or Expression with + resultType string). + :type aad_resource_id: object + :param aad_service_principal_credential_type: Specify the credential type + (key or cert) is used for service principal. Possible values include: + 'ServicePrincipalKey', 'ServicePrincipalCert' + :type aad_service_principal_credential_type: str or + ~azure.mgmt.datafactory.models.ODataAadServicePrincipalCredentialType + :param service_principal_key: Specify the secret of your application + registered in Azure Active Directory. Type: string (or Expression with + resultType string). + :type service_principal_key: ~azure.mgmt.datafactory.models.SecretBase + :param service_principal_embedded_cert: Specify the base64 encoded + certificate of your application registered in Azure Active Directory. + Type: string (or Expression with resultType string). + :type service_principal_embedded_cert: + ~azure.mgmt.datafactory.models.SecretBase + :param service_principal_embedded_cert_password: Specify the password of + your certificate if your certificate has a password and you are using + AadServicePrincipal authentication. Type: string (or Expression with + resultType string). + :type service_principal_embedded_cert_password: + ~azure.mgmt.datafactory.models.SecretBase :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -54,6 +87,7 @@ class ODataLinkedService(LinkedService): _validation = { 'type': {'required': True}, 'url': {'required': True}, + 'authentication_type': {'required': True}, } _attribute_map = { @@ -67,6 +101,13 @@ class ODataLinkedService(LinkedService): 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, 'user_name': {'key': 'typeProperties.userName', 'type': 'object'}, 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, + 'tenant': {'key': 'typeProperties.tenant', 'type': 'object'}, + 'service_principal_id': {'key': 'typeProperties.servicePrincipalId', 'type': 'object'}, + 'aad_resource_id': {'key': 'typeProperties.aadResourceId', 'type': 'object'}, + 'aad_service_principal_credential_type': {'key': 'typeProperties.aadServicePrincipalCredentialType', 'type': 'str'}, + 'service_principal_key': {'key': 'typeProperties.servicePrincipalKey', 'type': 'SecretBase'}, + 'service_principal_embedded_cert': {'key': 'typeProperties.servicePrincipalEmbeddedCert', 'type': 'SecretBase'}, + 'service_principal_embedded_cert_password': {'key': 'typeProperties.servicePrincipalEmbeddedCertPassword', 'type': 'SecretBase'}, 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } @@ -76,5 +117,12 @@ def __init__(self, **kwargs): self.authentication_type = kwargs.get('authentication_type', None) self.user_name = kwargs.get('user_name', None) self.password = kwargs.get('password', None) + self.tenant = kwargs.get('tenant', None) + self.service_principal_id = kwargs.get('service_principal_id', None) + self.aad_resource_id = kwargs.get('aad_resource_id', None) + self.aad_service_principal_credential_type = kwargs.get('aad_service_principal_credential_type', None) + self.service_principal_key = kwargs.get('service_principal_key', None) + self.service_principal_embedded_cert = kwargs.get('service_principal_embedded_cert', None) + self.service_principal_embedded_cert_password = kwargs.get('service_principal_embedded_cert_password', None) self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'OData' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odata_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odata_linked_service_py3.py index 688bb4e4ffda..80088701b50b 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odata_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odata_linked_service_py3.py @@ -36,8 +36,9 @@ class ODataLinkedService(LinkedService): :param url: Required. The URL of the OData service endpoint. Type: string (or Expression with resultType string). :type url: object - :param authentication_type: Type of authentication used to connect to the - OData service. Possible values include: 'Basic', 'Anonymous' + :param authentication_type: Required. Type of authentication used to + connect to the OData service. Possible values include: 'Anonymous', + 'Basic', 'Windows', 'AadServicePrincipal', 'ManagedServiceIdentity' :type authentication_type: str or ~azure.mgmt.datafactory.models.ODataAuthenticationType :param user_name: User name of the OData service. Type: string (or @@ -45,6 +46,38 @@ class ODataLinkedService(LinkedService): :type user_name: object :param password: Password of the OData service. :type password: ~azure.mgmt.datafactory.models.SecretBase + :param tenant: Specify the tenant information (domain name or tenant ID) + under which your application resides. Type: string (or Expression with + resultType string). + :type tenant: object + :param service_principal_id: Specify the application id of your + application registered in Azure Active Directory. Type: string (or + Expression with resultType string). + :type service_principal_id: object + :param aad_resource_id: Specify the resource you are requesting + authorization to use Directory. Type: string (or Expression with + resultType string). + :type aad_resource_id: object + :param aad_service_principal_credential_type: Specify the credential type + (key or cert) is used for service principal. Possible values include: + 'ServicePrincipalKey', 'ServicePrincipalCert' + :type aad_service_principal_credential_type: str or + ~azure.mgmt.datafactory.models.ODataAadServicePrincipalCredentialType + :param service_principal_key: Specify the secret of your application + registered in Azure Active Directory. Type: string (or Expression with + resultType string). + :type service_principal_key: ~azure.mgmt.datafactory.models.SecretBase + :param service_principal_embedded_cert: Specify the base64 encoded + certificate of your application registered in Azure Active Directory. + Type: string (or Expression with resultType string). + :type service_principal_embedded_cert: + ~azure.mgmt.datafactory.models.SecretBase + :param service_principal_embedded_cert_password: Specify the password of + your certificate if your certificate has a password and you are using + AadServicePrincipal authentication. Type: string (or Expression with + resultType string). + :type service_principal_embedded_cert_password: + ~azure.mgmt.datafactory.models.SecretBase :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -54,6 +87,7 @@ class ODataLinkedService(LinkedService): _validation = { 'type': {'required': True}, 'url': {'required': True}, + 'authentication_type': {'required': True}, } _attribute_map = { @@ -67,14 +101,28 @@ class ODataLinkedService(LinkedService): 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, 'user_name': {'key': 'typeProperties.userName', 'type': 'object'}, 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, + 'tenant': {'key': 'typeProperties.tenant', 'type': 'object'}, + 'service_principal_id': {'key': 'typeProperties.servicePrincipalId', 'type': 'object'}, + 'aad_resource_id': {'key': 'typeProperties.aadResourceId', 'type': 'object'}, + 'aad_service_principal_credential_type': {'key': 'typeProperties.aadServicePrincipalCredentialType', 'type': 'str'}, + 'service_principal_key': {'key': 'typeProperties.servicePrincipalKey', 'type': 'SecretBase'}, + 'service_principal_embedded_cert': {'key': 'typeProperties.servicePrincipalEmbeddedCert', 'type': 'SecretBase'}, + 'service_principal_embedded_cert_password': {'key': 'typeProperties.servicePrincipalEmbeddedCertPassword', 'type': 'SecretBase'}, 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, *, url, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, authentication_type=None, user_name=None, password=None, encrypted_credential=None, **kwargs) -> None: + def __init__(self, *, url, authentication_type, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, user_name=None, password=None, tenant=None, service_principal_id=None, aad_resource_id=None, aad_service_principal_credential_type=None, service_principal_key=None, service_principal_embedded_cert=None, service_principal_embedded_cert_password=None, encrypted_credential=None, **kwargs) -> None: super(ODataLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) self.url = url self.authentication_type = authentication_type self.user_name = user_name self.password = password + self.tenant = tenant + self.service_principal_id = service_principal_id + self.aad_resource_id = aad_resource_id + self.aad_service_principal_credential_type = aad_service_principal_credential_type + self.service_principal_key = service_principal_key + self.service_principal_embedded_cert = service_principal_embedded_cert + self.service_principal_embedded_cert_password = service_principal_embedded_cert_password self.encrypted_credential = encrypted_credential self.type = 'OData' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_service_cloud_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_service_cloud_linked_service.py new file mode 100644 index 000000000000..fae3d3a7a8d4 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_service_cloud_linked_service.py @@ -0,0 +1,95 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class OracleServiceCloudLinkedService(LinkedService): + """Oracle Service Cloud linked service. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param host: Required. The URL of the Oracle Service Cloud instance. + :type host: object + :param username: Required. The user name that you use to access Oracle + Service Cloud server. + :type username: object + :param password: Required. The password corresponding to the user name + that you provided in the username key. + :type password: ~azure.mgmt.datafactory.models.SecretBase + :param use_encrypted_endpoints: Specifies whether the data source + endpoints are encrypted using HTTPS. The default value is true. Type: + boolean (or Expression with resultType boolean). + :type use_encrypted_endpoints: object + :param use_host_verification: Specifies whether to require the host name + in the server's certificate to match the host name of the server when + connecting over SSL. The default value is true. Type: boolean (or + Expression with resultType boolean). + :type use_host_verification: object + :param use_peer_verification: Specifies whether to verify the identity of + the server when connecting over SSL. The default value is true. Type: + boolean (or Expression with resultType boolean). + :type use_peer_verification: object + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'host': {'required': True}, + 'username': {'required': True}, + 'password': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'host': {'key': 'typeProperties.host', 'type': 'object'}, + 'username': {'key': 'typeProperties.username', 'type': 'object'}, + 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, + 'use_encrypted_endpoints': {'key': 'typeProperties.useEncryptedEndpoints', 'type': 'object'}, + 'use_host_verification': {'key': 'typeProperties.useHostVerification', 'type': 'object'}, + 'use_peer_verification': {'key': 'typeProperties.usePeerVerification', 'type': 'object'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, **kwargs): + super(OracleServiceCloudLinkedService, self).__init__(**kwargs) + self.host = kwargs.get('host', None) + self.username = kwargs.get('username', None) + self.password = kwargs.get('password', None) + self.use_encrypted_endpoints = kwargs.get('use_encrypted_endpoints', None) + self.use_host_verification = kwargs.get('use_host_verification', None) + self.use_peer_verification = kwargs.get('use_peer_verification', None) + self.encrypted_credential = kwargs.get('encrypted_credential', None) + self.type = 'OracleServiceCloud' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_service_cloud_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_service_cloud_linked_service_py3.py new file mode 100644 index 000000000000..0936681e6ac5 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_service_cloud_linked_service_py3.py @@ -0,0 +1,95 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service_py3 import LinkedService + + +class OracleServiceCloudLinkedService(LinkedService): + """Oracle Service Cloud linked service. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param host: Required. The URL of the Oracle Service Cloud instance. + :type host: object + :param username: Required. The user name that you use to access Oracle + Service Cloud server. + :type username: object + :param password: Required. The password corresponding to the user name + that you provided in the username key. + :type password: ~azure.mgmt.datafactory.models.SecretBase + :param use_encrypted_endpoints: Specifies whether the data source + endpoints are encrypted using HTTPS. The default value is true. Type: + boolean (or Expression with resultType boolean). + :type use_encrypted_endpoints: object + :param use_host_verification: Specifies whether to require the host name + in the server's certificate to match the host name of the server when + connecting over SSL. The default value is true. Type: boolean (or + Expression with resultType boolean). + :type use_host_verification: object + :param use_peer_verification: Specifies whether to verify the identity of + the server when connecting over SSL. The default value is true. Type: + boolean (or Expression with resultType boolean). + :type use_peer_verification: object + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'host': {'required': True}, + 'username': {'required': True}, + 'password': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'host': {'key': 'typeProperties.host', 'type': 'object'}, + 'username': {'key': 'typeProperties.username', 'type': 'object'}, + 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, + 'use_encrypted_endpoints': {'key': 'typeProperties.useEncryptedEndpoints', 'type': 'object'}, + 'use_host_verification': {'key': 'typeProperties.useHostVerification', 'type': 'object'}, + 'use_peer_verification': {'key': 'typeProperties.usePeerVerification', 'type': 'object'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, *, host, username, password, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, use_encrypted_endpoints=None, use_host_verification=None, use_peer_verification=None, encrypted_credential=None, **kwargs) -> None: + super(OracleServiceCloudLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.host = host + self.username = username + self.password = password + self.use_encrypted_endpoints = use_encrypted_endpoints + self.use_host_verification = use_host_verification + self.use_peer_verification = use_peer_verification + self.encrypted_credential = encrypted_credential + self.type = 'OracleServiceCloud' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_service_cloud_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_service_cloud_object_dataset.py new file mode 100644 index 000000000000..4b842b975c70 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_service_cloud_object_dataset.py @@ -0,0 +1,62 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class OracleServiceCloudObjectDataset(Dataset): + """Oracle Service Cloud dataset. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Required. Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder + :param type: Required. Constant filled by server. + :type type: str + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, **kwargs): + super(OracleServiceCloudObjectDataset, self).__init__(**kwargs) + self.type = 'OracleServiceCloudObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_service_cloud_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_service_cloud_object_dataset_py3.py new file mode 100644 index 000000000000..2e1d1db41586 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_service_cloud_object_dataset_py3.py @@ -0,0 +1,62 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset_py3 import Dataset + + +class OracleServiceCloudObjectDataset(Dataset): + """Oracle Service Cloud dataset. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Required. Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder + :param type: Required. Constant filled by server. + :type type: str + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + super(OracleServiceCloudObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + self.type = 'OracleServiceCloudObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_service_cloud_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_service_cloud_source.py new file mode 100644 index 000000000000..1d362862c737 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_service_cloud_source.py @@ -0,0 +1,52 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class OracleServiceCloudSource(CopySource): + """A copy activity Oracle Service Cloud source. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Required. Constant filled by server. + :type type: str + :param query: A query to retrieve data from source. Type: string (or + Expression with resultType string). + :type query: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'query': {'key': 'query', 'type': 'object'}, + } + + def __init__(self, **kwargs): + super(OracleServiceCloudSource, self).__init__(**kwargs) + self.query = kwargs.get('query', None) + self.type = 'OracleServiceCloudSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_service_cloud_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_service_cloud_source_py3.py new file mode 100644 index 000000000000..2eebde6c3e50 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_service_cloud_source_py3.py @@ -0,0 +1,52 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source_py3 import CopySource + + +class OracleServiceCloudSource(CopySource): + """A copy activity Oracle Service Cloud source. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Required. Constant filled by server. + :type type: str + :param query: A query to retrieve data from source. Type: string (or + Expression with resultType string). + :type query: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'query': {'key': 'query', 'type': 'object'}, + } + + def __init__(self, *, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None, **kwargs) -> None: + super(OracleServiceCloudSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, **kwargs) + self.query = query + self.type = 'OracleServiceCloudSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/rest_service_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/rest_service_dataset.py new file mode 100644 index 000000000000..e3ec02526c91 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/rest_service_dataset.py @@ -0,0 +1,88 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class RestServiceDataset(Dataset): + """A Rest service dataset. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Required. Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder + :param type: Required. Constant filled by server. + :type type: str + :param relative_url: The relative URL to the resource that the RESTful API + provides. Type: string (or Expression with resultType string). + :type relative_url: object + :param request_method: The HTTP method used to call the RESTful API. The + default is GET. Type: string (or Expression with resultType string). + :type request_method: object + :param request_body: The HTTP request body to the RESTful API if + requestMethod is POST. Type: string (or Expression with resultType + string). + :type request_body: object + :param additional_headers: The additional HTTP headers in the request to + the RESTful API. Type: string (or Expression with resultType string). + :type additional_headers: object + :param pagination_rules: The pagination rules to compose next page + requests. + :type pagination_rules: object + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, + 'type': {'key': 'type', 'type': 'str'}, + 'relative_url': {'key': 'typeProperties.relativeUrl', 'type': 'object'}, + 'request_method': {'key': 'typeProperties.requestMethod', 'type': 'object'}, + 'request_body': {'key': 'typeProperties.requestBody', 'type': 'object'}, + 'additional_headers': {'key': 'typeProperties.additionalHeaders', 'type': 'object'}, + 'pagination_rules': {'key': 'typeProperties.paginationRules', 'type': 'object'}, + } + + def __init__(self, **kwargs): + super(RestServiceDataset, self).__init__(**kwargs) + self.relative_url = kwargs.get('relative_url', None) + self.request_method = kwargs.get('request_method', None) + self.request_body = kwargs.get('request_body', None) + self.additional_headers = kwargs.get('additional_headers', None) + self.pagination_rules = kwargs.get('pagination_rules', None) + self.type = 'RestResource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/rest_service_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/rest_service_dataset_py3.py new file mode 100644 index 000000000000..0813a1f29624 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/rest_service_dataset_py3.py @@ -0,0 +1,88 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset_py3 import Dataset + + +class RestServiceDataset(Dataset): + """A Rest service dataset. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Required. Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param folder: The folder that this Dataset is in. If not specified, + Dataset will appear at the root level. + :type folder: ~azure.mgmt.datafactory.models.DatasetFolder + :param type: Required. Constant filled by server. + :type type: str + :param relative_url: The relative URL to the resource that the RESTful API + provides. Type: string (or Expression with resultType string). + :type relative_url: object + :param request_method: The HTTP method used to call the RESTful API. The + default is GET. Type: string (or Expression with resultType string). + :type request_method: object + :param request_body: The HTTP request body to the RESTful API if + requestMethod is POST. Type: string (or Expression with resultType + string). + :type request_body: object + :param additional_headers: The additional HTTP headers in the request to + the RESTful API. Type: string (or Expression with resultType string). + :type additional_headers: object + :param pagination_rules: The pagination rules to compose next page + requests. + :type pagination_rules: object + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, + 'type': {'key': 'type', 'type': 'str'}, + 'relative_url': {'key': 'typeProperties.relativeUrl', 'type': 'object'}, + 'request_method': {'key': 'typeProperties.requestMethod', 'type': 'object'}, + 'request_body': {'key': 'typeProperties.requestBody', 'type': 'object'}, + 'additional_headers': {'key': 'typeProperties.additionalHeaders', 'type': 'object'}, + 'pagination_rules': {'key': 'typeProperties.paginationRules', 'type': 'object'}, + } + + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, relative_url=None, request_method=None, request_body=None, additional_headers=None, pagination_rules=None, **kwargs) -> None: + super(RestServiceDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + self.relative_url = relative_url + self.request_method = request_method + self.request_body = request_body + self.additional_headers = additional_headers + self.pagination_rules = pagination_rules + self.type = 'RestResource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/rest_service_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/rest_service_linked_service.py new file mode 100644 index 000000000000..8515b41394e3 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/rest_service_linked_service.py @@ -0,0 +1,107 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class RestServiceLinkedService(LinkedService): + """Rest Service linked service. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param url: Required. The base URL of the REST service. + :type url: object + :param enable_server_certificate_validation: Whether to validate server + side SSL certificate when connecting to the endpoint.The default value is + true. Type: boolean (or Expression with resultType boolean). + :type enable_server_certificate_validation: object + :param authentication_type: Required. Type of authentication used to + connect to the REST service. Possible values include: 'Anonymous', + 'Basic', 'AadServicePrincial', 'ManagedServiceIdentity' + :type authentication_type: str or + ~azure.mgmt.datafactory.models.RestServiceAuthenticationType + :param user_name: The user name used in Basic authentication type. + :type user_name: object + :param password: The password used in Basic authentication type. + :type password: ~azure.mgmt.datafactory.models.SecretBase + :param service_principal_id: The application’s client ID used in + AadServicePrincipal authentication type. + :type service_principal_id: object + :param service_principal_key: The application’s key used in + AadServicePrincipal authentication type. + :type service_principal_key: ~azure.mgmt.datafactory.models.SecretBase + :param tenant: The tenant information (domain name or tenant ID) used in + AadServicePrincipal authentication type under which your application + resides. + :type tenant: object + :param aad_resource_id: The resource you are requesting authorization to + use. + :type aad_resource_id: object + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'url': {'required': True}, + 'authentication_type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'url': {'key': 'typeProperties.url', 'type': 'object'}, + 'enable_server_certificate_validation': {'key': 'typeProperties.enableServerCertificateValidation', 'type': 'object'}, + 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, + 'user_name': {'key': 'typeProperties.userName', 'type': 'object'}, + 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, + 'service_principal_id': {'key': 'typeProperties.servicePrincipalId', 'type': 'object'}, + 'service_principal_key': {'key': 'typeProperties.servicePrincipalKey', 'type': 'SecretBase'}, + 'tenant': {'key': 'typeProperties.tenant', 'type': 'object'}, + 'aad_resource_id': {'key': 'typeProperties.aadResourceId', 'type': 'object'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, **kwargs): + super(RestServiceLinkedService, self).__init__(**kwargs) + self.url = kwargs.get('url', None) + self.enable_server_certificate_validation = kwargs.get('enable_server_certificate_validation', None) + self.authentication_type = kwargs.get('authentication_type', None) + self.user_name = kwargs.get('user_name', None) + self.password = kwargs.get('password', None) + self.service_principal_id = kwargs.get('service_principal_id', None) + self.service_principal_key = kwargs.get('service_principal_key', None) + self.tenant = kwargs.get('tenant', None) + self.aad_resource_id = kwargs.get('aad_resource_id', None) + self.encrypted_credential = kwargs.get('encrypted_credential', None) + self.type = 'RestService' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/rest_service_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/rest_service_linked_service_py3.py new file mode 100644 index 000000000000..7e9f9efd0281 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/rest_service_linked_service_py3.py @@ -0,0 +1,107 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service_py3 import LinkedService + + +class RestServiceLinkedService(LinkedService): + """Rest Service linked service. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param url: Required. The base URL of the REST service. + :type url: object + :param enable_server_certificate_validation: Whether to validate server + side SSL certificate when connecting to the endpoint.The default value is + true. Type: boolean (or Expression with resultType boolean). + :type enable_server_certificate_validation: object + :param authentication_type: Required. Type of authentication used to + connect to the REST service. Possible values include: 'Anonymous', + 'Basic', 'AadServicePrincial', 'ManagedServiceIdentity' + :type authentication_type: str or + ~azure.mgmt.datafactory.models.RestServiceAuthenticationType + :param user_name: The user name used in Basic authentication type. + :type user_name: object + :param password: The password used in Basic authentication type. + :type password: ~azure.mgmt.datafactory.models.SecretBase + :param service_principal_id: The application’s client ID used in + AadServicePrincipal authentication type. + :type service_principal_id: object + :param service_principal_key: The application’s key used in + AadServicePrincipal authentication type. + :type service_principal_key: ~azure.mgmt.datafactory.models.SecretBase + :param tenant: The tenant information (domain name or tenant ID) used in + AadServicePrincipal authentication type under which your application + resides. + :type tenant: object + :param aad_resource_id: The resource you are requesting authorization to + use. + :type aad_resource_id: object + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'url': {'required': True}, + 'authentication_type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'url': {'key': 'typeProperties.url', 'type': 'object'}, + 'enable_server_certificate_validation': {'key': 'typeProperties.enableServerCertificateValidation', 'type': 'object'}, + 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, + 'user_name': {'key': 'typeProperties.userName', 'type': 'object'}, + 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, + 'service_principal_id': {'key': 'typeProperties.servicePrincipalId', 'type': 'object'}, + 'service_principal_key': {'key': 'typeProperties.servicePrincipalKey', 'type': 'SecretBase'}, + 'tenant': {'key': 'typeProperties.tenant', 'type': 'object'}, + 'aad_resource_id': {'key': 'typeProperties.aadResourceId', 'type': 'object'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, *, url, authentication_type, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, enable_server_certificate_validation=None, user_name=None, password=None, service_principal_id=None, service_principal_key=None, tenant=None, aad_resource_id=None, encrypted_credential=None, **kwargs) -> None: + super(RestServiceLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.url = url + self.enable_server_certificate_validation = enable_server_certificate_validation + self.authentication_type = authentication_type + self.user_name = user_name + self.password = password + self.service_principal_id = service_principal_id + self.service_principal_key = service_principal_key + self.tenant = tenant + self.aad_resource_id = aad_resource_id + self.encrypted_credential = encrypted_credential + self.type = 'RestService' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/rest_service_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/rest_service_source.py new file mode 100644 index 000000000000..9a62dcdbd682 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/rest_service_source.py @@ -0,0 +1,62 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class RestServiceSource(CopySource): + """A copy activity Rest service source. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Required. Constant filled by server. + :type type: str + :param http_request_timeout: The timeout (TimeSpan) to get an HTTP + response. It is the timeout to get a response, not the timeout to read + response data. Default value: 00:01:40. Type: string (or Expression with + resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type http_request_timeout: object + :param request_interval: The time to await before sending next page + request. Default value: 00:00:01. Type: string (or Expression with + resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type request_interval: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'http_request_timeout': {'key': 'httpRequestTimeout', 'type': 'object'}, + 'request_interval': {'key': 'requestInterval', 'type': 'object'}, + } + + def __init__(self, **kwargs): + super(RestServiceSource, self).__init__(**kwargs) + self.http_request_timeout = kwargs.get('http_request_timeout', None) + self.request_interval = kwargs.get('request_interval', None) + self.type = 'RestServiceSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/rest_service_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/rest_service_source_py3.py new file mode 100644 index 000000000000..85674d829a1f --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/rest_service_source_py3.py @@ -0,0 +1,62 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source_py3 import CopySource + + +class RestServiceSource(CopySource): + """A copy activity Rest service source. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Required. Constant filled by server. + :type type: str + :param http_request_timeout: The timeout (TimeSpan) to get an HTTP + response. It is the timeout to get a response, not the timeout to read + response data. Default value: 00:01:40. Type: string (or Expression with + resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type http_request_timeout: object + :param request_interval: The time to await before sending next page + request. Default value: 00:00:01. Type: string (or Expression with + resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type request_interval: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'http_request_timeout': {'key': 'httpRequestTimeout', 'type': 'object'}, + 'request_interval': {'key': 'requestInterval', 'type': 'object'}, + } + + def __init__(self, *, additional_properties=None, source_retry_count=None, source_retry_wait=None, http_request_timeout=None, request_interval=None, **kwargs) -> None: + super(RestServiceSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, **kwargs) + self.http_request_timeout = http_request_timeout + self.request_interval = request_interval + self.type = 'RestServiceSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/stored_procedure_parameter.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/stored_procedure_parameter.py index 63d2c3b80aa5..748cf7cba53c 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/stored_procedure_parameter.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/stored_procedure_parameter.py @@ -15,9 +15,7 @@ class StoredProcedureParameter(Model): """SQL stored procedure parameter. - All required parameters must be populated in order to send to Azure. - - :param value: Required. Stored procedure parameter value. Type: string (or + :param value: Stored procedure parameter value. Type: string (or Expression with resultType string). :type value: object :param type: Stored procedure parameter type. Possible values include: @@ -26,10 +24,6 @@ class StoredProcedureParameter(Model): ~azure.mgmt.datafactory.models.StoredProcedureParameterType """ - _validation = { - 'value': {'required': True}, - } - _attribute_map = { 'value': {'key': 'value', 'type': 'object'}, 'type': {'key': 'type', 'type': 'str'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/stored_procedure_parameter_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/stored_procedure_parameter_py3.py index f5c55777e0c3..bd967ce52876 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/stored_procedure_parameter_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/stored_procedure_parameter_py3.py @@ -15,9 +15,7 @@ class StoredProcedureParameter(Model): """SQL stored procedure parameter. - All required parameters must be populated in order to send to Azure. - - :param value: Required. Stored procedure parameter value. Type: string (or + :param value: Stored procedure parameter value. Type: string (or Expression with resultType string). :type value: object :param type: Stored procedure parameter type. Possible values include: @@ -26,16 +24,12 @@ class StoredProcedureParameter(Model): ~azure.mgmt.datafactory.models.StoredProcedureParameterType """ - _validation = { - 'value': {'required': True}, - } - _attribute_map = { 'value': {'key': 'value', 'type': 'object'}, 'type': {'key': 'type', 'type': 'str'}, } - def __init__(self, *, value, type=None, **kwargs) -> None: + def __init__(self, *, value=None, type=None, **kwargs) -> None: super(StoredProcedureParameter, self).__init__(**kwargs) self.value = value self.type = type From 59275e6f89bee9587a4708c3b3ab02639d3949d4 Mon Sep 17 00:00:00 2001 From: Azure SDK for Python bot Date: Wed, 24 Oct 2018 10:47:12 -0700 Subject: [PATCH 27/54] Generated from 6843f73390001df17df0c164e48a9eab640718f7 (#3672) Revert "Swagger changes for connectors: Google AdWords, Oracle Service Cloud, OData, Rest Service, Dynamics AX and support for wildcardPath. (#4258)" This reverts commit c7b860297ddf53c1f9ef501583c824fdd00cbc40. --- .../azure/mgmt/datafactory/models/__init__.py | 42 ------- .../datafactory/models/amazon_s3_dataset.py | 5 - .../models/amazon_s3_dataset_py3.py | 7 +- .../datafactory/models/azure_blob_dataset.py | 6 - .../models/azure_blob_dataset_py3.py | 8 +- .../models/azure_data_lake_store_dataset.py | 11 +- .../azure_data_lake_store_dataset_py3.py | 13 +- .../mgmt/datafactory/models/copy_source.py | 15 ++- .../datafactory/models/copy_source_py3.py | 15 ++- .../data_factory_management_client_enums.py | 25 +--- .../azure/mgmt/datafactory/models/dataset.py | 19 ++- .../mgmt/datafactory/models/dataset_py3.py | 19 ++- .../models/dynamics_ax_linked_service.py | 90 ------------- .../models/dynamics_ax_linked_service_py3.py | 90 ------------- .../models/dynamics_ax_resource_dataset.py | 68 ---------- .../dynamics_ax_resource_dataset_py3.py | 68 ---------- .../datafactory/models/dynamics_ax_source.py | 52 -------- .../models/dynamics_ax_source_py3.py | 52 -------- .../datafactory/models/file_share_dataset.py | 6 - .../models/file_share_dataset_py3.py | 8 +- .../models/google_ad_words_linked_service.py | 119 ------------------ .../google_ad_words_linked_service_py3.py | 119 ------------------ .../models/google_ad_words_object_dataset.py | 62 --------- .../google_ad_words_object_dataset_py3.py | 62 --------- .../models/google_ad_words_source.py | 52 -------- .../models/google_ad_words_source_py3.py | 52 -------- .../mgmt/datafactory/models/linked_service.py | 17 ++- .../datafactory/models/linked_service_py3.py | 17 ++- .../models/odata_linked_service.py | 52 +------- .../models/odata_linked_service_py3.py | 54 +------- .../oracle_service_cloud_linked_service.py | 95 -------------- ...oracle_service_cloud_linked_service_py3.py | 95 -------------- .../oracle_service_cloud_object_dataset.py | 62 --------- ...oracle_service_cloud_object_dataset_py3.py | 62 --------- .../models/oracle_service_cloud_source.py | 52 -------- .../models/oracle_service_cloud_source_py3.py | 52 -------- .../models/rest_service_dataset.py | 88 ------------- .../models/rest_service_dataset_py3.py | 88 ------------- .../models/rest_service_linked_service.py | 107 ---------------- .../models/rest_service_linked_service_py3.py | 107 ---------------- .../datafactory/models/rest_service_source.py | 62 --------- .../models/rest_service_source_py3.py | 62 --------- .../models/stored_procedure_parameter.py | 8 +- .../models/stored_procedure_parameter_py3.py | 10 +- 44 files changed, 79 insertions(+), 2096 deletions(-) delete mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_ax_linked_service.py delete mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_ax_linked_service_py3.py delete mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_ax_resource_dataset.py delete mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_ax_resource_dataset_py3.py delete mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_ax_source.py delete mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_ax_source_py3.py delete mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_ad_words_linked_service.py delete mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_ad_words_linked_service_py3.py delete mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_ad_words_object_dataset.py delete mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_ad_words_object_dataset_py3.py delete mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_ad_words_source.py delete mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_ad_words_source_py3.py delete mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_service_cloud_linked_service.py delete mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_service_cloud_linked_service_py3.py delete mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_service_cloud_object_dataset.py delete mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_service_cloud_object_dataset_py3.py delete mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_service_cloud_source.py delete mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_service_cloud_source_py3.py delete mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/rest_service_dataset.py delete mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/rest_service_dataset_py3.py delete mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/rest_service_linked_service.py delete mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/rest_service_linked_service_py3.py delete mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/rest_service_source.py delete mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/rest_service_source_py3.py diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py index ffdfed6c3c53..cab29db469a7 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py @@ -116,7 +116,6 @@ from .hbase_linked_service_py3 import HBaseLinkedService from .greenplum_linked_service_py3 import GreenplumLinkedService from .google_big_query_linked_service_py3 import GoogleBigQueryLinkedService - from .google_ad_words_linked_service_py3 import GoogleAdWordsLinkedService from .eloqua_linked_service_py3 import EloquaLinkedService from .drill_linked_service_py3 import DrillLinkedService from .couchbase_linked_service_py3 import CouchbaseLinkedService @@ -127,7 +126,6 @@ from .sap_bw_linked_service_py3 import SapBWLinkedService from .sftp_server_linked_service_py3 import SftpServerLinkedService from .ftp_server_linked_service_py3 import FtpServerLinkedService - from .rest_service_linked_service_py3 import RestServiceLinkedService from .http_linked_service_py3 import HttpLinkedService from .azure_search_linked_service_py3 import AzureSearchLinkedService from .custom_data_source_linked_service_py3 import CustomDataSourceLinkedService @@ -154,11 +152,9 @@ from .postgre_sql_linked_service_py3 import PostgreSqlLinkedService from .my_sql_linked_service_py3 import MySqlLinkedService from .azure_my_sql_linked_service_py3 import AzureMySqlLinkedService - from .oracle_service_cloud_linked_service_py3 import OracleServiceCloudLinkedService from .oracle_linked_service_py3 import OracleLinkedService from .file_server_linked_service_py3 import FileServerLinkedService from .hd_insight_linked_service_py3 import HDInsightLinkedService - from .dynamics_ax_linked_service_py3 import DynamicsAXLinkedService from .dynamics_linked_service_py3 import DynamicsLinkedService from .cosmos_db_linked_service_py3 import CosmosDbLinkedService from .azure_key_vault_linked_service_py3 import AzureKeyVaultLinkedService @@ -193,14 +189,12 @@ from .hbase_object_dataset_py3 import HBaseObjectDataset from .greenplum_table_dataset_py3 import GreenplumTableDataset from .google_big_query_object_dataset_py3 import GoogleBigQueryObjectDataset - from .google_ad_words_object_dataset_py3 import GoogleAdWordsObjectDataset from .eloqua_object_dataset_py3 import EloquaObjectDataset from .drill_table_dataset_py3 import DrillTableDataset from .couchbase_table_dataset_py3 import CouchbaseTableDataset from .concur_object_dataset_py3 import ConcurObjectDataset from .azure_postgre_sql_table_dataset_py3 import AzurePostgreSqlTableDataset from .amazon_mws_object_dataset_py3 import AmazonMWSObjectDataset - from .rest_service_dataset_py3 import RestServiceDataset from .dataset_zip_deflate_compression_py3 import DatasetZipDeflateCompression from .dataset_deflate_compression_py3 import DatasetDeflateCompression from .dataset_gzip_compression_py3 import DatasetGZipCompression @@ -221,13 +215,11 @@ from .salesforce_object_dataset_py3 import SalesforceObjectDataset from .relational_table_dataset_py3 import RelationalTableDataset from .azure_my_sql_table_dataset_py3 import AzureMySqlTableDataset - from .oracle_service_cloud_object_dataset_py3 import OracleServiceCloudObjectDataset from .oracle_table_dataset_py3 import OracleTableDataset from .odata_resource_dataset_py3 import ODataResourceDataset from .mongo_db_collection_dataset_py3 import MongoDbCollectionDataset from .file_share_dataset_py3 import FileShareDataset from .azure_data_lake_store_dataset_py3 import AzureDataLakeStoreDataset - from .dynamics_ax_resource_dataset_py3 import DynamicsAXResourceDataset from .dynamics_entity_dataset_py3 import DynamicsEntityDataset from .document_db_collection_dataset_py3 import DocumentDbCollectionDataset from .custom_dataset_py3 import CustomDataset @@ -274,20 +266,17 @@ from .hbase_source_py3 import HBaseSource from .greenplum_source_py3 import GreenplumSource from .google_big_query_source_py3 import GoogleBigQuerySource - from .google_ad_words_source_py3 import GoogleAdWordsSource from .eloqua_source_py3 import EloquaSource from .drill_source_py3 import DrillSource from .couchbase_source_py3 import CouchbaseSource from .concur_source_py3 import ConcurSource from .azure_postgre_sql_source_py3 import AzurePostgreSqlSource from .amazon_mws_source_py3 import AmazonMWSSource - from .rest_service_source_py3 import RestServiceSource from .http_source_py3 import HttpSource from .azure_data_lake_store_source_py3 import AzureDataLakeStoreSource from .mongo_db_source_py3 import MongoDbSource from .cassandra_source_py3 import CassandraSource from .web_source_py3 import WebSource - from .oracle_service_cloud_source_py3 import OracleServiceCloudSource from .oracle_source_py3 import OracleSource from .azure_my_sql_source_py3 import AzureMySqlSource from .distcp_settings_py3 import DistcpSettings @@ -300,7 +289,6 @@ from .sap_cloud_for_customer_source_py3 import SapCloudForCustomerSource from .salesforce_source_py3 import SalesforceSource from .relational_source_py3 import RelationalSource - from .dynamics_ax_source_py3 import DynamicsAXSource from .dynamics_source_py3 import DynamicsSource from .document_db_collection_source_py3 import DocumentDbCollectionSource from .blob_source_py3 import BlobSource @@ -480,7 +468,6 @@ from .hbase_linked_service import HBaseLinkedService from .greenplum_linked_service import GreenplumLinkedService from .google_big_query_linked_service import GoogleBigQueryLinkedService - from .google_ad_words_linked_service import GoogleAdWordsLinkedService from .eloqua_linked_service import EloquaLinkedService from .drill_linked_service import DrillLinkedService from .couchbase_linked_service import CouchbaseLinkedService @@ -491,7 +478,6 @@ from .sap_bw_linked_service import SapBWLinkedService from .sftp_server_linked_service import SftpServerLinkedService from .ftp_server_linked_service import FtpServerLinkedService - from .rest_service_linked_service import RestServiceLinkedService from .http_linked_service import HttpLinkedService from .azure_search_linked_service import AzureSearchLinkedService from .custom_data_source_linked_service import CustomDataSourceLinkedService @@ -518,11 +504,9 @@ from .postgre_sql_linked_service import PostgreSqlLinkedService from .my_sql_linked_service import MySqlLinkedService from .azure_my_sql_linked_service import AzureMySqlLinkedService - from .oracle_service_cloud_linked_service import OracleServiceCloudLinkedService from .oracle_linked_service import OracleLinkedService from .file_server_linked_service import FileServerLinkedService from .hd_insight_linked_service import HDInsightLinkedService - from .dynamics_ax_linked_service import DynamicsAXLinkedService from .dynamics_linked_service import DynamicsLinkedService from .cosmos_db_linked_service import CosmosDbLinkedService from .azure_key_vault_linked_service import AzureKeyVaultLinkedService @@ -557,14 +541,12 @@ from .hbase_object_dataset import HBaseObjectDataset from .greenplum_table_dataset import GreenplumTableDataset from .google_big_query_object_dataset import GoogleBigQueryObjectDataset - from .google_ad_words_object_dataset import GoogleAdWordsObjectDataset from .eloqua_object_dataset import EloquaObjectDataset from .drill_table_dataset import DrillTableDataset from .couchbase_table_dataset import CouchbaseTableDataset from .concur_object_dataset import ConcurObjectDataset from .azure_postgre_sql_table_dataset import AzurePostgreSqlTableDataset from .amazon_mws_object_dataset import AmazonMWSObjectDataset - from .rest_service_dataset import RestServiceDataset from .dataset_zip_deflate_compression import DatasetZipDeflateCompression from .dataset_deflate_compression import DatasetDeflateCompression from .dataset_gzip_compression import DatasetGZipCompression @@ -585,13 +567,11 @@ from .salesforce_object_dataset import SalesforceObjectDataset from .relational_table_dataset import RelationalTableDataset from .azure_my_sql_table_dataset import AzureMySqlTableDataset - from .oracle_service_cloud_object_dataset import OracleServiceCloudObjectDataset from .oracle_table_dataset import OracleTableDataset from .odata_resource_dataset import ODataResourceDataset from .mongo_db_collection_dataset import MongoDbCollectionDataset from .file_share_dataset import FileShareDataset from .azure_data_lake_store_dataset import AzureDataLakeStoreDataset - from .dynamics_ax_resource_dataset import DynamicsAXResourceDataset from .dynamics_entity_dataset import DynamicsEntityDataset from .document_db_collection_dataset import DocumentDbCollectionDataset from .custom_dataset import CustomDataset @@ -638,20 +618,17 @@ from .hbase_source import HBaseSource from .greenplum_source import GreenplumSource from .google_big_query_source import GoogleBigQuerySource - from .google_ad_words_source import GoogleAdWordsSource from .eloqua_source import EloquaSource from .drill_source import DrillSource from .couchbase_source import CouchbaseSource from .concur_source import ConcurSource from .azure_postgre_sql_source import AzurePostgreSqlSource from .amazon_mws_source import AmazonMWSSource - from .rest_service_source import RestServiceSource from .http_source import HttpSource from .azure_data_lake_store_source import AzureDataLakeStoreSource from .mongo_db_source import MongoDbSource from .cassandra_source import CassandraSource from .web_source import WebSource - from .oracle_service_cloud_source import OracleServiceCloudSource from .oracle_source import OracleSource from .azure_my_sql_source import AzureMySqlSource from .distcp_settings import DistcpSettings @@ -664,7 +641,6 @@ from .sap_cloud_for_customer_source import SapCloudForCustomerSource from .salesforce_source import SalesforceSource from .relational_source import RelationalSource - from .dynamics_ax_source import DynamicsAXSource from .dynamics_source import DynamicsSource from .document_db_collection_source import DocumentDbCollectionSource from .blob_source import BlobSource @@ -774,15 +750,12 @@ HiveAuthenticationType, HBaseAuthenticationType, GoogleBigQueryAuthenticationType, - GoogleAdWordsAuthenticationType, SapHanaAuthenticationType, SftpAuthenticationType, FtpAuthenticationType, - RestServiceAuthenticationType, HttpAuthenticationType, MongoDbAuthenticationType, ODataAuthenticationType, - ODataAadServicePrincipalCredentialType, TeradataAuthenticationType, Db2AuthenticationType, SybaseAuthenticationType, @@ -917,7 +890,6 @@ 'HBaseLinkedService', 'GreenplumLinkedService', 'GoogleBigQueryLinkedService', - 'GoogleAdWordsLinkedService', 'EloquaLinkedService', 'DrillLinkedService', 'CouchbaseLinkedService', @@ -928,7 +900,6 @@ 'SapBWLinkedService', 'SftpServerLinkedService', 'FtpServerLinkedService', - 'RestServiceLinkedService', 'HttpLinkedService', 'AzureSearchLinkedService', 'CustomDataSourceLinkedService', @@ -955,11 +926,9 @@ 'PostgreSqlLinkedService', 'MySqlLinkedService', 'AzureMySqlLinkedService', - 'OracleServiceCloudLinkedService', 'OracleLinkedService', 'FileServerLinkedService', 'HDInsightLinkedService', - 'DynamicsAXLinkedService', 'DynamicsLinkedService', 'CosmosDbLinkedService', 'AzureKeyVaultLinkedService', @@ -994,14 +963,12 @@ 'HBaseObjectDataset', 'GreenplumTableDataset', 'GoogleBigQueryObjectDataset', - 'GoogleAdWordsObjectDataset', 'EloquaObjectDataset', 'DrillTableDataset', 'CouchbaseTableDataset', 'ConcurObjectDataset', 'AzurePostgreSqlTableDataset', 'AmazonMWSObjectDataset', - 'RestServiceDataset', 'DatasetZipDeflateCompression', 'DatasetDeflateCompression', 'DatasetGZipCompression', @@ -1022,13 +989,11 @@ 'SalesforceObjectDataset', 'RelationalTableDataset', 'AzureMySqlTableDataset', - 'OracleServiceCloudObjectDataset', 'OracleTableDataset', 'ODataResourceDataset', 'MongoDbCollectionDataset', 'FileShareDataset', 'AzureDataLakeStoreDataset', - 'DynamicsAXResourceDataset', 'DynamicsEntityDataset', 'DocumentDbCollectionDataset', 'CustomDataset', @@ -1075,20 +1040,17 @@ 'HBaseSource', 'GreenplumSource', 'GoogleBigQuerySource', - 'GoogleAdWordsSource', 'EloquaSource', 'DrillSource', 'CouchbaseSource', 'ConcurSource', 'AzurePostgreSqlSource', 'AmazonMWSSource', - 'RestServiceSource', 'HttpSource', 'AzureDataLakeStoreSource', 'MongoDbSource', 'CassandraSource', 'WebSource', - 'OracleServiceCloudSource', 'OracleSource', 'AzureMySqlSource', 'DistcpSettings', @@ -1101,7 +1063,6 @@ 'SapCloudForCustomerSource', 'SalesforceSource', 'RelationalSource', - 'DynamicsAXSource', 'DynamicsSource', 'DocumentDbCollectionSource', 'BlobSource', @@ -1210,15 +1171,12 @@ 'HiveAuthenticationType', 'HBaseAuthenticationType', 'GoogleBigQueryAuthenticationType', - 'GoogleAdWordsAuthenticationType', 'SapHanaAuthenticationType', 'SftpAuthenticationType', 'FtpAuthenticationType', - 'RestServiceAuthenticationType', 'HttpAuthenticationType', 'MongoDbAuthenticationType', 'ODataAuthenticationType', - 'ODataAadServicePrincipalCredentialType', 'TeradataAuthenticationType', 'Db2AuthenticationType', 'SybaseAuthenticationType', diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_s3_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_s3_dataset.py index 403e0e7811c4..87540eb0eb11 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_s3_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_s3_dataset.py @@ -48,9 +48,6 @@ class AmazonS3Dataset(Dataset): :param prefix: The prefix filter for the S3 object name. Type: string (or Expression with resultType string). :type prefix: object - :param wildcard_path: The path of the S3 object with wildcard supported. - Type: string (or Expression with resultType string). - :type wildcard_path: object :param version: The version for the S3 object. Type: string (or Expression with resultType string). :type version: object @@ -79,7 +76,6 @@ class AmazonS3Dataset(Dataset): 'bucket_name': {'key': 'typeProperties.bucketName', 'type': 'object'}, 'key': {'key': 'typeProperties.key', 'type': 'object'}, 'prefix': {'key': 'typeProperties.prefix', 'type': 'object'}, - 'wildcard_path': {'key': 'typeProperties.wildcardPath', 'type': 'object'}, 'version': {'key': 'typeProperties.version', 'type': 'object'}, 'format': {'key': 'typeProperties.format', 'type': 'DatasetStorageFormat'}, 'compression': {'key': 'typeProperties.compression', 'type': 'DatasetCompression'}, @@ -90,7 +86,6 @@ def __init__(self, **kwargs): self.bucket_name = kwargs.get('bucket_name', None) self.key = kwargs.get('key', None) self.prefix = kwargs.get('prefix', None) - self.wildcard_path = kwargs.get('wildcard_path', None) self.version = kwargs.get('version', None) self.format = kwargs.get('format', None) self.compression = kwargs.get('compression', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_s3_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_s3_dataset_py3.py index 09a3130c8c70..c4b602da1303 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_s3_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_s3_dataset_py3.py @@ -48,9 +48,6 @@ class AmazonS3Dataset(Dataset): :param prefix: The prefix filter for the S3 object name. Type: string (or Expression with resultType string). :type prefix: object - :param wildcard_path: The path of the S3 object with wildcard supported. - Type: string (or Expression with resultType string). - :type wildcard_path: object :param version: The version for the S3 object. Type: string (or Expression with resultType string). :type version: object @@ -79,18 +76,16 @@ class AmazonS3Dataset(Dataset): 'bucket_name': {'key': 'typeProperties.bucketName', 'type': 'object'}, 'key': {'key': 'typeProperties.key', 'type': 'object'}, 'prefix': {'key': 'typeProperties.prefix', 'type': 'object'}, - 'wildcard_path': {'key': 'typeProperties.wildcardPath', 'type': 'object'}, 'version': {'key': 'typeProperties.version', 'type': 'object'}, 'format': {'key': 'typeProperties.format', 'type': 'DatasetStorageFormat'}, 'compression': {'key': 'typeProperties.compression', 'type': 'DatasetCompression'}, } - def __init__(self, *, linked_service_name, bucket_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, key=None, prefix=None, wildcard_path=None, version=None, format=None, compression=None, **kwargs) -> None: + def __init__(self, *, linked_service_name, bucket_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, key=None, prefix=None, version=None, format=None, compression=None, **kwargs) -> None: super(AmazonS3Dataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.bucket_name = bucket_name self.key = key self.prefix = prefix - self.wildcard_path = wildcard_path self.version = version self.format = format self.compression = compression diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_dataset.py index f3070d3d6b46..dfa31040d658 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_dataset.py @@ -48,10 +48,6 @@ class AzureBlobDataset(Dataset): :param file_name: The name of the Azure Blob. Type: string (or Expression with resultType string). :type file_name: object - :param wildcard_path: The whole path include file name of the Azure Blob - storage with wildcard supported. Type: string (or Expression with - resultType string). - :type wildcard_path: object :param format: The format of the Azure Blob storage. :type format: ~azure.mgmt.datafactory.models.DatasetStorageFormat :param compression: The data compression method used for the blob storage. @@ -75,7 +71,6 @@ class AzureBlobDataset(Dataset): 'folder_path': {'key': 'typeProperties.folderPath', 'type': 'object'}, 'table_root_location': {'key': 'typeProperties.tableRootLocation', 'type': 'object'}, 'file_name': {'key': 'typeProperties.fileName', 'type': 'object'}, - 'wildcard_path': {'key': 'typeProperties.wildcardPath', 'type': 'object'}, 'format': {'key': 'typeProperties.format', 'type': 'DatasetStorageFormat'}, 'compression': {'key': 'typeProperties.compression', 'type': 'DatasetCompression'}, } @@ -85,7 +80,6 @@ def __init__(self, **kwargs): self.folder_path = kwargs.get('folder_path', None) self.table_root_location = kwargs.get('table_root_location', None) self.file_name = kwargs.get('file_name', None) - self.wildcard_path = kwargs.get('wildcard_path', None) self.format = kwargs.get('format', None) self.compression = kwargs.get('compression', None) self.type = 'AzureBlob' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_dataset_py3.py index f84716fd9548..adef2dbd6def 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_dataset_py3.py @@ -48,10 +48,6 @@ class AzureBlobDataset(Dataset): :param file_name: The name of the Azure Blob. Type: string (or Expression with resultType string). :type file_name: object - :param wildcard_path: The whole path include file name of the Azure Blob - storage with wildcard supported. Type: string (or Expression with - resultType string). - :type wildcard_path: object :param format: The format of the Azure Blob storage. :type format: ~azure.mgmt.datafactory.models.DatasetStorageFormat :param compression: The data compression method used for the blob storage. @@ -75,17 +71,15 @@ class AzureBlobDataset(Dataset): 'folder_path': {'key': 'typeProperties.folderPath', 'type': 'object'}, 'table_root_location': {'key': 'typeProperties.tableRootLocation', 'type': 'object'}, 'file_name': {'key': 'typeProperties.fileName', 'type': 'object'}, - 'wildcard_path': {'key': 'typeProperties.wildcardPath', 'type': 'object'}, 'format': {'key': 'typeProperties.format', 'type': 'DatasetStorageFormat'}, 'compression': {'key': 'typeProperties.compression', 'type': 'DatasetCompression'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, folder_path=None, table_root_location=None, file_name=None, wildcard_path=None, format=None, compression=None, **kwargs) -> None: + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, folder_path=None, table_root_location=None, file_name=None, format=None, compression=None, **kwargs) -> None: super(AzureBlobDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.folder_path = folder_path self.table_root_location = table_root_location self.file_name = file_name - self.wildcard_path = wildcard_path self.format = format self.compression = compression self.type = 'AzureBlob' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_dataset.py index 5dccec2d1bfe..0d3d9edf45d9 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_dataset.py @@ -39,18 +39,14 @@ class AzureDataLakeStoreDataset(Dataset): :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str - :param folder_path: Path to the folder in the Azure Data Lake Store. Type: - string (or Expression with resultType string). + :param folder_path: Required. Path to the folder in the Azure Data Lake + Store. Type: string (or Expression with resultType string). :type folder_path: object :param file_name: The name of the file in the Azure Data Lake Store. Type: string (or Expression with resultType string). :type file_name: object :param format: The format of the Data Lake Store. :type format: ~azure.mgmt.datafactory.models.DatasetStorageFormat - :param wildcard_path: The whole path include file name in the Azure Data - Lake Store with wildcard supported. Type: string (or Expression with - resultType string). - :type wildcard_path: object :param compression: The data compression method used for the item(s) in the Azure Data Lake Store. :type compression: ~azure.mgmt.datafactory.models.DatasetCompression @@ -59,6 +55,7 @@ class AzureDataLakeStoreDataset(Dataset): _validation = { 'linked_service_name': {'required': True}, 'type': {'required': True}, + 'folder_path': {'required': True}, } _attribute_map = { @@ -73,7 +70,6 @@ class AzureDataLakeStoreDataset(Dataset): 'folder_path': {'key': 'typeProperties.folderPath', 'type': 'object'}, 'file_name': {'key': 'typeProperties.fileName', 'type': 'object'}, 'format': {'key': 'typeProperties.format', 'type': 'DatasetStorageFormat'}, - 'wildcard_path': {'key': 'typeProperties.wildcardPath', 'type': 'object'}, 'compression': {'key': 'typeProperties.compression', 'type': 'DatasetCompression'}, } @@ -82,6 +78,5 @@ def __init__(self, **kwargs): self.folder_path = kwargs.get('folder_path', None) self.file_name = kwargs.get('file_name', None) self.format = kwargs.get('format', None) - self.wildcard_path = kwargs.get('wildcard_path', None) self.compression = kwargs.get('compression', None) self.type = 'AzureDataLakeStoreFile' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_dataset_py3.py index 98b5432f9cf3..8d108c944dcb 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_dataset_py3.py @@ -39,18 +39,14 @@ class AzureDataLakeStoreDataset(Dataset): :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str - :param folder_path: Path to the folder in the Azure Data Lake Store. Type: - string (or Expression with resultType string). + :param folder_path: Required. Path to the folder in the Azure Data Lake + Store. Type: string (or Expression with resultType string). :type folder_path: object :param file_name: The name of the file in the Azure Data Lake Store. Type: string (or Expression with resultType string). :type file_name: object :param format: The format of the Data Lake Store. :type format: ~azure.mgmt.datafactory.models.DatasetStorageFormat - :param wildcard_path: The whole path include file name in the Azure Data - Lake Store with wildcard supported. Type: string (or Expression with - resultType string). - :type wildcard_path: object :param compression: The data compression method used for the item(s) in the Azure Data Lake Store. :type compression: ~azure.mgmt.datafactory.models.DatasetCompression @@ -59,6 +55,7 @@ class AzureDataLakeStoreDataset(Dataset): _validation = { 'linked_service_name': {'required': True}, 'type': {'required': True}, + 'folder_path': {'required': True}, } _attribute_map = { @@ -73,15 +70,13 @@ class AzureDataLakeStoreDataset(Dataset): 'folder_path': {'key': 'typeProperties.folderPath', 'type': 'object'}, 'file_name': {'key': 'typeProperties.fileName', 'type': 'object'}, 'format': {'key': 'typeProperties.format', 'type': 'DatasetStorageFormat'}, - 'wildcard_path': {'key': 'typeProperties.wildcardPath', 'type': 'object'}, 'compression': {'key': 'typeProperties.compression', 'type': 'DatasetCompression'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, folder_path=None, file_name=None, format=None, wildcard_path=None, compression=None, **kwargs) -> None: + def __init__(self, *, linked_service_name, folder_path, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, file_name=None, format=None, compression=None, **kwargs) -> None: super(AzureDataLakeStoreDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.folder_path = folder_path self.file_name = file_name self.format = format - self.wildcard_path = wildcard_path self.compression = compression self.type = 'AzureDataLakeStoreFile' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_source.py index 8ad510c03d95..9a11107fc8e8 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_source.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_source.py @@ -22,13 +22,12 @@ class CopySource(Model): QuickBooksSource, PrestoSource, PhoenixSource, PaypalSource, MarketoSource, MariaDBSource, MagentoSource, JiraSource, ImpalaSource, HubspotSource, HiveSource, HBaseSource, GreenplumSource, GoogleBigQuerySource, - GoogleAdWordsSource, EloquaSource, DrillSource, CouchbaseSource, - ConcurSource, AzurePostgreSqlSource, AmazonMWSSource, RestServiceSource, - HttpSource, AzureDataLakeStoreSource, MongoDbSource, CassandraSource, - WebSource, OracleServiceCloudSource, OracleSource, AzureMySqlSource, - HdfsSource, FileSystemSource, SqlDWSource, SqlSource, SapEccSource, - SapCloudForCustomerSource, SalesforceSource, RelationalSource, - DynamicsAXSource, DynamicsSource, DocumentDbCollectionSource, BlobSource, + EloquaSource, DrillSource, CouchbaseSource, ConcurSource, + AzurePostgreSqlSource, AmazonMWSSource, HttpSource, + AzureDataLakeStoreSource, MongoDbSource, CassandraSource, WebSource, + OracleSource, AzureMySqlSource, HdfsSource, FileSystemSource, SqlDWSource, + SqlSource, SapEccSource, SapCloudForCustomerSource, SalesforceSource, + RelationalSource, DynamicsSource, DocumentDbCollectionSource, BlobSource, AzureTableSource All required parameters must be populated in order to send to Azure. @@ -59,7 +58,7 @@ class CopySource(Model): } _subtype_map = { - 'type': {'AmazonRedshiftSource': 'AmazonRedshiftSource', 'ResponsysSource': 'ResponsysSource', 'SalesforceMarketingCloudSource': 'SalesforceMarketingCloudSource', 'VerticaSource': 'VerticaSource', 'NetezzaSource': 'NetezzaSource', 'ZohoSource': 'ZohoSource', 'XeroSource': 'XeroSource', 'SquareSource': 'SquareSource', 'SparkSource': 'SparkSource', 'ShopifySource': 'ShopifySource', 'ServiceNowSource': 'ServiceNowSource', 'QuickBooksSource': 'QuickBooksSource', 'PrestoSource': 'PrestoSource', 'PhoenixSource': 'PhoenixSource', 'PaypalSource': 'PaypalSource', 'MarketoSource': 'MarketoSource', 'MariaDBSource': 'MariaDBSource', 'MagentoSource': 'MagentoSource', 'JiraSource': 'JiraSource', 'ImpalaSource': 'ImpalaSource', 'HubspotSource': 'HubspotSource', 'HiveSource': 'HiveSource', 'HBaseSource': 'HBaseSource', 'GreenplumSource': 'GreenplumSource', 'GoogleBigQuerySource': 'GoogleBigQuerySource', 'GoogleAdWordsSource': 'GoogleAdWordsSource', 'EloquaSource': 'EloquaSource', 'DrillSource': 'DrillSource', 'CouchbaseSource': 'CouchbaseSource', 'ConcurSource': 'ConcurSource', 'AzurePostgreSqlSource': 'AzurePostgreSqlSource', 'AmazonMWSSource': 'AmazonMWSSource', 'RestServiceSource': 'RestServiceSource', 'HttpSource': 'HttpSource', 'AzureDataLakeStoreSource': 'AzureDataLakeStoreSource', 'MongoDbSource': 'MongoDbSource', 'CassandraSource': 'CassandraSource', 'WebSource': 'WebSource', 'OracleServiceCloudSource': 'OracleServiceCloudSource', 'OracleSource': 'OracleSource', 'AzureMySqlSource': 'AzureMySqlSource', 'HdfsSource': 'HdfsSource', 'FileSystemSource': 'FileSystemSource', 'SqlDWSource': 'SqlDWSource', 'SqlSource': 'SqlSource', 'SapEccSource': 'SapEccSource', 'SapCloudForCustomerSource': 'SapCloudForCustomerSource', 'SalesforceSource': 'SalesforceSource', 'RelationalSource': 'RelationalSource', 'DynamicsAXSource': 'DynamicsAXSource', 'DynamicsSource': 'DynamicsSource', 'DocumentDbCollectionSource': 'DocumentDbCollectionSource', 'BlobSource': 'BlobSource', 'AzureTableSource': 'AzureTableSource'} + 'type': {'AmazonRedshiftSource': 'AmazonRedshiftSource', 'ResponsysSource': 'ResponsysSource', 'SalesforceMarketingCloudSource': 'SalesforceMarketingCloudSource', 'VerticaSource': 'VerticaSource', 'NetezzaSource': 'NetezzaSource', 'ZohoSource': 'ZohoSource', 'XeroSource': 'XeroSource', 'SquareSource': 'SquareSource', 'SparkSource': 'SparkSource', 'ShopifySource': 'ShopifySource', 'ServiceNowSource': 'ServiceNowSource', 'QuickBooksSource': 'QuickBooksSource', 'PrestoSource': 'PrestoSource', 'PhoenixSource': 'PhoenixSource', 'PaypalSource': 'PaypalSource', 'MarketoSource': 'MarketoSource', 'MariaDBSource': 'MariaDBSource', 'MagentoSource': 'MagentoSource', 'JiraSource': 'JiraSource', 'ImpalaSource': 'ImpalaSource', 'HubspotSource': 'HubspotSource', 'HiveSource': 'HiveSource', 'HBaseSource': 'HBaseSource', 'GreenplumSource': 'GreenplumSource', 'GoogleBigQuerySource': 'GoogleBigQuerySource', 'EloquaSource': 'EloquaSource', 'DrillSource': 'DrillSource', 'CouchbaseSource': 'CouchbaseSource', 'ConcurSource': 'ConcurSource', 'AzurePostgreSqlSource': 'AzurePostgreSqlSource', 'AmazonMWSSource': 'AmazonMWSSource', 'HttpSource': 'HttpSource', 'AzureDataLakeStoreSource': 'AzureDataLakeStoreSource', 'MongoDbSource': 'MongoDbSource', 'CassandraSource': 'CassandraSource', 'WebSource': 'WebSource', 'OracleSource': 'OracleSource', 'AzureMySqlSource': 'AzureMySqlSource', 'HdfsSource': 'HdfsSource', 'FileSystemSource': 'FileSystemSource', 'SqlDWSource': 'SqlDWSource', 'SqlSource': 'SqlSource', 'SapEccSource': 'SapEccSource', 'SapCloudForCustomerSource': 'SapCloudForCustomerSource', 'SalesforceSource': 'SalesforceSource', 'RelationalSource': 'RelationalSource', 'DynamicsSource': 'DynamicsSource', 'DocumentDbCollectionSource': 'DocumentDbCollectionSource', 'BlobSource': 'BlobSource', 'AzureTableSource': 'AzureTableSource'} } def __init__(self, **kwargs): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_source_py3.py index b2d6beed13ce..7c1a96b2897a 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_source_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_source_py3.py @@ -22,13 +22,12 @@ class CopySource(Model): QuickBooksSource, PrestoSource, PhoenixSource, PaypalSource, MarketoSource, MariaDBSource, MagentoSource, JiraSource, ImpalaSource, HubspotSource, HiveSource, HBaseSource, GreenplumSource, GoogleBigQuerySource, - GoogleAdWordsSource, EloquaSource, DrillSource, CouchbaseSource, - ConcurSource, AzurePostgreSqlSource, AmazonMWSSource, RestServiceSource, - HttpSource, AzureDataLakeStoreSource, MongoDbSource, CassandraSource, - WebSource, OracleServiceCloudSource, OracleSource, AzureMySqlSource, - HdfsSource, FileSystemSource, SqlDWSource, SqlSource, SapEccSource, - SapCloudForCustomerSource, SalesforceSource, RelationalSource, - DynamicsAXSource, DynamicsSource, DocumentDbCollectionSource, BlobSource, + EloquaSource, DrillSource, CouchbaseSource, ConcurSource, + AzurePostgreSqlSource, AmazonMWSSource, HttpSource, + AzureDataLakeStoreSource, MongoDbSource, CassandraSource, WebSource, + OracleSource, AzureMySqlSource, HdfsSource, FileSystemSource, SqlDWSource, + SqlSource, SapEccSource, SapCloudForCustomerSource, SalesforceSource, + RelationalSource, DynamicsSource, DocumentDbCollectionSource, BlobSource, AzureTableSource All required parameters must be populated in order to send to Azure. @@ -59,7 +58,7 @@ class CopySource(Model): } _subtype_map = { - 'type': {'AmazonRedshiftSource': 'AmazonRedshiftSource', 'ResponsysSource': 'ResponsysSource', 'SalesforceMarketingCloudSource': 'SalesforceMarketingCloudSource', 'VerticaSource': 'VerticaSource', 'NetezzaSource': 'NetezzaSource', 'ZohoSource': 'ZohoSource', 'XeroSource': 'XeroSource', 'SquareSource': 'SquareSource', 'SparkSource': 'SparkSource', 'ShopifySource': 'ShopifySource', 'ServiceNowSource': 'ServiceNowSource', 'QuickBooksSource': 'QuickBooksSource', 'PrestoSource': 'PrestoSource', 'PhoenixSource': 'PhoenixSource', 'PaypalSource': 'PaypalSource', 'MarketoSource': 'MarketoSource', 'MariaDBSource': 'MariaDBSource', 'MagentoSource': 'MagentoSource', 'JiraSource': 'JiraSource', 'ImpalaSource': 'ImpalaSource', 'HubspotSource': 'HubspotSource', 'HiveSource': 'HiveSource', 'HBaseSource': 'HBaseSource', 'GreenplumSource': 'GreenplumSource', 'GoogleBigQuerySource': 'GoogleBigQuerySource', 'GoogleAdWordsSource': 'GoogleAdWordsSource', 'EloquaSource': 'EloquaSource', 'DrillSource': 'DrillSource', 'CouchbaseSource': 'CouchbaseSource', 'ConcurSource': 'ConcurSource', 'AzurePostgreSqlSource': 'AzurePostgreSqlSource', 'AmazonMWSSource': 'AmazonMWSSource', 'RestServiceSource': 'RestServiceSource', 'HttpSource': 'HttpSource', 'AzureDataLakeStoreSource': 'AzureDataLakeStoreSource', 'MongoDbSource': 'MongoDbSource', 'CassandraSource': 'CassandraSource', 'WebSource': 'WebSource', 'OracleServiceCloudSource': 'OracleServiceCloudSource', 'OracleSource': 'OracleSource', 'AzureMySqlSource': 'AzureMySqlSource', 'HdfsSource': 'HdfsSource', 'FileSystemSource': 'FileSystemSource', 'SqlDWSource': 'SqlDWSource', 'SqlSource': 'SqlSource', 'SapEccSource': 'SapEccSource', 'SapCloudForCustomerSource': 'SapCloudForCustomerSource', 'SalesforceSource': 'SalesforceSource', 'RelationalSource': 'RelationalSource', 'DynamicsAXSource': 'DynamicsAXSource', 'DynamicsSource': 'DynamicsSource', 'DocumentDbCollectionSource': 'DocumentDbCollectionSource', 'BlobSource': 'BlobSource', 'AzureTableSource': 'AzureTableSource'} + 'type': {'AmazonRedshiftSource': 'AmazonRedshiftSource', 'ResponsysSource': 'ResponsysSource', 'SalesforceMarketingCloudSource': 'SalesforceMarketingCloudSource', 'VerticaSource': 'VerticaSource', 'NetezzaSource': 'NetezzaSource', 'ZohoSource': 'ZohoSource', 'XeroSource': 'XeroSource', 'SquareSource': 'SquareSource', 'SparkSource': 'SparkSource', 'ShopifySource': 'ShopifySource', 'ServiceNowSource': 'ServiceNowSource', 'QuickBooksSource': 'QuickBooksSource', 'PrestoSource': 'PrestoSource', 'PhoenixSource': 'PhoenixSource', 'PaypalSource': 'PaypalSource', 'MarketoSource': 'MarketoSource', 'MariaDBSource': 'MariaDBSource', 'MagentoSource': 'MagentoSource', 'JiraSource': 'JiraSource', 'ImpalaSource': 'ImpalaSource', 'HubspotSource': 'HubspotSource', 'HiveSource': 'HiveSource', 'HBaseSource': 'HBaseSource', 'GreenplumSource': 'GreenplumSource', 'GoogleBigQuerySource': 'GoogleBigQuerySource', 'EloquaSource': 'EloquaSource', 'DrillSource': 'DrillSource', 'CouchbaseSource': 'CouchbaseSource', 'ConcurSource': 'ConcurSource', 'AzurePostgreSqlSource': 'AzurePostgreSqlSource', 'AmazonMWSSource': 'AmazonMWSSource', 'HttpSource': 'HttpSource', 'AzureDataLakeStoreSource': 'AzureDataLakeStoreSource', 'MongoDbSource': 'MongoDbSource', 'CassandraSource': 'CassandraSource', 'WebSource': 'WebSource', 'OracleSource': 'OracleSource', 'AzureMySqlSource': 'AzureMySqlSource', 'HdfsSource': 'HdfsSource', 'FileSystemSource': 'FileSystemSource', 'SqlDWSource': 'SqlDWSource', 'SqlSource': 'SqlSource', 'SapEccSource': 'SapEccSource', 'SapCloudForCustomerSource': 'SapCloudForCustomerSource', 'SalesforceSource': 'SalesforceSource', 'RelationalSource': 'RelationalSource', 'DynamicsSource': 'DynamicsSource', 'DocumentDbCollectionSource': 'DocumentDbCollectionSource', 'BlobSource': 'BlobSource', 'AzureTableSource': 'AzureTableSource'} } def __init__(self, *, additional_properties=None, source_retry_count=None, source_retry_wait=None, **kwargs) -> None: diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_factory_management_client_enums.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_factory_management_client_enums.py index 2b837286cc12..aa76fe05ddc0 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_factory_management_client_enums.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_factory_management_client_enums.py @@ -240,12 +240,6 @@ class GoogleBigQueryAuthenticationType(str, Enum): user_authentication = "UserAuthentication" -class GoogleAdWordsAuthenticationType(str, Enum): - - service_authentication = "ServiceAuthentication" - user_authentication = "UserAuthentication" - - class SapHanaAuthenticationType(str, Enum): basic = "Basic" @@ -264,14 +258,6 @@ class FtpAuthenticationType(str, Enum): anonymous = "Anonymous" -class RestServiceAuthenticationType(str, Enum): - - anonymous = "Anonymous" - basic = "Basic" - aad_service_princial = "AadServicePrincial" - managed_service_identity = "ManagedServiceIdentity" - - class HttpAuthenticationType(str, Enum): basic = "Basic" @@ -289,17 +275,8 @@ class MongoDbAuthenticationType(str, Enum): class ODataAuthenticationType(str, Enum): - anonymous = "Anonymous" basic = "Basic" - windows = "Windows" - aad_service_principal = "AadServicePrincipal" - managed_service_identity = "ManagedServiceIdentity" - - -class ODataAadServicePrincipalCredentialType(str, Enum): - - service_principal_key = "ServicePrincipalKey" - service_principal_cert = "ServicePrincipalCert" + anonymous = "Anonymous" class TeradataAuthenticationType(str, Enum): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset.py index 2cefbb683276..626015fd5d11 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset.py @@ -26,15 +26,14 @@ class Dataset(Model): MariaDBTableDataset, MagentoObjectDataset, JiraObjectDataset, ImpalaObjectDataset, HubspotObjectDataset, HiveObjectDataset, HBaseObjectDataset, GreenplumTableDataset, GoogleBigQueryObjectDataset, - GoogleAdWordsObjectDataset, EloquaObjectDataset, DrillTableDataset, - CouchbaseTableDataset, ConcurObjectDataset, AzurePostgreSqlTableDataset, - AmazonMWSObjectDataset, RestServiceDataset, HttpDataset, - AzureSearchIndexDataset, WebTableDataset, SqlServerTableDataset, - SapEccResourceDataset, SapCloudForCustomerResourceDataset, - SalesforceObjectDataset, RelationalTableDataset, AzureMySqlTableDataset, - OracleServiceCloudObjectDataset, OracleTableDataset, ODataResourceDataset, - MongoDbCollectionDataset, FileShareDataset, AzureDataLakeStoreDataset, - DynamicsAXResourceDataset, DynamicsEntityDataset, + EloquaObjectDataset, DrillTableDataset, CouchbaseTableDataset, + ConcurObjectDataset, AzurePostgreSqlTableDataset, AmazonMWSObjectDataset, + HttpDataset, AzureSearchIndexDataset, WebTableDataset, + SqlServerTableDataset, SapEccResourceDataset, + SapCloudForCustomerResourceDataset, SalesforceObjectDataset, + RelationalTableDataset, AzureMySqlTableDataset, OracleTableDataset, + ODataResourceDataset, MongoDbCollectionDataset, FileShareDataset, + AzureDataLakeStoreDataset, DynamicsEntityDataset, DocumentDbCollectionDataset, CustomDataset, CassandraTableDataset, AzureSqlDWTableDataset, AzureSqlTableDataset, AzureTableDataset, AzureBlobDataset, AmazonS3Dataset @@ -82,7 +81,7 @@ class Dataset(Model): } _subtype_map = { - 'type': {'ResponsysObject': 'ResponsysObjectDataset', 'SalesforceMarketingCloudObject': 'SalesforceMarketingCloudObjectDataset', 'VerticaTable': 'VerticaTableDataset', 'NetezzaTable': 'NetezzaTableDataset', 'ZohoObject': 'ZohoObjectDataset', 'XeroObject': 'XeroObjectDataset', 'SquareObject': 'SquareObjectDataset', 'SparkObject': 'SparkObjectDataset', 'ShopifyObject': 'ShopifyObjectDataset', 'ServiceNowObject': 'ServiceNowObjectDataset', 'QuickBooksObject': 'QuickBooksObjectDataset', 'PrestoObject': 'PrestoObjectDataset', 'PhoenixObject': 'PhoenixObjectDataset', 'PaypalObject': 'PaypalObjectDataset', 'MarketoObject': 'MarketoObjectDataset', 'MariaDBTable': 'MariaDBTableDataset', 'MagentoObject': 'MagentoObjectDataset', 'JiraObject': 'JiraObjectDataset', 'ImpalaObject': 'ImpalaObjectDataset', 'HubspotObject': 'HubspotObjectDataset', 'HiveObject': 'HiveObjectDataset', 'HBaseObject': 'HBaseObjectDataset', 'GreenplumTable': 'GreenplumTableDataset', 'GoogleBigQueryObject': 'GoogleBigQueryObjectDataset', 'GoogleAdWordsObject': 'GoogleAdWordsObjectDataset', 'EloquaObject': 'EloquaObjectDataset', 'DrillTable': 'DrillTableDataset', 'CouchbaseTable': 'CouchbaseTableDataset', 'ConcurObject': 'ConcurObjectDataset', 'AzurePostgreSqlTable': 'AzurePostgreSqlTableDataset', 'AmazonMWSObject': 'AmazonMWSObjectDataset', 'RestResource': 'RestServiceDataset', 'HttpFile': 'HttpDataset', 'AzureSearchIndex': 'AzureSearchIndexDataset', 'WebTable': 'WebTableDataset', 'SqlServerTable': 'SqlServerTableDataset', 'SapEccResource': 'SapEccResourceDataset', 'SapCloudForCustomerResource': 'SapCloudForCustomerResourceDataset', 'SalesforceObject': 'SalesforceObjectDataset', 'RelationalTable': 'RelationalTableDataset', 'AzureMySqlTable': 'AzureMySqlTableDataset', 'OracleServiceCloudObject': 'OracleServiceCloudObjectDataset', 'OracleTable': 'OracleTableDataset', 'ODataResource': 'ODataResourceDataset', 'MongoDbCollection': 'MongoDbCollectionDataset', 'FileShare': 'FileShareDataset', 'AzureDataLakeStoreFile': 'AzureDataLakeStoreDataset', 'DynamicsAXResource': 'DynamicsAXResourceDataset', 'DynamicsEntity': 'DynamicsEntityDataset', 'DocumentDbCollection': 'DocumentDbCollectionDataset', 'CustomDataset': 'CustomDataset', 'CassandraTable': 'CassandraTableDataset', 'AzureSqlDWTable': 'AzureSqlDWTableDataset', 'AzureSqlTable': 'AzureSqlTableDataset', 'AzureTable': 'AzureTableDataset', 'AzureBlob': 'AzureBlobDataset', 'AmazonS3Object': 'AmazonS3Dataset'} + 'type': {'ResponsysObject': 'ResponsysObjectDataset', 'SalesforceMarketingCloudObject': 'SalesforceMarketingCloudObjectDataset', 'VerticaTable': 'VerticaTableDataset', 'NetezzaTable': 'NetezzaTableDataset', 'ZohoObject': 'ZohoObjectDataset', 'XeroObject': 'XeroObjectDataset', 'SquareObject': 'SquareObjectDataset', 'SparkObject': 'SparkObjectDataset', 'ShopifyObject': 'ShopifyObjectDataset', 'ServiceNowObject': 'ServiceNowObjectDataset', 'QuickBooksObject': 'QuickBooksObjectDataset', 'PrestoObject': 'PrestoObjectDataset', 'PhoenixObject': 'PhoenixObjectDataset', 'PaypalObject': 'PaypalObjectDataset', 'MarketoObject': 'MarketoObjectDataset', 'MariaDBTable': 'MariaDBTableDataset', 'MagentoObject': 'MagentoObjectDataset', 'JiraObject': 'JiraObjectDataset', 'ImpalaObject': 'ImpalaObjectDataset', 'HubspotObject': 'HubspotObjectDataset', 'HiveObject': 'HiveObjectDataset', 'HBaseObject': 'HBaseObjectDataset', 'GreenplumTable': 'GreenplumTableDataset', 'GoogleBigQueryObject': 'GoogleBigQueryObjectDataset', 'EloquaObject': 'EloquaObjectDataset', 'DrillTable': 'DrillTableDataset', 'CouchbaseTable': 'CouchbaseTableDataset', 'ConcurObject': 'ConcurObjectDataset', 'AzurePostgreSqlTable': 'AzurePostgreSqlTableDataset', 'AmazonMWSObject': 'AmazonMWSObjectDataset', 'HttpFile': 'HttpDataset', 'AzureSearchIndex': 'AzureSearchIndexDataset', 'WebTable': 'WebTableDataset', 'SqlServerTable': 'SqlServerTableDataset', 'SapEccResource': 'SapEccResourceDataset', 'SapCloudForCustomerResource': 'SapCloudForCustomerResourceDataset', 'SalesforceObject': 'SalesforceObjectDataset', 'RelationalTable': 'RelationalTableDataset', 'AzureMySqlTable': 'AzureMySqlTableDataset', 'OracleTable': 'OracleTableDataset', 'ODataResource': 'ODataResourceDataset', 'MongoDbCollection': 'MongoDbCollectionDataset', 'FileShare': 'FileShareDataset', 'AzureDataLakeStoreFile': 'AzureDataLakeStoreDataset', 'DynamicsEntity': 'DynamicsEntityDataset', 'DocumentDbCollection': 'DocumentDbCollectionDataset', 'CustomDataset': 'CustomDataset', 'CassandraTable': 'CassandraTableDataset', 'AzureSqlDWTable': 'AzureSqlDWTableDataset', 'AzureSqlTable': 'AzureSqlTableDataset', 'AzureTable': 'AzureTableDataset', 'AzureBlob': 'AzureBlobDataset', 'AmazonS3Object': 'AmazonS3Dataset'} } def __init__(self, **kwargs): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_py3.py index 5551144b5fa6..3d67a1328cd4 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_py3.py @@ -26,15 +26,14 @@ class Dataset(Model): MariaDBTableDataset, MagentoObjectDataset, JiraObjectDataset, ImpalaObjectDataset, HubspotObjectDataset, HiveObjectDataset, HBaseObjectDataset, GreenplumTableDataset, GoogleBigQueryObjectDataset, - GoogleAdWordsObjectDataset, EloquaObjectDataset, DrillTableDataset, - CouchbaseTableDataset, ConcurObjectDataset, AzurePostgreSqlTableDataset, - AmazonMWSObjectDataset, RestServiceDataset, HttpDataset, - AzureSearchIndexDataset, WebTableDataset, SqlServerTableDataset, - SapEccResourceDataset, SapCloudForCustomerResourceDataset, - SalesforceObjectDataset, RelationalTableDataset, AzureMySqlTableDataset, - OracleServiceCloudObjectDataset, OracleTableDataset, ODataResourceDataset, - MongoDbCollectionDataset, FileShareDataset, AzureDataLakeStoreDataset, - DynamicsAXResourceDataset, DynamicsEntityDataset, + EloquaObjectDataset, DrillTableDataset, CouchbaseTableDataset, + ConcurObjectDataset, AzurePostgreSqlTableDataset, AmazonMWSObjectDataset, + HttpDataset, AzureSearchIndexDataset, WebTableDataset, + SqlServerTableDataset, SapEccResourceDataset, + SapCloudForCustomerResourceDataset, SalesforceObjectDataset, + RelationalTableDataset, AzureMySqlTableDataset, OracleTableDataset, + ODataResourceDataset, MongoDbCollectionDataset, FileShareDataset, + AzureDataLakeStoreDataset, DynamicsEntityDataset, DocumentDbCollectionDataset, CustomDataset, CassandraTableDataset, AzureSqlDWTableDataset, AzureSqlTableDataset, AzureTableDataset, AzureBlobDataset, AmazonS3Dataset @@ -82,7 +81,7 @@ class Dataset(Model): } _subtype_map = { - 'type': {'ResponsysObject': 'ResponsysObjectDataset', 'SalesforceMarketingCloudObject': 'SalesforceMarketingCloudObjectDataset', 'VerticaTable': 'VerticaTableDataset', 'NetezzaTable': 'NetezzaTableDataset', 'ZohoObject': 'ZohoObjectDataset', 'XeroObject': 'XeroObjectDataset', 'SquareObject': 'SquareObjectDataset', 'SparkObject': 'SparkObjectDataset', 'ShopifyObject': 'ShopifyObjectDataset', 'ServiceNowObject': 'ServiceNowObjectDataset', 'QuickBooksObject': 'QuickBooksObjectDataset', 'PrestoObject': 'PrestoObjectDataset', 'PhoenixObject': 'PhoenixObjectDataset', 'PaypalObject': 'PaypalObjectDataset', 'MarketoObject': 'MarketoObjectDataset', 'MariaDBTable': 'MariaDBTableDataset', 'MagentoObject': 'MagentoObjectDataset', 'JiraObject': 'JiraObjectDataset', 'ImpalaObject': 'ImpalaObjectDataset', 'HubspotObject': 'HubspotObjectDataset', 'HiveObject': 'HiveObjectDataset', 'HBaseObject': 'HBaseObjectDataset', 'GreenplumTable': 'GreenplumTableDataset', 'GoogleBigQueryObject': 'GoogleBigQueryObjectDataset', 'GoogleAdWordsObject': 'GoogleAdWordsObjectDataset', 'EloquaObject': 'EloquaObjectDataset', 'DrillTable': 'DrillTableDataset', 'CouchbaseTable': 'CouchbaseTableDataset', 'ConcurObject': 'ConcurObjectDataset', 'AzurePostgreSqlTable': 'AzurePostgreSqlTableDataset', 'AmazonMWSObject': 'AmazonMWSObjectDataset', 'RestResource': 'RestServiceDataset', 'HttpFile': 'HttpDataset', 'AzureSearchIndex': 'AzureSearchIndexDataset', 'WebTable': 'WebTableDataset', 'SqlServerTable': 'SqlServerTableDataset', 'SapEccResource': 'SapEccResourceDataset', 'SapCloudForCustomerResource': 'SapCloudForCustomerResourceDataset', 'SalesforceObject': 'SalesforceObjectDataset', 'RelationalTable': 'RelationalTableDataset', 'AzureMySqlTable': 'AzureMySqlTableDataset', 'OracleServiceCloudObject': 'OracleServiceCloudObjectDataset', 'OracleTable': 'OracleTableDataset', 'ODataResource': 'ODataResourceDataset', 'MongoDbCollection': 'MongoDbCollectionDataset', 'FileShare': 'FileShareDataset', 'AzureDataLakeStoreFile': 'AzureDataLakeStoreDataset', 'DynamicsAXResource': 'DynamicsAXResourceDataset', 'DynamicsEntity': 'DynamicsEntityDataset', 'DocumentDbCollection': 'DocumentDbCollectionDataset', 'CustomDataset': 'CustomDataset', 'CassandraTable': 'CassandraTableDataset', 'AzureSqlDWTable': 'AzureSqlDWTableDataset', 'AzureSqlTable': 'AzureSqlTableDataset', 'AzureTable': 'AzureTableDataset', 'AzureBlob': 'AzureBlobDataset', 'AmazonS3Object': 'AmazonS3Dataset'} + 'type': {'ResponsysObject': 'ResponsysObjectDataset', 'SalesforceMarketingCloudObject': 'SalesforceMarketingCloudObjectDataset', 'VerticaTable': 'VerticaTableDataset', 'NetezzaTable': 'NetezzaTableDataset', 'ZohoObject': 'ZohoObjectDataset', 'XeroObject': 'XeroObjectDataset', 'SquareObject': 'SquareObjectDataset', 'SparkObject': 'SparkObjectDataset', 'ShopifyObject': 'ShopifyObjectDataset', 'ServiceNowObject': 'ServiceNowObjectDataset', 'QuickBooksObject': 'QuickBooksObjectDataset', 'PrestoObject': 'PrestoObjectDataset', 'PhoenixObject': 'PhoenixObjectDataset', 'PaypalObject': 'PaypalObjectDataset', 'MarketoObject': 'MarketoObjectDataset', 'MariaDBTable': 'MariaDBTableDataset', 'MagentoObject': 'MagentoObjectDataset', 'JiraObject': 'JiraObjectDataset', 'ImpalaObject': 'ImpalaObjectDataset', 'HubspotObject': 'HubspotObjectDataset', 'HiveObject': 'HiveObjectDataset', 'HBaseObject': 'HBaseObjectDataset', 'GreenplumTable': 'GreenplumTableDataset', 'GoogleBigQueryObject': 'GoogleBigQueryObjectDataset', 'EloquaObject': 'EloquaObjectDataset', 'DrillTable': 'DrillTableDataset', 'CouchbaseTable': 'CouchbaseTableDataset', 'ConcurObject': 'ConcurObjectDataset', 'AzurePostgreSqlTable': 'AzurePostgreSqlTableDataset', 'AmazonMWSObject': 'AmazonMWSObjectDataset', 'HttpFile': 'HttpDataset', 'AzureSearchIndex': 'AzureSearchIndexDataset', 'WebTable': 'WebTableDataset', 'SqlServerTable': 'SqlServerTableDataset', 'SapEccResource': 'SapEccResourceDataset', 'SapCloudForCustomerResource': 'SapCloudForCustomerResourceDataset', 'SalesforceObject': 'SalesforceObjectDataset', 'RelationalTable': 'RelationalTableDataset', 'AzureMySqlTable': 'AzureMySqlTableDataset', 'OracleTable': 'OracleTableDataset', 'ODataResource': 'ODataResourceDataset', 'MongoDbCollection': 'MongoDbCollectionDataset', 'FileShare': 'FileShareDataset', 'AzureDataLakeStoreFile': 'AzureDataLakeStoreDataset', 'DynamicsEntity': 'DynamicsEntityDataset', 'DocumentDbCollection': 'DocumentDbCollectionDataset', 'CustomDataset': 'CustomDataset', 'CassandraTable': 'CassandraTableDataset', 'AzureSqlDWTable': 'AzureSqlDWTableDataset', 'AzureSqlTable': 'AzureSqlTableDataset', 'AzureTable': 'AzureTableDataset', 'AzureBlob': 'AzureBlobDataset', 'AmazonS3Object': 'AmazonS3Dataset'} } def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_ax_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_ax_linked_service.py deleted file mode 100644 index f28a5071f400..000000000000 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_ax_linked_service.py +++ /dev/null @@ -1,90 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for -# license information. -# -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is -# regenerated. -# -------------------------------------------------------------------------- - -from .linked_service import LinkedService - - -class DynamicsAXLinkedService(LinkedService): - """Dynamics AX linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] - :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.mgmt.datafactory.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.mgmt.datafactory.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the - Dataset. - :type annotations: list[object] - :param type: Required. Constant filled by server. - :type type: str - :param url: Required. The URL of Dynamics AX OData API. Type: string (or - Expression with resultType string). - :type url: object - :param tenant: Specify the tenant information (domain name or tenant ID) - under which your application resides. Type: string (or Expression with - resultType string). - :type tenant: object - :param service_principal_id: Specify the application id of your - application registered in Azure Active Directory. Type: string (or - Expression with resultType string). - :type service_principal_id: object - :param aad_resource_id: Specify the resource you are requesting - authorization to use Directory. Type: string (or Expression with - resultType string). - :type aad_resource_id: object - :param service_principal_key: Specify the secret of your application - registered in Azure Active Directory. Type: string (or Expression with - resultType string). - :type service_principal_key: ~azure.mgmt.datafactory.models.SecretBase - :param encrypted_credential: The encrypted credential used for - authentication. Credentials are encrypted using the integration runtime - credential manager. Either encryptedCredential or username/password must - be provided. Type: string (or Expression with resultType string). - :type encrypted_credential: object - """ - - _validation = { - 'type': {'required': True}, - 'url': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'type': {'key': 'type', 'type': 'str'}, - 'url': {'key': 'typeProperties.url', 'type': 'object'}, - 'tenant': {'key': 'typeProperties.tenant', 'type': 'object'}, - 'service_principal_id': {'key': 'typeProperties.servicePrincipalId', 'type': 'object'}, - 'aad_resource_id': {'key': 'typeProperties.aadResourceId', 'type': 'object'}, - 'service_principal_key': {'key': 'typeProperties.servicePrincipalKey', 'type': 'SecretBase'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__(self, **kwargs): - super(DynamicsAXLinkedService, self).__init__(**kwargs) - self.url = kwargs.get('url', None) - self.tenant = kwargs.get('tenant', None) - self.service_principal_id = kwargs.get('service_principal_id', None) - self.aad_resource_id = kwargs.get('aad_resource_id', None) - self.service_principal_key = kwargs.get('service_principal_key', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - self.type = 'DynamicsAX' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_ax_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_ax_linked_service_py3.py deleted file mode 100644 index 3875023c438e..000000000000 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_ax_linked_service_py3.py +++ /dev/null @@ -1,90 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for -# license information. -# -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is -# regenerated. -# -------------------------------------------------------------------------- - -from .linked_service_py3 import LinkedService - - -class DynamicsAXLinkedService(LinkedService): - """Dynamics AX linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] - :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.mgmt.datafactory.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.mgmt.datafactory.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the - Dataset. - :type annotations: list[object] - :param type: Required. Constant filled by server. - :type type: str - :param url: Required. The URL of Dynamics AX OData API. Type: string (or - Expression with resultType string). - :type url: object - :param tenant: Specify the tenant information (domain name or tenant ID) - under which your application resides. Type: string (or Expression with - resultType string). - :type tenant: object - :param service_principal_id: Specify the application id of your - application registered in Azure Active Directory. Type: string (or - Expression with resultType string). - :type service_principal_id: object - :param aad_resource_id: Specify the resource you are requesting - authorization to use Directory. Type: string (or Expression with - resultType string). - :type aad_resource_id: object - :param service_principal_key: Specify the secret of your application - registered in Azure Active Directory. Type: string (or Expression with - resultType string). - :type service_principal_key: ~azure.mgmt.datafactory.models.SecretBase - :param encrypted_credential: The encrypted credential used for - authentication. Credentials are encrypted using the integration runtime - credential manager. Either encryptedCredential or username/password must - be provided. Type: string (or Expression with resultType string). - :type encrypted_credential: object - """ - - _validation = { - 'type': {'required': True}, - 'url': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'type': {'key': 'type', 'type': 'str'}, - 'url': {'key': 'typeProperties.url', 'type': 'object'}, - 'tenant': {'key': 'typeProperties.tenant', 'type': 'object'}, - 'service_principal_id': {'key': 'typeProperties.servicePrincipalId', 'type': 'object'}, - 'aad_resource_id': {'key': 'typeProperties.aadResourceId', 'type': 'object'}, - 'service_principal_key': {'key': 'typeProperties.servicePrincipalKey', 'type': 'SecretBase'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__(self, *, url, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, tenant=None, service_principal_id=None, aad_resource_id=None, service_principal_key=None, encrypted_credential=None, **kwargs) -> None: - super(DynamicsAXLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.url = url - self.tenant = tenant - self.service_principal_id = service_principal_id - self.aad_resource_id = aad_resource_id - self.service_principal_key = service_principal_key - self.encrypted_credential = encrypted_credential - self.type = 'DynamicsAX' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_ax_resource_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_ax_resource_dataset.py deleted file mode 100644 index 2603fb97543c..000000000000 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_ax_resource_dataset.py +++ /dev/null @@ -1,68 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for -# license information. -# -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is -# regenerated. -# -------------------------------------------------------------------------- - -from .dataset import Dataset - - -class DynamicsAXResourceDataset(Dataset): - """The path of the Dynamics AX OData entity. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: - array (or Expression with resultType array), itemType: DatasetDataElement. - :type structure: object - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.mgmt.datafactory.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.mgmt.datafactory.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the - Dataset. - :type annotations: list[object] - :param folder: The folder that this Dataset is in. If not specified, - Dataset will appear at the root level. - :type folder: ~azure.mgmt.datafactory.models.DatasetFolder - :param type: Required. Constant filled by server. - :type type: str - :param path: Required. The path of the Dynamics AX OData entity. Type: - string (or Expression with resultType string). - :type path: str - """ - - _validation = { - 'linked_service_name': {'required': True}, - 'type': {'required': True}, - 'path': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'type': {'key': 'type', 'type': 'str'}, - 'path': {'key': 'typeProperties.path', 'type': 'str'}, - } - - def __init__(self, **kwargs): - super(DynamicsAXResourceDataset, self).__init__(**kwargs) - self.path = kwargs.get('path', None) - self.type = 'DynamicsAXResource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_ax_resource_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_ax_resource_dataset_py3.py deleted file mode 100644 index 9cd8602941e7..000000000000 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_ax_resource_dataset_py3.py +++ /dev/null @@ -1,68 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for -# license information. -# -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is -# regenerated. -# -------------------------------------------------------------------------- - -from .dataset_py3 import Dataset - - -class DynamicsAXResourceDataset(Dataset): - """The path of the Dynamics AX OData entity. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: - array (or Expression with resultType array), itemType: DatasetDataElement. - :type structure: object - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.mgmt.datafactory.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.mgmt.datafactory.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the - Dataset. - :type annotations: list[object] - :param folder: The folder that this Dataset is in. If not specified, - Dataset will appear at the root level. - :type folder: ~azure.mgmt.datafactory.models.DatasetFolder - :param type: Required. Constant filled by server. - :type type: str - :param path: Required. The path of the Dynamics AX OData entity. Type: - string (or Expression with resultType string). - :type path: str - """ - - _validation = { - 'linked_service_name': {'required': True}, - 'type': {'required': True}, - 'path': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'type': {'key': 'type', 'type': 'str'}, - 'path': {'key': 'typeProperties.path', 'type': 'str'}, - } - - def __init__(self, *, linked_service_name, path: str, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: - super(DynamicsAXResourceDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.path = path - self.type = 'DynamicsAXResource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_ax_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_ax_source.py deleted file mode 100644 index 60ef05c0bd0c..000000000000 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_ax_source.py +++ /dev/null @@ -1,52 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for -# license information. -# -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is -# regenerated. -# -------------------------------------------------------------------------- - -from .copy_source import CopySource - - -class DynamicsAXSource(CopySource): - """A copy activity Dynamics AX source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] - :param source_retry_count: Source retry count. Type: integer (or - Expression with resultType integer). - :type source_retry_count: object - :param source_retry_wait: Source retry wait. Type: string (or Expression - with resultType string), pattern: - ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: object - :param type: Required. Constant filled by server. - :type type: str - :param query: A query to retrieve data from source. Type: string (or - Expression with resultType string). - :type query: object - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'type': {'key': 'type', 'type': 'str'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__(self, **kwargs): - super(DynamicsAXSource, self).__init__(**kwargs) - self.query = kwargs.get('query', None) - self.type = 'DynamicsAXSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_ax_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_ax_source_py3.py deleted file mode 100644 index 0aa064facbf2..000000000000 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_ax_source_py3.py +++ /dev/null @@ -1,52 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for -# license information. -# -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is -# regenerated. -# -------------------------------------------------------------------------- - -from .copy_source_py3 import CopySource - - -class DynamicsAXSource(CopySource): - """A copy activity Dynamics AX source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] - :param source_retry_count: Source retry count. Type: integer (or - Expression with resultType integer). - :type source_retry_count: object - :param source_retry_wait: Source retry wait. Type: string (or Expression - with resultType string), pattern: - ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: object - :param type: Required. Constant filled by server. - :type type: str - :param query: A query to retrieve data from source. Type: string (or - Expression with resultType string). - :type query: object - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'type': {'key': 'type', 'type': 'str'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__(self, *, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None, **kwargs) -> None: - super(DynamicsAXSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, **kwargs) - self.query = query - self.type = 'DynamicsAXSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_share_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_share_dataset.py index f28535bb54d3..d5cbf4ac6822 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_share_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_share_dataset.py @@ -45,10 +45,6 @@ class FileShareDataset(Dataset): :param file_name: The name of the on-premises file system. Type: string (or Expression with resultType string). :type file_name: object - :param wildcard_path: The whole path include file name of the on-premises - file system with wildcard supported. Type: string (or Expression with - resultType string). - :type wildcard_path: object :param format: The format of the files. :type format: ~azure.mgmt.datafactory.models.DatasetStorageFormat :param file_filter: Specify a filter to be used to select a subset of @@ -75,7 +71,6 @@ class FileShareDataset(Dataset): 'type': {'key': 'type', 'type': 'str'}, 'folder_path': {'key': 'typeProperties.folderPath', 'type': 'object'}, 'file_name': {'key': 'typeProperties.fileName', 'type': 'object'}, - 'wildcard_path': {'key': 'typeProperties.wildcardPath', 'type': 'object'}, 'format': {'key': 'typeProperties.format', 'type': 'DatasetStorageFormat'}, 'file_filter': {'key': 'typeProperties.fileFilter', 'type': 'object'}, 'compression': {'key': 'typeProperties.compression', 'type': 'DatasetCompression'}, @@ -85,7 +80,6 @@ def __init__(self, **kwargs): super(FileShareDataset, self).__init__(**kwargs) self.folder_path = kwargs.get('folder_path', None) self.file_name = kwargs.get('file_name', None) - self.wildcard_path = kwargs.get('wildcard_path', None) self.format = kwargs.get('format', None) self.file_filter = kwargs.get('file_filter', None) self.compression = kwargs.get('compression', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_share_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_share_dataset_py3.py index ae96956c9290..fe980181543e 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_share_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_share_dataset_py3.py @@ -45,10 +45,6 @@ class FileShareDataset(Dataset): :param file_name: The name of the on-premises file system. Type: string (or Expression with resultType string). :type file_name: object - :param wildcard_path: The whole path include file name of the on-premises - file system with wildcard supported. Type: string (or Expression with - resultType string). - :type wildcard_path: object :param format: The format of the files. :type format: ~azure.mgmt.datafactory.models.DatasetStorageFormat :param file_filter: Specify a filter to be used to select a subset of @@ -75,17 +71,15 @@ class FileShareDataset(Dataset): 'type': {'key': 'type', 'type': 'str'}, 'folder_path': {'key': 'typeProperties.folderPath', 'type': 'object'}, 'file_name': {'key': 'typeProperties.fileName', 'type': 'object'}, - 'wildcard_path': {'key': 'typeProperties.wildcardPath', 'type': 'object'}, 'format': {'key': 'typeProperties.format', 'type': 'DatasetStorageFormat'}, 'file_filter': {'key': 'typeProperties.fileFilter', 'type': 'object'}, 'compression': {'key': 'typeProperties.compression', 'type': 'DatasetCompression'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, folder_path=None, file_name=None, wildcard_path=None, format=None, file_filter=None, compression=None, **kwargs) -> None: + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, folder_path=None, file_name=None, format=None, file_filter=None, compression=None, **kwargs) -> None: super(FileShareDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.folder_path = folder_path self.file_name = file_name - self.wildcard_path = wildcard_path self.format = format self.file_filter = file_filter self.compression = compression diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_ad_words_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_ad_words_linked_service.py deleted file mode 100644 index 9f36df58d576..000000000000 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_ad_words_linked_service.py +++ /dev/null @@ -1,119 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for -# license information. -# -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is -# regenerated. -# -------------------------------------------------------------------------- - -from .linked_service import LinkedService - - -class GoogleAdWordsLinkedService(LinkedService): - """Google Awords service linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] - :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.mgmt.datafactory.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.mgmt.datafactory.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the - Dataset. - :type annotations: list[object] - :param type: Required. Constant filled by server. - :type type: str - :param client_customer_id: Required. The Client customer ID of the AdWords - account that you want to fetch report data for. - :type client_customer_id: object - :param developer_token: Required. The developer token associated with the - manager account that you use to grant access to the AdWords API. - :type developer_token: ~azure.mgmt.datafactory.models.SecretBase - :param authentication_type: Required. The OAuth 2.0 authentication - mechanism used for authentication. ServiceAuthentication can only be used - on self-hosted IR. Possible values include: 'ServiceAuthentication', - 'UserAuthentication' - :type authentication_type: str or - ~azure.mgmt.datafactory.models.GoogleAdWordsAuthenticationType - :param refresh_token: The refresh token obtained from Google for - authorizing access to AdWords for UserAuthentication. - :type refresh_token: ~azure.mgmt.datafactory.models.SecretBase - :param client_id: The client id of the google application used to acquire - the refresh token. - :type client_id: ~azure.mgmt.datafactory.models.SecretBase - :param client_secret: The client secret of the google application used to - acquire the refresh token. - :type client_secret: ~azure.mgmt.datafactory.models.SecretBase - :param email: The service account email ID that is used for - ServiceAuthentication and can only be used on self-hosted IR. - :type email: object - :param key_file_path: The full path to the .p12 key file that is used to - authenticate the service account email address and can only be used on - self-hosted IR. - :type key_file_path: object - :param trusted_cert_path: The full path of the .pem file containing - trusted CA certificates for verifying the server when connecting over SSL. - This property can only be set when using SSL on self-hosted IR. The - default value is the cacerts.pem file installed with the IR. - :type trusted_cert_path: object - :param use_system_trust_store: Specifies whether to use a CA certificate - from the system trust store or from a specified PEM file. The default - value is false. - :type use_system_trust_store: object - :param encrypted_credential: The encrypted credential used for - authentication. Credentials are encrypted using the integration runtime - credential manager. Type: string (or Expression with resultType string). - :type encrypted_credential: object - """ - - _validation = { - 'type': {'required': True}, - 'client_customer_id': {'required': True}, - 'developer_token': {'required': True}, - 'authentication_type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'type': {'key': 'type', 'type': 'str'}, - 'client_customer_id': {'key': 'typeProperties.clientCustomerID', 'type': 'object'}, - 'developer_token': {'key': 'typeProperties.developerToken', 'type': 'SecretBase'}, - 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, - 'refresh_token': {'key': 'typeProperties.refreshToken', 'type': 'SecretBase'}, - 'client_id': {'key': 'typeProperties.clientId', 'type': 'SecretBase'}, - 'client_secret': {'key': 'typeProperties.clientSecret', 'type': 'SecretBase'}, - 'email': {'key': 'typeProperties.email', 'type': 'object'}, - 'key_file_path': {'key': 'typeProperties.keyFilePath', 'type': 'object'}, - 'trusted_cert_path': {'key': 'typeProperties.trustedCertPath', 'type': 'object'}, - 'use_system_trust_store': {'key': 'typeProperties.useSystemTrustStore', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__(self, **kwargs): - super(GoogleAdWordsLinkedService, self).__init__(**kwargs) - self.client_customer_id = kwargs.get('client_customer_id', None) - self.developer_token = kwargs.get('developer_token', None) - self.authentication_type = kwargs.get('authentication_type', None) - self.refresh_token = kwargs.get('refresh_token', None) - self.client_id = kwargs.get('client_id', None) - self.client_secret = kwargs.get('client_secret', None) - self.email = kwargs.get('email', None) - self.key_file_path = kwargs.get('key_file_path', None) - self.trusted_cert_path = kwargs.get('trusted_cert_path', None) - self.use_system_trust_store = kwargs.get('use_system_trust_store', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - self.type = 'GoogleAdWords' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_ad_words_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_ad_words_linked_service_py3.py deleted file mode 100644 index ff5ffe6df381..000000000000 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_ad_words_linked_service_py3.py +++ /dev/null @@ -1,119 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for -# license information. -# -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is -# regenerated. -# -------------------------------------------------------------------------- - -from .linked_service_py3 import LinkedService - - -class GoogleAdWordsLinkedService(LinkedService): - """Google Awords service linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] - :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.mgmt.datafactory.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.mgmt.datafactory.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the - Dataset. - :type annotations: list[object] - :param type: Required. Constant filled by server. - :type type: str - :param client_customer_id: Required. The Client customer ID of the AdWords - account that you want to fetch report data for. - :type client_customer_id: object - :param developer_token: Required. The developer token associated with the - manager account that you use to grant access to the AdWords API. - :type developer_token: ~azure.mgmt.datafactory.models.SecretBase - :param authentication_type: Required. The OAuth 2.0 authentication - mechanism used for authentication. ServiceAuthentication can only be used - on self-hosted IR. Possible values include: 'ServiceAuthentication', - 'UserAuthentication' - :type authentication_type: str or - ~azure.mgmt.datafactory.models.GoogleAdWordsAuthenticationType - :param refresh_token: The refresh token obtained from Google for - authorizing access to AdWords for UserAuthentication. - :type refresh_token: ~azure.mgmt.datafactory.models.SecretBase - :param client_id: The client id of the google application used to acquire - the refresh token. - :type client_id: ~azure.mgmt.datafactory.models.SecretBase - :param client_secret: The client secret of the google application used to - acquire the refresh token. - :type client_secret: ~azure.mgmt.datafactory.models.SecretBase - :param email: The service account email ID that is used for - ServiceAuthentication and can only be used on self-hosted IR. - :type email: object - :param key_file_path: The full path to the .p12 key file that is used to - authenticate the service account email address and can only be used on - self-hosted IR. - :type key_file_path: object - :param trusted_cert_path: The full path of the .pem file containing - trusted CA certificates for verifying the server when connecting over SSL. - This property can only be set when using SSL on self-hosted IR. The - default value is the cacerts.pem file installed with the IR. - :type trusted_cert_path: object - :param use_system_trust_store: Specifies whether to use a CA certificate - from the system trust store or from a specified PEM file. The default - value is false. - :type use_system_trust_store: object - :param encrypted_credential: The encrypted credential used for - authentication. Credentials are encrypted using the integration runtime - credential manager. Type: string (or Expression with resultType string). - :type encrypted_credential: object - """ - - _validation = { - 'type': {'required': True}, - 'client_customer_id': {'required': True}, - 'developer_token': {'required': True}, - 'authentication_type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'type': {'key': 'type', 'type': 'str'}, - 'client_customer_id': {'key': 'typeProperties.clientCustomerID', 'type': 'object'}, - 'developer_token': {'key': 'typeProperties.developerToken', 'type': 'SecretBase'}, - 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, - 'refresh_token': {'key': 'typeProperties.refreshToken', 'type': 'SecretBase'}, - 'client_id': {'key': 'typeProperties.clientId', 'type': 'SecretBase'}, - 'client_secret': {'key': 'typeProperties.clientSecret', 'type': 'SecretBase'}, - 'email': {'key': 'typeProperties.email', 'type': 'object'}, - 'key_file_path': {'key': 'typeProperties.keyFilePath', 'type': 'object'}, - 'trusted_cert_path': {'key': 'typeProperties.trustedCertPath', 'type': 'object'}, - 'use_system_trust_store': {'key': 'typeProperties.useSystemTrustStore', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__(self, *, client_customer_id, developer_token, authentication_type, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, refresh_token=None, client_id=None, client_secret=None, email=None, key_file_path=None, trusted_cert_path=None, use_system_trust_store=None, encrypted_credential=None, **kwargs) -> None: - super(GoogleAdWordsLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.client_customer_id = client_customer_id - self.developer_token = developer_token - self.authentication_type = authentication_type - self.refresh_token = refresh_token - self.client_id = client_id - self.client_secret = client_secret - self.email = email - self.key_file_path = key_file_path - self.trusted_cert_path = trusted_cert_path - self.use_system_trust_store = use_system_trust_store - self.encrypted_credential = encrypted_credential - self.type = 'GoogleAdWords' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_ad_words_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_ad_words_object_dataset.py deleted file mode 100644 index 0daede4e239a..000000000000 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_ad_words_object_dataset.py +++ /dev/null @@ -1,62 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for -# license information. -# -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is -# regenerated. -# -------------------------------------------------------------------------- - -from .dataset import Dataset - - -class GoogleAdWordsObjectDataset(Dataset): - """Google AdWords service dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: - array (or Expression with resultType array), itemType: DatasetDataElement. - :type structure: object - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.mgmt.datafactory.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.mgmt.datafactory.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the - Dataset. - :type annotations: list[object] - :param folder: The folder that this Dataset is in. If not specified, - Dataset will appear at the root level. - :type folder: ~azure.mgmt.datafactory.models.DatasetFolder - :param type: Required. Constant filled by server. - :type type: str - """ - - _validation = { - 'linked_service_name': {'required': True}, - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'type': {'key': 'type', 'type': 'str'}, - } - - def __init__(self, **kwargs): - super(GoogleAdWordsObjectDataset, self).__init__(**kwargs) - self.type = 'GoogleAdWordsObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_ad_words_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_ad_words_object_dataset_py3.py deleted file mode 100644 index 7c3fb4379041..000000000000 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_ad_words_object_dataset_py3.py +++ /dev/null @@ -1,62 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for -# license information. -# -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is -# regenerated. -# -------------------------------------------------------------------------- - -from .dataset_py3 import Dataset - - -class GoogleAdWordsObjectDataset(Dataset): - """Google AdWords service dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: - array (or Expression with resultType array), itemType: DatasetDataElement. - :type structure: object - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.mgmt.datafactory.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.mgmt.datafactory.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the - Dataset. - :type annotations: list[object] - :param folder: The folder that this Dataset is in. If not specified, - Dataset will appear at the root level. - :type folder: ~azure.mgmt.datafactory.models.DatasetFolder - :param type: Required. Constant filled by server. - :type type: str - """ - - _validation = { - 'linked_service_name': {'required': True}, - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'type': {'key': 'type', 'type': 'str'}, - } - - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: - super(GoogleAdWordsObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'GoogleAdWordsObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_ad_words_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_ad_words_source.py deleted file mode 100644 index ccd21de4a364..000000000000 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_ad_words_source.py +++ /dev/null @@ -1,52 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for -# license information. -# -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is -# regenerated. -# -------------------------------------------------------------------------- - -from .copy_source import CopySource - - -class GoogleAdWordsSource(CopySource): - """A copy activity Google AdWords service source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] - :param source_retry_count: Source retry count. Type: integer (or - Expression with resultType integer). - :type source_retry_count: object - :param source_retry_wait: Source retry wait. Type: string (or Expression - with resultType string), pattern: - ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: object - :param type: Required. Constant filled by server. - :type type: str - :param query: A query to retrieve data from source. Type: string (or - Expression with resultType string). - :type query: object - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'type': {'key': 'type', 'type': 'str'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__(self, **kwargs): - super(GoogleAdWordsSource, self).__init__(**kwargs) - self.query = kwargs.get('query', None) - self.type = 'GoogleAdWordsSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_ad_words_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_ad_words_source_py3.py deleted file mode 100644 index 6e421dddc45f..000000000000 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_ad_words_source_py3.py +++ /dev/null @@ -1,52 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for -# license information. -# -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is -# regenerated. -# -------------------------------------------------------------------------- - -from .copy_source_py3 import CopySource - - -class GoogleAdWordsSource(CopySource): - """A copy activity Google AdWords service source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] - :param source_retry_count: Source retry count. Type: integer (or - Expression with resultType integer). - :type source_retry_count: object - :param source_retry_wait: Source retry wait. Type: string (or Expression - with resultType string), pattern: - ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: object - :param type: Required. Constant filled by server. - :type type: str - :param query: A query to retrieve data from source. Type: string (or - Expression with resultType string). - :type query: object - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'type': {'key': 'type', 'type': 'str'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__(self, *, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None, **kwargs) -> None: - super(GoogleAdWordsSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, **kwargs) - self.query = query - self.type = 'GoogleAdWordsSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service.py index c74f1320eb95..3dec3c0dcc24 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service.py @@ -28,20 +28,19 @@ class LinkedService(Model): MariaDBLinkedService, MagentoLinkedService, JiraLinkedService, ImpalaLinkedService, HubspotLinkedService, HiveLinkedService, HBaseLinkedService, GreenplumLinkedService, GoogleBigQueryLinkedService, - GoogleAdWordsLinkedService, EloquaLinkedService, DrillLinkedService, - CouchbaseLinkedService, ConcurLinkedService, AzurePostgreSqlLinkedService, - AmazonMWSLinkedService, SapHanaLinkedService, SapBWLinkedService, - SftpServerLinkedService, FtpServerLinkedService, RestServiceLinkedService, - HttpLinkedService, AzureSearchLinkedService, CustomDataSourceLinkedService, - AmazonRedshiftLinkedService, AmazonS3LinkedService, SapEccLinkedService, + EloquaLinkedService, DrillLinkedService, CouchbaseLinkedService, + ConcurLinkedService, AzurePostgreSqlLinkedService, AmazonMWSLinkedService, + SapHanaLinkedService, SapBWLinkedService, SftpServerLinkedService, + FtpServerLinkedService, HttpLinkedService, AzureSearchLinkedService, + CustomDataSourceLinkedService, AmazonRedshiftLinkedService, + AmazonS3LinkedService, SapEccLinkedService, SapCloudForCustomerLinkedService, SalesforceLinkedService, AzureDataLakeStoreLinkedService, MongoDbLinkedService, CassandraLinkedService, WebLinkedService, ODataLinkedService, HdfsLinkedService, OdbcLinkedService, AzureMLLinkedService, TeradataLinkedService, Db2LinkedService, SybaseLinkedService, PostgreSqlLinkedService, MySqlLinkedService, AzureMySqlLinkedService, - OracleServiceCloudLinkedService, OracleLinkedService, - FileServerLinkedService, HDInsightLinkedService, DynamicsAXLinkedService, + OracleLinkedService, FileServerLinkedService, HDInsightLinkedService, DynamicsLinkedService, CosmosDbLinkedService, AzureKeyVaultLinkedService, AzureBatchLinkedService, AzureSqlDatabaseLinkedService, SqlServerLinkedService, AzureSqlDWLinkedService, @@ -82,7 +81,7 @@ class LinkedService(Model): } _subtype_map = { - 'type': {'Responsys': 'ResponsysLinkedService', 'AzureDatabricks': 'AzureDatabricksLinkedService', 'AzureDataLakeAnalytics': 'AzureDataLakeAnalyticsLinkedService', 'HDInsightOnDemand': 'HDInsightOnDemandLinkedService', 'SalesforceMarketingCloud': 'SalesforceMarketingCloudLinkedService', 'Netezza': 'NetezzaLinkedService', 'Vertica': 'VerticaLinkedService', 'Zoho': 'ZohoLinkedService', 'Xero': 'XeroLinkedService', 'Square': 'SquareLinkedService', 'Spark': 'SparkLinkedService', 'Shopify': 'ShopifyLinkedService', 'ServiceNow': 'ServiceNowLinkedService', 'QuickBooks': 'QuickBooksLinkedService', 'Presto': 'PrestoLinkedService', 'Phoenix': 'PhoenixLinkedService', 'Paypal': 'PaypalLinkedService', 'Marketo': 'MarketoLinkedService', 'MariaDB': 'MariaDBLinkedService', 'Magento': 'MagentoLinkedService', 'Jira': 'JiraLinkedService', 'Impala': 'ImpalaLinkedService', 'Hubspot': 'HubspotLinkedService', 'Hive': 'HiveLinkedService', 'HBase': 'HBaseLinkedService', 'Greenplum': 'GreenplumLinkedService', 'GoogleBigQuery': 'GoogleBigQueryLinkedService', 'GoogleAdWords': 'GoogleAdWordsLinkedService', 'Eloqua': 'EloquaLinkedService', 'Drill': 'DrillLinkedService', 'Couchbase': 'CouchbaseLinkedService', 'Concur': 'ConcurLinkedService', 'AzurePostgreSql': 'AzurePostgreSqlLinkedService', 'AmazonMWS': 'AmazonMWSLinkedService', 'SapHana': 'SapHanaLinkedService', 'SapBW': 'SapBWLinkedService', 'Sftp': 'SftpServerLinkedService', 'FtpServer': 'FtpServerLinkedService', 'RestService': 'RestServiceLinkedService', 'HttpServer': 'HttpLinkedService', 'AzureSearch': 'AzureSearchLinkedService', 'CustomDataSource': 'CustomDataSourceLinkedService', 'AmazonRedshift': 'AmazonRedshiftLinkedService', 'AmazonS3': 'AmazonS3LinkedService', 'SapEcc': 'SapEccLinkedService', 'SapCloudForCustomer': 'SapCloudForCustomerLinkedService', 'Salesforce': 'SalesforceLinkedService', 'AzureDataLakeStore': 'AzureDataLakeStoreLinkedService', 'MongoDb': 'MongoDbLinkedService', 'Cassandra': 'CassandraLinkedService', 'Web': 'WebLinkedService', 'OData': 'ODataLinkedService', 'Hdfs': 'HdfsLinkedService', 'Odbc': 'OdbcLinkedService', 'AzureML': 'AzureMLLinkedService', 'Teradata': 'TeradataLinkedService', 'Db2': 'Db2LinkedService', 'Sybase': 'SybaseLinkedService', 'PostgreSql': 'PostgreSqlLinkedService', 'MySql': 'MySqlLinkedService', 'AzureMySql': 'AzureMySqlLinkedService', 'OracleServiceCloud': 'OracleServiceCloudLinkedService', 'Oracle': 'OracleLinkedService', 'FileServer': 'FileServerLinkedService', 'HDInsight': 'HDInsightLinkedService', 'DynamicsAX': 'DynamicsAXLinkedService', 'Dynamics': 'DynamicsLinkedService', 'CosmosDb': 'CosmosDbLinkedService', 'AzureKeyVault': 'AzureKeyVaultLinkedService', 'AzureBatch': 'AzureBatchLinkedService', 'AzureSqlDatabase': 'AzureSqlDatabaseLinkedService', 'SqlServer': 'SqlServerLinkedService', 'AzureSqlDW': 'AzureSqlDWLinkedService', 'AzureTableStorage': 'AzureTableStorageLinkedService', 'AzureBlobStorage': 'AzureBlobStorageLinkedService', 'AzureStorage': 'AzureStorageLinkedService'} + 'type': {'Responsys': 'ResponsysLinkedService', 'AzureDatabricks': 'AzureDatabricksLinkedService', 'AzureDataLakeAnalytics': 'AzureDataLakeAnalyticsLinkedService', 'HDInsightOnDemand': 'HDInsightOnDemandLinkedService', 'SalesforceMarketingCloud': 'SalesforceMarketingCloudLinkedService', 'Netezza': 'NetezzaLinkedService', 'Vertica': 'VerticaLinkedService', 'Zoho': 'ZohoLinkedService', 'Xero': 'XeroLinkedService', 'Square': 'SquareLinkedService', 'Spark': 'SparkLinkedService', 'Shopify': 'ShopifyLinkedService', 'ServiceNow': 'ServiceNowLinkedService', 'QuickBooks': 'QuickBooksLinkedService', 'Presto': 'PrestoLinkedService', 'Phoenix': 'PhoenixLinkedService', 'Paypal': 'PaypalLinkedService', 'Marketo': 'MarketoLinkedService', 'MariaDB': 'MariaDBLinkedService', 'Magento': 'MagentoLinkedService', 'Jira': 'JiraLinkedService', 'Impala': 'ImpalaLinkedService', 'Hubspot': 'HubspotLinkedService', 'Hive': 'HiveLinkedService', 'HBase': 'HBaseLinkedService', 'Greenplum': 'GreenplumLinkedService', 'GoogleBigQuery': 'GoogleBigQueryLinkedService', 'Eloqua': 'EloquaLinkedService', 'Drill': 'DrillLinkedService', 'Couchbase': 'CouchbaseLinkedService', 'Concur': 'ConcurLinkedService', 'AzurePostgreSql': 'AzurePostgreSqlLinkedService', 'AmazonMWS': 'AmazonMWSLinkedService', 'SapHana': 'SapHanaLinkedService', 'SapBW': 'SapBWLinkedService', 'Sftp': 'SftpServerLinkedService', 'FtpServer': 'FtpServerLinkedService', 'HttpServer': 'HttpLinkedService', 'AzureSearch': 'AzureSearchLinkedService', 'CustomDataSource': 'CustomDataSourceLinkedService', 'AmazonRedshift': 'AmazonRedshiftLinkedService', 'AmazonS3': 'AmazonS3LinkedService', 'SapEcc': 'SapEccLinkedService', 'SapCloudForCustomer': 'SapCloudForCustomerLinkedService', 'Salesforce': 'SalesforceLinkedService', 'AzureDataLakeStore': 'AzureDataLakeStoreLinkedService', 'MongoDb': 'MongoDbLinkedService', 'Cassandra': 'CassandraLinkedService', 'Web': 'WebLinkedService', 'OData': 'ODataLinkedService', 'Hdfs': 'HdfsLinkedService', 'Odbc': 'OdbcLinkedService', 'AzureML': 'AzureMLLinkedService', 'Teradata': 'TeradataLinkedService', 'Db2': 'Db2LinkedService', 'Sybase': 'SybaseLinkedService', 'PostgreSql': 'PostgreSqlLinkedService', 'MySql': 'MySqlLinkedService', 'AzureMySql': 'AzureMySqlLinkedService', 'Oracle': 'OracleLinkedService', 'FileServer': 'FileServerLinkedService', 'HDInsight': 'HDInsightLinkedService', 'Dynamics': 'DynamicsLinkedService', 'CosmosDb': 'CosmosDbLinkedService', 'AzureKeyVault': 'AzureKeyVaultLinkedService', 'AzureBatch': 'AzureBatchLinkedService', 'AzureSqlDatabase': 'AzureSqlDatabaseLinkedService', 'SqlServer': 'SqlServerLinkedService', 'AzureSqlDW': 'AzureSqlDWLinkedService', 'AzureTableStorage': 'AzureTableStorageLinkedService', 'AzureBlobStorage': 'AzureBlobStorageLinkedService', 'AzureStorage': 'AzureStorageLinkedService'} } def __init__(self, **kwargs): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service_py3.py index ba49c525e830..885af520bcda 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service_py3.py @@ -28,20 +28,19 @@ class LinkedService(Model): MariaDBLinkedService, MagentoLinkedService, JiraLinkedService, ImpalaLinkedService, HubspotLinkedService, HiveLinkedService, HBaseLinkedService, GreenplumLinkedService, GoogleBigQueryLinkedService, - GoogleAdWordsLinkedService, EloquaLinkedService, DrillLinkedService, - CouchbaseLinkedService, ConcurLinkedService, AzurePostgreSqlLinkedService, - AmazonMWSLinkedService, SapHanaLinkedService, SapBWLinkedService, - SftpServerLinkedService, FtpServerLinkedService, RestServiceLinkedService, - HttpLinkedService, AzureSearchLinkedService, CustomDataSourceLinkedService, - AmazonRedshiftLinkedService, AmazonS3LinkedService, SapEccLinkedService, + EloquaLinkedService, DrillLinkedService, CouchbaseLinkedService, + ConcurLinkedService, AzurePostgreSqlLinkedService, AmazonMWSLinkedService, + SapHanaLinkedService, SapBWLinkedService, SftpServerLinkedService, + FtpServerLinkedService, HttpLinkedService, AzureSearchLinkedService, + CustomDataSourceLinkedService, AmazonRedshiftLinkedService, + AmazonS3LinkedService, SapEccLinkedService, SapCloudForCustomerLinkedService, SalesforceLinkedService, AzureDataLakeStoreLinkedService, MongoDbLinkedService, CassandraLinkedService, WebLinkedService, ODataLinkedService, HdfsLinkedService, OdbcLinkedService, AzureMLLinkedService, TeradataLinkedService, Db2LinkedService, SybaseLinkedService, PostgreSqlLinkedService, MySqlLinkedService, AzureMySqlLinkedService, - OracleServiceCloudLinkedService, OracleLinkedService, - FileServerLinkedService, HDInsightLinkedService, DynamicsAXLinkedService, + OracleLinkedService, FileServerLinkedService, HDInsightLinkedService, DynamicsLinkedService, CosmosDbLinkedService, AzureKeyVaultLinkedService, AzureBatchLinkedService, AzureSqlDatabaseLinkedService, SqlServerLinkedService, AzureSqlDWLinkedService, @@ -82,7 +81,7 @@ class LinkedService(Model): } _subtype_map = { - 'type': {'Responsys': 'ResponsysLinkedService', 'AzureDatabricks': 'AzureDatabricksLinkedService', 'AzureDataLakeAnalytics': 'AzureDataLakeAnalyticsLinkedService', 'HDInsightOnDemand': 'HDInsightOnDemandLinkedService', 'SalesforceMarketingCloud': 'SalesforceMarketingCloudLinkedService', 'Netezza': 'NetezzaLinkedService', 'Vertica': 'VerticaLinkedService', 'Zoho': 'ZohoLinkedService', 'Xero': 'XeroLinkedService', 'Square': 'SquareLinkedService', 'Spark': 'SparkLinkedService', 'Shopify': 'ShopifyLinkedService', 'ServiceNow': 'ServiceNowLinkedService', 'QuickBooks': 'QuickBooksLinkedService', 'Presto': 'PrestoLinkedService', 'Phoenix': 'PhoenixLinkedService', 'Paypal': 'PaypalLinkedService', 'Marketo': 'MarketoLinkedService', 'MariaDB': 'MariaDBLinkedService', 'Magento': 'MagentoLinkedService', 'Jira': 'JiraLinkedService', 'Impala': 'ImpalaLinkedService', 'Hubspot': 'HubspotLinkedService', 'Hive': 'HiveLinkedService', 'HBase': 'HBaseLinkedService', 'Greenplum': 'GreenplumLinkedService', 'GoogleBigQuery': 'GoogleBigQueryLinkedService', 'GoogleAdWords': 'GoogleAdWordsLinkedService', 'Eloqua': 'EloquaLinkedService', 'Drill': 'DrillLinkedService', 'Couchbase': 'CouchbaseLinkedService', 'Concur': 'ConcurLinkedService', 'AzurePostgreSql': 'AzurePostgreSqlLinkedService', 'AmazonMWS': 'AmazonMWSLinkedService', 'SapHana': 'SapHanaLinkedService', 'SapBW': 'SapBWLinkedService', 'Sftp': 'SftpServerLinkedService', 'FtpServer': 'FtpServerLinkedService', 'RestService': 'RestServiceLinkedService', 'HttpServer': 'HttpLinkedService', 'AzureSearch': 'AzureSearchLinkedService', 'CustomDataSource': 'CustomDataSourceLinkedService', 'AmazonRedshift': 'AmazonRedshiftLinkedService', 'AmazonS3': 'AmazonS3LinkedService', 'SapEcc': 'SapEccLinkedService', 'SapCloudForCustomer': 'SapCloudForCustomerLinkedService', 'Salesforce': 'SalesforceLinkedService', 'AzureDataLakeStore': 'AzureDataLakeStoreLinkedService', 'MongoDb': 'MongoDbLinkedService', 'Cassandra': 'CassandraLinkedService', 'Web': 'WebLinkedService', 'OData': 'ODataLinkedService', 'Hdfs': 'HdfsLinkedService', 'Odbc': 'OdbcLinkedService', 'AzureML': 'AzureMLLinkedService', 'Teradata': 'TeradataLinkedService', 'Db2': 'Db2LinkedService', 'Sybase': 'SybaseLinkedService', 'PostgreSql': 'PostgreSqlLinkedService', 'MySql': 'MySqlLinkedService', 'AzureMySql': 'AzureMySqlLinkedService', 'OracleServiceCloud': 'OracleServiceCloudLinkedService', 'Oracle': 'OracleLinkedService', 'FileServer': 'FileServerLinkedService', 'HDInsight': 'HDInsightLinkedService', 'DynamicsAX': 'DynamicsAXLinkedService', 'Dynamics': 'DynamicsLinkedService', 'CosmosDb': 'CosmosDbLinkedService', 'AzureKeyVault': 'AzureKeyVaultLinkedService', 'AzureBatch': 'AzureBatchLinkedService', 'AzureSqlDatabase': 'AzureSqlDatabaseLinkedService', 'SqlServer': 'SqlServerLinkedService', 'AzureSqlDW': 'AzureSqlDWLinkedService', 'AzureTableStorage': 'AzureTableStorageLinkedService', 'AzureBlobStorage': 'AzureBlobStorageLinkedService', 'AzureStorage': 'AzureStorageLinkedService'} + 'type': {'Responsys': 'ResponsysLinkedService', 'AzureDatabricks': 'AzureDatabricksLinkedService', 'AzureDataLakeAnalytics': 'AzureDataLakeAnalyticsLinkedService', 'HDInsightOnDemand': 'HDInsightOnDemandLinkedService', 'SalesforceMarketingCloud': 'SalesforceMarketingCloudLinkedService', 'Netezza': 'NetezzaLinkedService', 'Vertica': 'VerticaLinkedService', 'Zoho': 'ZohoLinkedService', 'Xero': 'XeroLinkedService', 'Square': 'SquareLinkedService', 'Spark': 'SparkLinkedService', 'Shopify': 'ShopifyLinkedService', 'ServiceNow': 'ServiceNowLinkedService', 'QuickBooks': 'QuickBooksLinkedService', 'Presto': 'PrestoLinkedService', 'Phoenix': 'PhoenixLinkedService', 'Paypal': 'PaypalLinkedService', 'Marketo': 'MarketoLinkedService', 'MariaDB': 'MariaDBLinkedService', 'Magento': 'MagentoLinkedService', 'Jira': 'JiraLinkedService', 'Impala': 'ImpalaLinkedService', 'Hubspot': 'HubspotLinkedService', 'Hive': 'HiveLinkedService', 'HBase': 'HBaseLinkedService', 'Greenplum': 'GreenplumLinkedService', 'GoogleBigQuery': 'GoogleBigQueryLinkedService', 'Eloqua': 'EloquaLinkedService', 'Drill': 'DrillLinkedService', 'Couchbase': 'CouchbaseLinkedService', 'Concur': 'ConcurLinkedService', 'AzurePostgreSql': 'AzurePostgreSqlLinkedService', 'AmazonMWS': 'AmazonMWSLinkedService', 'SapHana': 'SapHanaLinkedService', 'SapBW': 'SapBWLinkedService', 'Sftp': 'SftpServerLinkedService', 'FtpServer': 'FtpServerLinkedService', 'HttpServer': 'HttpLinkedService', 'AzureSearch': 'AzureSearchLinkedService', 'CustomDataSource': 'CustomDataSourceLinkedService', 'AmazonRedshift': 'AmazonRedshiftLinkedService', 'AmazonS3': 'AmazonS3LinkedService', 'SapEcc': 'SapEccLinkedService', 'SapCloudForCustomer': 'SapCloudForCustomerLinkedService', 'Salesforce': 'SalesforceLinkedService', 'AzureDataLakeStore': 'AzureDataLakeStoreLinkedService', 'MongoDb': 'MongoDbLinkedService', 'Cassandra': 'CassandraLinkedService', 'Web': 'WebLinkedService', 'OData': 'ODataLinkedService', 'Hdfs': 'HdfsLinkedService', 'Odbc': 'OdbcLinkedService', 'AzureML': 'AzureMLLinkedService', 'Teradata': 'TeradataLinkedService', 'Db2': 'Db2LinkedService', 'Sybase': 'SybaseLinkedService', 'PostgreSql': 'PostgreSqlLinkedService', 'MySql': 'MySqlLinkedService', 'AzureMySql': 'AzureMySqlLinkedService', 'Oracle': 'OracleLinkedService', 'FileServer': 'FileServerLinkedService', 'HDInsight': 'HDInsightLinkedService', 'Dynamics': 'DynamicsLinkedService', 'CosmosDb': 'CosmosDbLinkedService', 'AzureKeyVault': 'AzureKeyVaultLinkedService', 'AzureBatch': 'AzureBatchLinkedService', 'AzureSqlDatabase': 'AzureSqlDatabaseLinkedService', 'SqlServer': 'SqlServerLinkedService', 'AzureSqlDW': 'AzureSqlDWLinkedService', 'AzureTableStorage': 'AzureTableStorageLinkedService', 'AzureBlobStorage': 'AzureBlobStorageLinkedService', 'AzureStorage': 'AzureStorageLinkedService'} } def __init__(self, *, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, **kwargs) -> None: diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odata_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odata_linked_service.py index 2c2fbab04c62..9a7edca9ddb1 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odata_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odata_linked_service.py @@ -36,9 +36,8 @@ class ODataLinkedService(LinkedService): :param url: Required. The URL of the OData service endpoint. Type: string (or Expression with resultType string). :type url: object - :param authentication_type: Required. Type of authentication used to - connect to the OData service. Possible values include: 'Anonymous', - 'Basic', 'Windows', 'AadServicePrincipal', 'ManagedServiceIdentity' + :param authentication_type: Type of authentication used to connect to the + OData service. Possible values include: 'Basic', 'Anonymous' :type authentication_type: str or ~azure.mgmt.datafactory.models.ODataAuthenticationType :param user_name: User name of the OData service. Type: string (or @@ -46,38 +45,6 @@ class ODataLinkedService(LinkedService): :type user_name: object :param password: Password of the OData service. :type password: ~azure.mgmt.datafactory.models.SecretBase - :param tenant: Specify the tenant information (domain name or tenant ID) - under which your application resides. Type: string (or Expression with - resultType string). - :type tenant: object - :param service_principal_id: Specify the application id of your - application registered in Azure Active Directory. Type: string (or - Expression with resultType string). - :type service_principal_id: object - :param aad_resource_id: Specify the resource you are requesting - authorization to use Directory. Type: string (or Expression with - resultType string). - :type aad_resource_id: object - :param aad_service_principal_credential_type: Specify the credential type - (key or cert) is used for service principal. Possible values include: - 'ServicePrincipalKey', 'ServicePrincipalCert' - :type aad_service_principal_credential_type: str or - ~azure.mgmt.datafactory.models.ODataAadServicePrincipalCredentialType - :param service_principal_key: Specify the secret of your application - registered in Azure Active Directory. Type: string (or Expression with - resultType string). - :type service_principal_key: ~azure.mgmt.datafactory.models.SecretBase - :param service_principal_embedded_cert: Specify the base64 encoded - certificate of your application registered in Azure Active Directory. - Type: string (or Expression with resultType string). - :type service_principal_embedded_cert: - ~azure.mgmt.datafactory.models.SecretBase - :param service_principal_embedded_cert_password: Specify the password of - your certificate if your certificate has a password and you are using - AadServicePrincipal authentication. Type: string (or Expression with - resultType string). - :type service_principal_embedded_cert_password: - ~azure.mgmt.datafactory.models.SecretBase :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -87,7 +54,6 @@ class ODataLinkedService(LinkedService): _validation = { 'type': {'required': True}, 'url': {'required': True}, - 'authentication_type': {'required': True}, } _attribute_map = { @@ -101,13 +67,6 @@ class ODataLinkedService(LinkedService): 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, 'user_name': {'key': 'typeProperties.userName', 'type': 'object'}, 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'tenant': {'key': 'typeProperties.tenant', 'type': 'object'}, - 'service_principal_id': {'key': 'typeProperties.servicePrincipalId', 'type': 'object'}, - 'aad_resource_id': {'key': 'typeProperties.aadResourceId', 'type': 'object'}, - 'aad_service_principal_credential_type': {'key': 'typeProperties.aadServicePrincipalCredentialType', 'type': 'str'}, - 'service_principal_key': {'key': 'typeProperties.servicePrincipalKey', 'type': 'SecretBase'}, - 'service_principal_embedded_cert': {'key': 'typeProperties.servicePrincipalEmbeddedCert', 'type': 'SecretBase'}, - 'service_principal_embedded_cert_password': {'key': 'typeProperties.servicePrincipalEmbeddedCertPassword', 'type': 'SecretBase'}, 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } @@ -117,12 +76,5 @@ def __init__(self, **kwargs): self.authentication_type = kwargs.get('authentication_type', None) self.user_name = kwargs.get('user_name', None) self.password = kwargs.get('password', None) - self.tenant = kwargs.get('tenant', None) - self.service_principal_id = kwargs.get('service_principal_id', None) - self.aad_resource_id = kwargs.get('aad_resource_id', None) - self.aad_service_principal_credential_type = kwargs.get('aad_service_principal_credential_type', None) - self.service_principal_key = kwargs.get('service_principal_key', None) - self.service_principal_embedded_cert = kwargs.get('service_principal_embedded_cert', None) - self.service_principal_embedded_cert_password = kwargs.get('service_principal_embedded_cert_password', None) self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'OData' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odata_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odata_linked_service_py3.py index 80088701b50b..688bb4e4ffda 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odata_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odata_linked_service_py3.py @@ -36,9 +36,8 @@ class ODataLinkedService(LinkedService): :param url: Required. The URL of the OData service endpoint. Type: string (or Expression with resultType string). :type url: object - :param authentication_type: Required. Type of authentication used to - connect to the OData service. Possible values include: 'Anonymous', - 'Basic', 'Windows', 'AadServicePrincipal', 'ManagedServiceIdentity' + :param authentication_type: Type of authentication used to connect to the + OData service. Possible values include: 'Basic', 'Anonymous' :type authentication_type: str or ~azure.mgmt.datafactory.models.ODataAuthenticationType :param user_name: User name of the OData service. Type: string (or @@ -46,38 +45,6 @@ class ODataLinkedService(LinkedService): :type user_name: object :param password: Password of the OData service. :type password: ~azure.mgmt.datafactory.models.SecretBase - :param tenant: Specify the tenant information (domain name or tenant ID) - under which your application resides. Type: string (or Expression with - resultType string). - :type tenant: object - :param service_principal_id: Specify the application id of your - application registered in Azure Active Directory. Type: string (or - Expression with resultType string). - :type service_principal_id: object - :param aad_resource_id: Specify the resource you are requesting - authorization to use Directory. Type: string (or Expression with - resultType string). - :type aad_resource_id: object - :param aad_service_principal_credential_type: Specify the credential type - (key or cert) is used for service principal. Possible values include: - 'ServicePrincipalKey', 'ServicePrincipalCert' - :type aad_service_principal_credential_type: str or - ~azure.mgmt.datafactory.models.ODataAadServicePrincipalCredentialType - :param service_principal_key: Specify the secret of your application - registered in Azure Active Directory. Type: string (or Expression with - resultType string). - :type service_principal_key: ~azure.mgmt.datafactory.models.SecretBase - :param service_principal_embedded_cert: Specify the base64 encoded - certificate of your application registered in Azure Active Directory. - Type: string (or Expression with resultType string). - :type service_principal_embedded_cert: - ~azure.mgmt.datafactory.models.SecretBase - :param service_principal_embedded_cert_password: Specify the password of - your certificate if your certificate has a password and you are using - AadServicePrincipal authentication. Type: string (or Expression with - resultType string). - :type service_principal_embedded_cert_password: - ~azure.mgmt.datafactory.models.SecretBase :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -87,7 +54,6 @@ class ODataLinkedService(LinkedService): _validation = { 'type': {'required': True}, 'url': {'required': True}, - 'authentication_type': {'required': True}, } _attribute_map = { @@ -101,28 +67,14 @@ class ODataLinkedService(LinkedService): 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, 'user_name': {'key': 'typeProperties.userName', 'type': 'object'}, 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'tenant': {'key': 'typeProperties.tenant', 'type': 'object'}, - 'service_principal_id': {'key': 'typeProperties.servicePrincipalId', 'type': 'object'}, - 'aad_resource_id': {'key': 'typeProperties.aadResourceId', 'type': 'object'}, - 'aad_service_principal_credential_type': {'key': 'typeProperties.aadServicePrincipalCredentialType', 'type': 'str'}, - 'service_principal_key': {'key': 'typeProperties.servicePrincipalKey', 'type': 'SecretBase'}, - 'service_principal_embedded_cert': {'key': 'typeProperties.servicePrincipalEmbeddedCert', 'type': 'SecretBase'}, - 'service_principal_embedded_cert_password': {'key': 'typeProperties.servicePrincipalEmbeddedCertPassword', 'type': 'SecretBase'}, 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, *, url, authentication_type, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, user_name=None, password=None, tenant=None, service_principal_id=None, aad_resource_id=None, aad_service_principal_credential_type=None, service_principal_key=None, service_principal_embedded_cert=None, service_principal_embedded_cert_password=None, encrypted_credential=None, **kwargs) -> None: + def __init__(self, *, url, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, authentication_type=None, user_name=None, password=None, encrypted_credential=None, **kwargs) -> None: super(ODataLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) self.url = url self.authentication_type = authentication_type self.user_name = user_name self.password = password - self.tenant = tenant - self.service_principal_id = service_principal_id - self.aad_resource_id = aad_resource_id - self.aad_service_principal_credential_type = aad_service_principal_credential_type - self.service_principal_key = service_principal_key - self.service_principal_embedded_cert = service_principal_embedded_cert - self.service_principal_embedded_cert_password = service_principal_embedded_cert_password self.encrypted_credential = encrypted_credential self.type = 'OData' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_service_cloud_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_service_cloud_linked_service.py deleted file mode 100644 index fae3d3a7a8d4..000000000000 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_service_cloud_linked_service.py +++ /dev/null @@ -1,95 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for -# license information. -# -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is -# regenerated. -# -------------------------------------------------------------------------- - -from .linked_service import LinkedService - - -class OracleServiceCloudLinkedService(LinkedService): - """Oracle Service Cloud linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] - :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.mgmt.datafactory.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.mgmt.datafactory.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the - Dataset. - :type annotations: list[object] - :param type: Required. Constant filled by server. - :type type: str - :param host: Required. The URL of the Oracle Service Cloud instance. - :type host: object - :param username: Required. The user name that you use to access Oracle - Service Cloud server. - :type username: object - :param password: Required. The password corresponding to the user name - that you provided in the username key. - :type password: ~azure.mgmt.datafactory.models.SecretBase - :param use_encrypted_endpoints: Specifies whether the data source - endpoints are encrypted using HTTPS. The default value is true. Type: - boolean (or Expression with resultType boolean). - :type use_encrypted_endpoints: object - :param use_host_verification: Specifies whether to require the host name - in the server's certificate to match the host name of the server when - connecting over SSL. The default value is true. Type: boolean (or - Expression with resultType boolean). - :type use_host_verification: object - :param use_peer_verification: Specifies whether to verify the identity of - the server when connecting over SSL. The default value is true. Type: - boolean (or Expression with resultType boolean). - :type use_peer_verification: object - :param encrypted_credential: The encrypted credential used for - authentication. Credentials are encrypted using the integration runtime - credential manager. Type: string (or Expression with resultType string). - :type encrypted_credential: object - """ - - _validation = { - 'type': {'required': True}, - 'host': {'required': True}, - 'username': {'required': True}, - 'password': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'type': {'key': 'type', 'type': 'str'}, - 'host': {'key': 'typeProperties.host', 'type': 'object'}, - 'username': {'key': 'typeProperties.username', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'use_encrypted_endpoints': {'key': 'typeProperties.useEncryptedEndpoints', 'type': 'object'}, - 'use_host_verification': {'key': 'typeProperties.useHostVerification', 'type': 'object'}, - 'use_peer_verification': {'key': 'typeProperties.usePeerVerification', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__(self, **kwargs): - super(OracleServiceCloudLinkedService, self).__init__(**kwargs) - self.host = kwargs.get('host', None) - self.username = kwargs.get('username', None) - self.password = kwargs.get('password', None) - self.use_encrypted_endpoints = kwargs.get('use_encrypted_endpoints', None) - self.use_host_verification = kwargs.get('use_host_verification', None) - self.use_peer_verification = kwargs.get('use_peer_verification', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - self.type = 'OracleServiceCloud' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_service_cloud_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_service_cloud_linked_service_py3.py deleted file mode 100644 index 0936681e6ac5..000000000000 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_service_cloud_linked_service_py3.py +++ /dev/null @@ -1,95 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for -# license information. -# -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is -# regenerated. -# -------------------------------------------------------------------------- - -from .linked_service_py3 import LinkedService - - -class OracleServiceCloudLinkedService(LinkedService): - """Oracle Service Cloud linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] - :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.mgmt.datafactory.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.mgmt.datafactory.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the - Dataset. - :type annotations: list[object] - :param type: Required. Constant filled by server. - :type type: str - :param host: Required. The URL of the Oracle Service Cloud instance. - :type host: object - :param username: Required. The user name that you use to access Oracle - Service Cloud server. - :type username: object - :param password: Required. The password corresponding to the user name - that you provided in the username key. - :type password: ~azure.mgmt.datafactory.models.SecretBase - :param use_encrypted_endpoints: Specifies whether the data source - endpoints are encrypted using HTTPS. The default value is true. Type: - boolean (or Expression with resultType boolean). - :type use_encrypted_endpoints: object - :param use_host_verification: Specifies whether to require the host name - in the server's certificate to match the host name of the server when - connecting over SSL. The default value is true. Type: boolean (or - Expression with resultType boolean). - :type use_host_verification: object - :param use_peer_verification: Specifies whether to verify the identity of - the server when connecting over SSL. The default value is true. Type: - boolean (or Expression with resultType boolean). - :type use_peer_verification: object - :param encrypted_credential: The encrypted credential used for - authentication. Credentials are encrypted using the integration runtime - credential manager. Type: string (or Expression with resultType string). - :type encrypted_credential: object - """ - - _validation = { - 'type': {'required': True}, - 'host': {'required': True}, - 'username': {'required': True}, - 'password': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'type': {'key': 'type', 'type': 'str'}, - 'host': {'key': 'typeProperties.host', 'type': 'object'}, - 'username': {'key': 'typeProperties.username', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'use_encrypted_endpoints': {'key': 'typeProperties.useEncryptedEndpoints', 'type': 'object'}, - 'use_host_verification': {'key': 'typeProperties.useHostVerification', 'type': 'object'}, - 'use_peer_verification': {'key': 'typeProperties.usePeerVerification', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__(self, *, host, username, password, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, use_encrypted_endpoints=None, use_host_verification=None, use_peer_verification=None, encrypted_credential=None, **kwargs) -> None: - super(OracleServiceCloudLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.host = host - self.username = username - self.password = password - self.use_encrypted_endpoints = use_encrypted_endpoints - self.use_host_verification = use_host_verification - self.use_peer_verification = use_peer_verification - self.encrypted_credential = encrypted_credential - self.type = 'OracleServiceCloud' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_service_cloud_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_service_cloud_object_dataset.py deleted file mode 100644 index 4b842b975c70..000000000000 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_service_cloud_object_dataset.py +++ /dev/null @@ -1,62 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for -# license information. -# -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is -# regenerated. -# -------------------------------------------------------------------------- - -from .dataset import Dataset - - -class OracleServiceCloudObjectDataset(Dataset): - """Oracle Service Cloud dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: - array (or Expression with resultType array), itemType: DatasetDataElement. - :type structure: object - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.mgmt.datafactory.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.mgmt.datafactory.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the - Dataset. - :type annotations: list[object] - :param folder: The folder that this Dataset is in. If not specified, - Dataset will appear at the root level. - :type folder: ~azure.mgmt.datafactory.models.DatasetFolder - :param type: Required. Constant filled by server. - :type type: str - """ - - _validation = { - 'linked_service_name': {'required': True}, - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'type': {'key': 'type', 'type': 'str'}, - } - - def __init__(self, **kwargs): - super(OracleServiceCloudObjectDataset, self).__init__(**kwargs) - self.type = 'OracleServiceCloudObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_service_cloud_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_service_cloud_object_dataset_py3.py deleted file mode 100644 index 2e1d1db41586..000000000000 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_service_cloud_object_dataset_py3.py +++ /dev/null @@ -1,62 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for -# license information. -# -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is -# regenerated. -# -------------------------------------------------------------------------- - -from .dataset_py3 import Dataset - - -class OracleServiceCloudObjectDataset(Dataset): - """Oracle Service Cloud dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: - array (or Expression with resultType array), itemType: DatasetDataElement. - :type structure: object - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.mgmt.datafactory.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.mgmt.datafactory.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the - Dataset. - :type annotations: list[object] - :param folder: The folder that this Dataset is in. If not specified, - Dataset will appear at the root level. - :type folder: ~azure.mgmt.datafactory.models.DatasetFolder - :param type: Required. Constant filled by server. - :type type: str - """ - - _validation = { - 'linked_service_name': {'required': True}, - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'type': {'key': 'type', 'type': 'str'}, - } - - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: - super(OracleServiceCloudObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'OracleServiceCloudObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_service_cloud_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_service_cloud_source.py deleted file mode 100644 index 1d362862c737..000000000000 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_service_cloud_source.py +++ /dev/null @@ -1,52 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for -# license information. -# -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is -# regenerated. -# -------------------------------------------------------------------------- - -from .copy_source import CopySource - - -class OracleServiceCloudSource(CopySource): - """A copy activity Oracle Service Cloud source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] - :param source_retry_count: Source retry count. Type: integer (or - Expression with resultType integer). - :type source_retry_count: object - :param source_retry_wait: Source retry wait. Type: string (or Expression - with resultType string), pattern: - ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: object - :param type: Required. Constant filled by server. - :type type: str - :param query: A query to retrieve data from source. Type: string (or - Expression with resultType string). - :type query: object - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'type': {'key': 'type', 'type': 'str'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__(self, **kwargs): - super(OracleServiceCloudSource, self).__init__(**kwargs) - self.query = kwargs.get('query', None) - self.type = 'OracleServiceCloudSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_service_cloud_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_service_cloud_source_py3.py deleted file mode 100644 index 2eebde6c3e50..000000000000 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_service_cloud_source_py3.py +++ /dev/null @@ -1,52 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for -# license information. -# -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is -# regenerated. -# -------------------------------------------------------------------------- - -from .copy_source_py3 import CopySource - - -class OracleServiceCloudSource(CopySource): - """A copy activity Oracle Service Cloud source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] - :param source_retry_count: Source retry count. Type: integer (or - Expression with resultType integer). - :type source_retry_count: object - :param source_retry_wait: Source retry wait. Type: string (or Expression - with resultType string), pattern: - ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: object - :param type: Required. Constant filled by server. - :type type: str - :param query: A query to retrieve data from source. Type: string (or - Expression with resultType string). - :type query: object - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'type': {'key': 'type', 'type': 'str'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__(self, *, additional_properties=None, source_retry_count=None, source_retry_wait=None, query=None, **kwargs) -> None: - super(OracleServiceCloudSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, **kwargs) - self.query = query - self.type = 'OracleServiceCloudSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/rest_service_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/rest_service_dataset.py deleted file mode 100644 index e3ec02526c91..000000000000 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/rest_service_dataset.py +++ /dev/null @@ -1,88 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for -# license information. -# -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is -# regenerated. -# -------------------------------------------------------------------------- - -from .dataset import Dataset - - -class RestServiceDataset(Dataset): - """A Rest service dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: - array (or Expression with resultType array), itemType: DatasetDataElement. - :type structure: object - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.mgmt.datafactory.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.mgmt.datafactory.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the - Dataset. - :type annotations: list[object] - :param folder: The folder that this Dataset is in. If not specified, - Dataset will appear at the root level. - :type folder: ~azure.mgmt.datafactory.models.DatasetFolder - :param type: Required. Constant filled by server. - :type type: str - :param relative_url: The relative URL to the resource that the RESTful API - provides. Type: string (or Expression with resultType string). - :type relative_url: object - :param request_method: The HTTP method used to call the RESTful API. The - default is GET. Type: string (or Expression with resultType string). - :type request_method: object - :param request_body: The HTTP request body to the RESTful API if - requestMethod is POST. Type: string (or Expression with resultType - string). - :type request_body: object - :param additional_headers: The additional HTTP headers in the request to - the RESTful API. Type: string (or Expression with resultType string). - :type additional_headers: object - :param pagination_rules: The pagination rules to compose next page - requests. - :type pagination_rules: object - """ - - _validation = { - 'linked_service_name': {'required': True}, - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'type': {'key': 'type', 'type': 'str'}, - 'relative_url': {'key': 'typeProperties.relativeUrl', 'type': 'object'}, - 'request_method': {'key': 'typeProperties.requestMethod', 'type': 'object'}, - 'request_body': {'key': 'typeProperties.requestBody', 'type': 'object'}, - 'additional_headers': {'key': 'typeProperties.additionalHeaders', 'type': 'object'}, - 'pagination_rules': {'key': 'typeProperties.paginationRules', 'type': 'object'}, - } - - def __init__(self, **kwargs): - super(RestServiceDataset, self).__init__(**kwargs) - self.relative_url = kwargs.get('relative_url', None) - self.request_method = kwargs.get('request_method', None) - self.request_body = kwargs.get('request_body', None) - self.additional_headers = kwargs.get('additional_headers', None) - self.pagination_rules = kwargs.get('pagination_rules', None) - self.type = 'RestResource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/rest_service_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/rest_service_dataset_py3.py deleted file mode 100644 index 0813a1f29624..000000000000 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/rest_service_dataset_py3.py +++ /dev/null @@ -1,88 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for -# license information. -# -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is -# regenerated. -# -------------------------------------------------------------------------- - -from .dataset_py3 import Dataset - - -class RestServiceDataset(Dataset): - """A Rest service dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: - array (or Expression with resultType array), itemType: DatasetDataElement. - :type structure: object - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.mgmt.datafactory.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.mgmt.datafactory.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the - Dataset. - :type annotations: list[object] - :param folder: The folder that this Dataset is in. If not specified, - Dataset will appear at the root level. - :type folder: ~azure.mgmt.datafactory.models.DatasetFolder - :param type: Required. Constant filled by server. - :type type: str - :param relative_url: The relative URL to the resource that the RESTful API - provides. Type: string (or Expression with resultType string). - :type relative_url: object - :param request_method: The HTTP method used to call the RESTful API. The - default is GET. Type: string (or Expression with resultType string). - :type request_method: object - :param request_body: The HTTP request body to the RESTful API if - requestMethod is POST. Type: string (or Expression with resultType - string). - :type request_body: object - :param additional_headers: The additional HTTP headers in the request to - the RESTful API. Type: string (or Expression with resultType string). - :type additional_headers: object - :param pagination_rules: The pagination rules to compose next page - requests. - :type pagination_rules: object - """ - - _validation = { - 'linked_service_name': {'required': True}, - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'type': {'key': 'type', 'type': 'str'}, - 'relative_url': {'key': 'typeProperties.relativeUrl', 'type': 'object'}, - 'request_method': {'key': 'typeProperties.requestMethod', 'type': 'object'}, - 'request_body': {'key': 'typeProperties.requestBody', 'type': 'object'}, - 'additional_headers': {'key': 'typeProperties.additionalHeaders', 'type': 'object'}, - 'pagination_rules': {'key': 'typeProperties.paginationRules', 'type': 'object'}, - } - - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, relative_url=None, request_method=None, request_body=None, additional_headers=None, pagination_rules=None, **kwargs) -> None: - super(RestServiceDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.relative_url = relative_url - self.request_method = request_method - self.request_body = request_body - self.additional_headers = additional_headers - self.pagination_rules = pagination_rules - self.type = 'RestResource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/rest_service_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/rest_service_linked_service.py deleted file mode 100644 index 8515b41394e3..000000000000 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/rest_service_linked_service.py +++ /dev/null @@ -1,107 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for -# license information. -# -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is -# regenerated. -# -------------------------------------------------------------------------- - -from .linked_service import LinkedService - - -class RestServiceLinkedService(LinkedService): - """Rest Service linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] - :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.mgmt.datafactory.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.mgmt.datafactory.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the - Dataset. - :type annotations: list[object] - :param type: Required. Constant filled by server. - :type type: str - :param url: Required. The base URL of the REST service. - :type url: object - :param enable_server_certificate_validation: Whether to validate server - side SSL certificate when connecting to the endpoint.The default value is - true. Type: boolean (or Expression with resultType boolean). - :type enable_server_certificate_validation: object - :param authentication_type: Required. Type of authentication used to - connect to the REST service. Possible values include: 'Anonymous', - 'Basic', 'AadServicePrincial', 'ManagedServiceIdentity' - :type authentication_type: str or - ~azure.mgmt.datafactory.models.RestServiceAuthenticationType - :param user_name: The user name used in Basic authentication type. - :type user_name: object - :param password: The password used in Basic authentication type. - :type password: ~azure.mgmt.datafactory.models.SecretBase - :param service_principal_id: The application’s client ID used in - AadServicePrincipal authentication type. - :type service_principal_id: object - :param service_principal_key: The application’s key used in - AadServicePrincipal authentication type. - :type service_principal_key: ~azure.mgmt.datafactory.models.SecretBase - :param tenant: The tenant information (domain name or tenant ID) used in - AadServicePrincipal authentication type under which your application - resides. - :type tenant: object - :param aad_resource_id: The resource you are requesting authorization to - use. - :type aad_resource_id: object - :param encrypted_credential: The encrypted credential used for - authentication. Credentials are encrypted using the integration runtime - credential manager. Type: string (or Expression with resultType string). - :type encrypted_credential: object - """ - - _validation = { - 'type': {'required': True}, - 'url': {'required': True}, - 'authentication_type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'type': {'key': 'type', 'type': 'str'}, - 'url': {'key': 'typeProperties.url', 'type': 'object'}, - 'enable_server_certificate_validation': {'key': 'typeProperties.enableServerCertificateValidation', 'type': 'object'}, - 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, - 'user_name': {'key': 'typeProperties.userName', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'service_principal_id': {'key': 'typeProperties.servicePrincipalId', 'type': 'object'}, - 'service_principal_key': {'key': 'typeProperties.servicePrincipalKey', 'type': 'SecretBase'}, - 'tenant': {'key': 'typeProperties.tenant', 'type': 'object'}, - 'aad_resource_id': {'key': 'typeProperties.aadResourceId', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__(self, **kwargs): - super(RestServiceLinkedService, self).__init__(**kwargs) - self.url = kwargs.get('url', None) - self.enable_server_certificate_validation = kwargs.get('enable_server_certificate_validation', None) - self.authentication_type = kwargs.get('authentication_type', None) - self.user_name = kwargs.get('user_name', None) - self.password = kwargs.get('password', None) - self.service_principal_id = kwargs.get('service_principal_id', None) - self.service_principal_key = kwargs.get('service_principal_key', None) - self.tenant = kwargs.get('tenant', None) - self.aad_resource_id = kwargs.get('aad_resource_id', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - self.type = 'RestService' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/rest_service_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/rest_service_linked_service_py3.py deleted file mode 100644 index 7e9f9efd0281..000000000000 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/rest_service_linked_service_py3.py +++ /dev/null @@ -1,107 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for -# license information. -# -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is -# regenerated. -# -------------------------------------------------------------------------- - -from .linked_service_py3 import LinkedService - - -class RestServiceLinkedService(LinkedService): - """Rest Service linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] - :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.mgmt.datafactory.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.mgmt.datafactory.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the - Dataset. - :type annotations: list[object] - :param type: Required. Constant filled by server. - :type type: str - :param url: Required. The base URL of the REST service. - :type url: object - :param enable_server_certificate_validation: Whether to validate server - side SSL certificate when connecting to the endpoint.The default value is - true. Type: boolean (or Expression with resultType boolean). - :type enable_server_certificate_validation: object - :param authentication_type: Required. Type of authentication used to - connect to the REST service. Possible values include: 'Anonymous', - 'Basic', 'AadServicePrincial', 'ManagedServiceIdentity' - :type authentication_type: str or - ~azure.mgmt.datafactory.models.RestServiceAuthenticationType - :param user_name: The user name used in Basic authentication type. - :type user_name: object - :param password: The password used in Basic authentication type. - :type password: ~azure.mgmt.datafactory.models.SecretBase - :param service_principal_id: The application’s client ID used in - AadServicePrincipal authentication type. - :type service_principal_id: object - :param service_principal_key: The application’s key used in - AadServicePrincipal authentication type. - :type service_principal_key: ~azure.mgmt.datafactory.models.SecretBase - :param tenant: The tenant information (domain name or tenant ID) used in - AadServicePrincipal authentication type under which your application - resides. - :type tenant: object - :param aad_resource_id: The resource you are requesting authorization to - use. - :type aad_resource_id: object - :param encrypted_credential: The encrypted credential used for - authentication. Credentials are encrypted using the integration runtime - credential manager. Type: string (or Expression with resultType string). - :type encrypted_credential: object - """ - - _validation = { - 'type': {'required': True}, - 'url': {'required': True}, - 'authentication_type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'type': {'key': 'type', 'type': 'str'}, - 'url': {'key': 'typeProperties.url', 'type': 'object'}, - 'enable_server_certificate_validation': {'key': 'typeProperties.enableServerCertificateValidation', 'type': 'object'}, - 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, - 'user_name': {'key': 'typeProperties.userName', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'service_principal_id': {'key': 'typeProperties.servicePrincipalId', 'type': 'object'}, - 'service_principal_key': {'key': 'typeProperties.servicePrincipalKey', 'type': 'SecretBase'}, - 'tenant': {'key': 'typeProperties.tenant', 'type': 'object'}, - 'aad_resource_id': {'key': 'typeProperties.aadResourceId', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__(self, *, url, authentication_type, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, enable_server_certificate_validation=None, user_name=None, password=None, service_principal_id=None, service_principal_key=None, tenant=None, aad_resource_id=None, encrypted_credential=None, **kwargs) -> None: - super(RestServiceLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.url = url - self.enable_server_certificate_validation = enable_server_certificate_validation - self.authentication_type = authentication_type - self.user_name = user_name - self.password = password - self.service_principal_id = service_principal_id - self.service_principal_key = service_principal_key - self.tenant = tenant - self.aad_resource_id = aad_resource_id - self.encrypted_credential = encrypted_credential - self.type = 'RestService' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/rest_service_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/rest_service_source.py deleted file mode 100644 index 9a62dcdbd682..000000000000 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/rest_service_source.py +++ /dev/null @@ -1,62 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for -# license information. -# -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is -# regenerated. -# -------------------------------------------------------------------------- - -from .copy_source import CopySource - - -class RestServiceSource(CopySource): - """A copy activity Rest service source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] - :param source_retry_count: Source retry count. Type: integer (or - Expression with resultType integer). - :type source_retry_count: object - :param source_retry_wait: Source retry wait. Type: string (or Expression - with resultType string), pattern: - ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: object - :param type: Required. Constant filled by server. - :type type: str - :param http_request_timeout: The timeout (TimeSpan) to get an HTTP - response. It is the timeout to get a response, not the timeout to read - response data. Default value: 00:01:40. Type: string (or Expression with - resultType string), pattern: - ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type http_request_timeout: object - :param request_interval: The time to await before sending next page - request. Default value: 00:00:01. Type: string (or Expression with - resultType string), pattern: - ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type request_interval: object - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'type': {'key': 'type', 'type': 'str'}, - 'http_request_timeout': {'key': 'httpRequestTimeout', 'type': 'object'}, - 'request_interval': {'key': 'requestInterval', 'type': 'object'}, - } - - def __init__(self, **kwargs): - super(RestServiceSource, self).__init__(**kwargs) - self.http_request_timeout = kwargs.get('http_request_timeout', None) - self.request_interval = kwargs.get('request_interval', None) - self.type = 'RestServiceSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/rest_service_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/rest_service_source_py3.py deleted file mode 100644 index 85674d829a1f..000000000000 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/rest_service_source_py3.py +++ /dev/null @@ -1,62 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for -# license information. -# -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is -# regenerated. -# -------------------------------------------------------------------------- - -from .copy_source_py3 import CopySource - - -class RestServiceSource(CopySource): - """A copy activity Rest service source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] - :param source_retry_count: Source retry count. Type: integer (or - Expression with resultType integer). - :type source_retry_count: object - :param source_retry_wait: Source retry wait. Type: string (or Expression - with resultType string), pattern: - ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: object - :param type: Required. Constant filled by server. - :type type: str - :param http_request_timeout: The timeout (TimeSpan) to get an HTTP - response. It is the timeout to get a response, not the timeout to read - response data. Default value: 00:01:40. Type: string (or Expression with - resultType string), pattern: - ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type http_request_timeout: object - :param request_interval: The time to await before sending next page - request. Default value: 00:00:01. Type: string (or Expression with - resultType string), pattern: - ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type request_interval: object - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'type': {'key': 'type', 'type': 'str'}, - 'http_request_timeout': {'key': 'httpRequestTimeout', 'type': 'object'}, - 'request_interval': {'key': 'requestInterval', 'type': 'object'}, - } - - def __init__(self, *, additional_properties=None, source_retry_count=None, source_retry_wait=None, http_request_timeout=None, request_interval=None, **kwargs) -> None: - super(RestServiceSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, **kwargs) - self.http_request_timeout = http_request_timeout - self.request_interval = request_interval - self.type = 'RestServiceSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/stored_procedure_parameter.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/stored_procedure_parameter.py index 748cf7cba53c..63d2c3b80aa5 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/stored_procedure_parameter.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/stored_procedure_parameter.py @@ -15,7 +15,9 @@ class StoredProcedureParameter(Model): """SQL stored procedure parameter. - :param value: Stored procedure parameter value. Type: string (or + All required parameters must be populated in order to send to Azure. + + :param value: Required. Stored procedure parameter value. Type: string (or Expression with resultType string). :type value: object :param type: Stored procedure parameter type. Possible values include: @@ -24,6 +26,10 @@ class StoredProcedureParameter(Model): ~azure.mgmt.datafactory.models.StoredProcedureParameterType """ + _validation = { + 'value': {'required': True}, + } + _attribute_map = { 'value': {'key': 'value', 'type': 'object'}, 'type': {'key': 'type', 'type': 'str'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/stored_procedure_parameter_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/stored_procedure_parameter_py3.py index bd967ce52876..f5c55777e0c3 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/stored_procedure_parameter_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/stored_procedure_parameter_py3.py @@ -15,7 +15,9 @@ class StoredProcedureParameter(Model): """SQL stored procedure parameter. - :param value: Stored procedure parameter value. Type: string (or + All required parameters must be populated in order to send to Azure. + + :param value: Required. Stored procedure parameter value. Type: string (or Expression with resultType string). :type value: object :param type: Stored procedure parameter type. Possible values include: @@ -24,12 +26,16 @@ class StoredProcedureParameter(Model): ~azure.mgmt.datafactory.models.StoredProcedureParameterType """ + _validation = { + 'value': {'required': True}, + } + _attribute_map = { 'value': {'key': 'value', 'type': 'object'}, 'type': {'key': 'type', 'type': 'str'}, } - def __init__(self, *, value=None, type=None, **kwargs) -> None: + def __init__(self, *, value, type=None, **kwargs) -> None: super(StoredProcedureParameter, self).__init__(**kwargs) self.value = value self.type = type From a0ce8ce7e78c8b4c09134dbd4a551658b80cbe4d Mon Sep 17 00:00:00 2001 From: Azure SDK for Python bot Date: Mon, 5 Nov 2018 17:52:40 -0800 Subject: [PATCH 28/54] [AutoPR datafactory/resource-manager] Enable delete activity (#3734) * Generated from 5e54670508b90d66ed341ab44aa8dd19d2644a1d Enable delete activity * Generated from cf65697ea020da33e927c257d64339b5504733fa Enable delete activity. * Generated from 5ff291ba3fd2e3bda21c89a3b159531d56a75fbb Update --- .../azure/mgmt/datafactory/models/__init__.py | 3 + .../datafactory/models/delete_activity.py | 69 +++++++++++++++++++ .../datafactory/models/delete_activity_py3.py | 69 +++++++++++++++++++ .../datafactory/models/execution_activity.py | 4 +- .../models/execution_activity_py3.py | 4 +- 5 files changed, 145 insertions(+), 4 deletions(-) create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/delete_activity.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/delete_activity_py3.py diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py index cab29db469a7..4970f33f96e1 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py @@ -295,6 +295,7 @@ from .azure_table_source_py3 import AzureTableSource from .copy_source_py3 import CopySource from .lookup_activity_py3 import LookupActivity + from .delete_activity_py3 import DeleteActivity from .sql_server_stored_procedure_activity_py3 import SqlServerStoredProcedureActivity from .custom_activity_reference_object_py3 import CustomActivityReferenceObject from .custom_activity_py3 import CustomActivity @@ -647,6 +648,7 @@ from .azure_table_source import AzureTableSource from .copy_source import CopySource from .lookup_activity import LookupActivity + from .delete_activity import DeleteActivity from .sql_server_stored_procedure_activity import SqlServerStoredProcedureActivity from .custom_activity_reference_object import CustomActivityReferenceObject from .custom_activity import CustomActivity @@ -1069,6 +1071,7 @@ 'AzureTableSource', 'CopySource', 'LookupActivity', + 'DeleteActivity', 'SqlServerStoredProcedureActivity', 'CustomActivityReferenceObject', 'CustomActivity', diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/delete_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/delete_activity.py new file mode 100644 index 000000000000..62b88a50185f --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/delete_activity.py @@ -0,0 +1,69 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .execution_activity import ExecutionActivity + + +class DeleteActivity(ExecutionActivity): + """Delete activity. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param name: Required. Activity name. + :type name: str + :param description: Activity description. + :type description: str + :param depends_on: Activity depends on condition. + :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param user_properties: Activity user properties. + :type user_properties: list[~azure.mgmt.datafactory.models.UserProperty] + :param type: Required. Constant filled by server. + :type type: str + :param linked_service_name: Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param policy: Activity policy. + :type policy: ~azure.mgmt.datafactory.models.ActivityPolicy + :param recursive: If true, files under the folder path will be deleted + recursively. Default is true. Type: boolean (or Expression with resultType + boolean). + :type recursive: object + :param dataset: Required. Delete activity dataset reference. + :type dataset: ~azure.mgmt.datafactory.models.DatasetReference + """ + + _validation = { + 'name': {'required': True}, + 'type': {'required': True}, + 'dataset': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'name': {'key': 'name', 'type': 'str'}, + 'description': {'key': 'description', 'type': 'str'}, + 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, + 'recursive': {'key': 'typeProperties.recursive', 'type': 'object'}, + 'dataset': {'key': 'typeProperties.dataset', 'type': 'DatasetReference'}, + } + + def __init__(self, **kwargs): + super(DeleteActivity, self).__init__(**kwargs) + self.recursive = kwargs.get('recursive', None) + self.dataset = kwargs.get('dataset', None) + self.type = 'Delete' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/delete_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/delete_activity_py3.py new file mode 100644 index 000000000000..f05ce0832ca8 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/delete_activity_py3.py @@ -0,0 +1,69 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .execution_activity_py3 import ExecutionActivity + + +class DeleteActivity(ExecutionActivity): + """Delete activity. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param name: Required. Activity name. + :type name: str + :param description: Activity description. + :type description: str + :param depends_on: Activity depends on condition. + :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param user_properties: Activity user properties. + :type user_properties: list[~azure.mgmt.datafactory.models.UserProperty] + :param type: Required. Constant filled by server. + :type type: str + :param linked_service_name: Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param policy: Activity policy. + :type policy: ~azure.mgmt.datafactory.models.ActivityPolicy + :param recursive: If true, files under the folder path will be deleted + recursively. Default is true. Type: boolean (or Expression with resultType + boolean). + :type recursive: object + :param dataset: Required. Delete activity dataset reference. + :type dataset: ~azure.mgmt.datafactory.models.DatasetReference + """ + + _validation = { + 'name': {'required': True}, + 'type': {'required': True}, + 'dataset': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'name': {'key': 'name', 'type': 'str'}, + 'description': {'key': 'description', 'type': 'str'}, + 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, + 'recursive': {'key': 'typeProperties.recursive', 'type': 'object'}, + 'dataset': {'key': 'typeProperties.dataset', 'type': 'DatasetReference'}, + } + + def __init__(self, *, name: str, dataset, additional_properties=None, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, recursive=None, **kwargs) -> None: + super(DeleteActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) + self.recursive = recursive + self.dataset = dataset + self.type = 'Delete' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execution_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execution_activity.py index a9362afc43a9..e061d595a3ba 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execution_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execution_activity.py @@ -19,7 +19,7 @@ class ExecutionActivity(Activity): sub-classes are: DatabricksSparkPythonActivity, DatabricksSparkJarActivity, DatabricksNotebookActivity, DataLakeAnalyticsUSQLActivity, AzureMLUpdateResourceActivity, AzureMLBatchExecutionActivity, - GetMetadataActivity, WebActivity, LookupActivity, + GetMetadataActivity, WebActivity, LookupActivity, DeleteActivity, SqlServerStoredProcedureActivity, CustomActivity, ExecuteSSISPackageActivity, HDInsightSparkActivity, HDInsightStreamingActivity, HDInsightMapReduceActivity, @@ -64,7 +64,7 @@ class ExecutionActivity(Activity): } _subtype_map = { - 'type': {'DatabricksSparkPython': 'DatabricksSparkPythonActivity', 'DatabricksSparkJar': 'DatabricksSparkJarActivity', 'DatabricksNotebook': 'DatabricksNotebookActivity', 'DataLakeAnalyticsU-SQL': 'DataLakeAnalyticsUSQLActivity', 'AzureMLUpdateResource': 'AzureMLUpdateResourceActivity', 'AzureMLBatchExecution': 'AzureMLBatchExecutionActivity', 'GetMetadata': 'GetMetadataActivity', 'WebActivity': 'WebActivity', 'Lookup': 'LookupActivity', 'SqlServerStoredProcedure': 'SqlServerStoredProcedureActivity', 'Custom': 'CustomActivity', 'ExecuteSSISPackage': 'ExecuteSSISPackageActivity', 'HDInsightSpark': 'HDInsightSparkActivity', 'HDInsightStreaming': 'HDInsightStreamingActivity', 'HDInsightMapReduce': 'HDInsightMapReduceActivity', 'HDInsightPig': 'HDInsightPigActivity', 'HDInsightHive': 'HDInsightHiveActivity', 'Copy': 'CopyActivity'} + 'type': {'DatabricksSparkPython': 'DatabricksSparkPythonActivity', 'DatabricksSparkJar': 'DatabricksSparkJarActivity', 'DatabricksNotebook': 'DatabricksNotebookActivity', 'DataLakeAnalyticsU-SQL': 'DataLakeAnalyticsUSQLActivity', 'AzureMLUpdateResource': 'AzureMLUpdateResourceActivity', 'AzureMLBatchExecution': 'AzureMLBatchExecutionActivity', 'GetMetadata': 'GetMetadataActivity', 'WebActivity': 'WebActivity', 'Lookup': 'LookupActivity', 'Delete': 'DeleteActivity', 'SqlServerStoredProcedure': 'SqlServerStoredProcedureActivity', 'Custom': 'CustomActivity', 'ExecuteSSISPackage': 'ExecuteSSISPackageActivity', 'HDInsightSpark': 'HDInsightSparkActivity', 'HDInsightStreaming': 'HDInsightStreamingActivity', 'HDInsightMapReduce': 'HDInsightMapReduceActivity', 'HDInsightPig': 'HDInsightPigActivity', 'HDInsightHive': 'HDInsightHiveActivity', 'Copy': 'CopyActivity'} } def __init__(self, **kwargs): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execution_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execution_activity_py3.py index 8762d742eb93..e3fcc573bc48 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execution_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execution_activity_py3.py @@ -19,7 +19,7 @@ class ExecutionActivity(Activity): sub-classes are: DatabricksSparkPythonActivity, DatabricksSparkJarActivity, DatabricksNotebookActivity, DataLakeAnalyticsUSQLActivity, AzureMLUpdateResourceActivity, AzureMLBatchExecutionActivity, - GetMetadataActivity, WebActivity, LookupActivity, + GetMetadataActivity, WebActivity, LookupActivity, DeleteActivity, SqlServerStoredProcedureActivity, CustomActivity, ExecuteSSISPackageActivity, HDInsightSparkActivity, HDInsightStreamingActivity, HDInsightMapReduceActivity, @@ -64,7 +64,7 @@ class ExecutionActivity(Activity): } _subtype_map = { - 'type': {'DatabricksSparkPython': 'DatabricksSparkPythonActivity', 'DatabricksSparkJar': 'DatabricksSparkJarActivity', 'DatabricksNotebook': 'DatabricksNotebookActivity', 'DataLakeAnalyticsU-SQL': 'DataLakeAnalyticsUSQLActivity', 'AzureMLUpdateResource': 'AzureMLUpdateResourceActivity', 'AzureMLBatchExecution': 'AzureMLBatchExecutionActivity', 'GetMetadata': 'GetMetadataActivity', 'WebActivity': 'WebActivity', 'Lookup': 'LookupActivity', 'SqlServerStoredProcedure': 'SqlServerStoredProcedureActivity', 'Custom': 'CustomActivity', 'ExecuteSSISPackage': 'ExecuteSSISPackageActivity', 'HDInsightSpark': 'HDInsightSparkActivity', 'HDInsightStreaming': 'HDInsightStreamingActivity', 'HDInsightMapReduce': 'HDInsightMapReduceActivity', 'HDInsightPig': 'HDInsightPigActivity', 'HDInsightHive': 'HDInsightHiveActivity', 'Copy': 'CopyActivity'} + 'type': {'DatabricksSparkPython': 'DatabricksSparkPythonActivity', 'DatabricksSparkJar': 'DatabricksSparkJarActivity', 'DatabricksNotebook': 'DatabricksNotebookActivity', 'DataLakeAnalyticsU-SQL': 'DataLakeAnalyticsUSQLActivity', 'AzureMLUpdateResource': 'AzureMLUpdateResourceActivity', 'AzureMLBatchExecution': 'AzureMLBatchExecutionActivity', 'GetMetadata': 'GetMetadataActivity', 'WebActivity': 'WebActivity', 'Lookup': 'LookupActivity', 'Delete': 'DeleteActivity', 'SqlServerStoredProcedure': 'SqlServerStoredProcedureActivity', 'Custom': 'CustomActivity', 'ExecuteSSISPackage': 'ExecuteSSISPackageActivity', 'HDInsightSpark': 'HDInsightSparkActivity', 'HDInsightStreaming': 'HDInsightStreamingActivity', 'HDInsightMapReduce': 'HDInsightMapReduceActivity', 'HDInsightPig': 'HDInsightPigActivity', 'HDInsightHive': 'HDInsightHiveActivity', 'Copy': 'CopyActivity'} } def __init__(self, *, name: str, additional_properties=None, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, **kwargs) -> None: From 843048b4387138d5ccfcc21c31f0d76ade314de4 Mon Sep 17 00:00:00 2001 From: Azure SDK for Python bot Date: Mon, 5 Nov 2018 18:43:14 -0800 Subject: [PATCH 29/54] [AutoPR datafactory/resource-manager] [DataFactory]Add sensitive fields extracted from connection string in swagger for AKV reference (#3730) * Generated from 5e66be9bba64b2b337e79273326b1758dbedb893 Add sensitive fields extracted from connection string in swagger for akv reference * Generated from f0f730dbcd8cc298ba9332b3e607b9a0e7a2d665 Change sasUri type to object as it supports string, SecureString and AzureKeyVaultSecretReference, same with connection string. * Generated from f0f730dbcd8cc298ba9332b3e607b9a0e7a2d665 Change sasUri type to object as it supports string, SecureString and AzureKeyVaultSecretReference, same with connection string. --- .../azure_blob_storage_linked_service.py | 19 ++++++++++++++--- .../azure_blob_storage_linked_service_py3.py | 21 +++++++++++++++---- .../models/azure_my_sql_linked_service.py | 6 ++++++ .../models/azure_my_sql_linked_service_py3.py | 8 ++++++- .../azure_postgre_sql_linked_service.py | 6 ++++++ .../azure_postgre_sql_linked_service_py3.py | 8 ++++++- .../azure_sql_database_linked_service.py | 6 ++++++ .../azure_sql_database_linked_service_py3.py | 8 ++++++- .../models/azure_sql_dw_linked_service.py | 6 ++++++ .../models/azure_sql_dw_linked_service_py3.py | 8 ++++++- .../models/azure_storage_linked_service.py | 19 ++++++++++++++--- .../azure_storage_linked_service_py3.py | 21 +++++++++++++++---- .../azure_table_storage_linked_service.py | 19 ++++++++++++++--- .../azure_table_storage_linked_service_py3.py | 21 +++++++++++++++---- .../models/cosmos_db_linked_service.py | 6 ++++++ .../models/cosmos_db_linked_service_py3.py | 8 ++++++- .../models/couchbase_linked_service.py | 6 ++++++ .../models/couchbase_linked_service_py3.py | 8 ++++++- .../models/drill_linked_service.py | 5 +++++ .../models/drill_linked_service_py3.py | 7 ++++++- .../models/greenplum_linked_service.py | 5 +++++ .../models/greenplum_linked_service_py3.py | 7 ++++++- .../models/maria_db_linked_service.py | 5 +++++ .../models/maria_db_linked_service_py3.py | 7 ++++++- .../models/my_sql_linked_service.py | 6 ++++++ .../models/my_sql_linked_service_py3.py | 8 ++++++- .../models/netezza_linked_service.py | 5 +++++ .../models/netezza_linked_service_py3.py | 7 ++++++- .../models/oracle_linked_service.py | 6 ++++++ .../models/oracle_linked_service_py3.py | 8 ++++++- .../models/postgre_sql_linked_service.py | 6 ++++++ .../models/postgre_sql_linked_service_py3.py | 8 ++++++- .../models/vertica_linked_service.py | 5 +++++ .../models/vertica_linked_service_py3.py | 7 ++++++- 34 files changed, 271 insertions(+), 35 deletions(-) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_storage_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_storage_linked_service.py index d9a129530f53..e4466c4ce9c9 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_storage_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_storage_linked_service.py @@ -37,9 +37,18 @@ class AzureBlobStorageLinkedService(LinkedService): with sasUri, serviceEndpoint property. Type: string, SecureString or AzureKeyVaultSecretReference. :type connection_string: object + :param account_key: The Azure key vault secret reference of accountKey in + connection string. + :type account_key: + ~azure.mgmt.datafactory.models.AzureKeyVaultSecretReference :param sas_uri: SAS URI of the Azure Blob Storage resource. It is mutually - exclusive with connectionString, serviceEndpoint property. - :type sas_uri: ~azure.mgmt.datafactory.models.SecretBase + exclusive with connectionString, serviceEndpoint property. Type: string, + SecureString or AzureKeyVaultSecretReference. + :type sas_uri: object + :param sas_token: The Azure key vault secret reference of sasToken in sas + uri. + :type sas_token: + ~azure.mgmt.datafactory.models.AzureKeyVaultSecretReference :param service_endpoint: Blob service endpoint of the Azure Blob Storage resource. It is mutually exclusive with connectionString, sasUri property. :type service_endpoint: str @@ -71,7 +80,9 @@ class AzureBlobStorageLinkedService(LinkedService): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'type': {'key': 'type', 'type': 'str'}, 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, - 'sas_uri': {'key': 'typeProperties.sasUri', 'type': 'SecretBase'}, + 'account_key': {'key': 'typeProperties.accountKey', 'type': 'AzureKeyVaultSecretReference'}, + 'sas_uri': {'key': 'typeProperties.sasUri', 'type': 'object'}, + 'sas_token': {'key': 'typeProperties.sasToken', 'type': 'AzureKeyVaultSecretReference'}, 'service_endpoint': {'key': 'typeProperties.serviceEndpoint', 'type': 'str'}, 'service_principal_id': {'key': 'typeProperties.servicePrincipalId', 'type': 'object'}, 'service_principal_key': {'key': 'typeProperties.servicePrincipalKey', 'type': 'SecretBase'}, @@ -82,7 +93,9 @@ class AzureBlobStorageLinkedService(LinkedService): def __init__(self, **kwargs): super(AzureBlobStorageLinkedService, self).__init__(**kwargs) self.connection_string = kwargs.get('connection_string', None) + self.account_key = kwargs.get('account_key', None) self.sas_uri = kwargs.get('sas_uri', None) + self.sas_token = kwargs.get('sas_token', None) self.service_endpoint = kwargs.get('service_endpoint', None) self.service_principal_id = kwargs.get('service_principal_id', None) self.service_principal_key = kwargs.get('service_principal_key', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_storage_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_storage_linked_service_py3.py index 39efbab9e83e..4587e0c95dad 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_storage_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_storage_linked_service_py3.py @@ -37,9 +37,18 @@ class AzureBlobStorageLinkedService(LinkedService): with sasUri, serviceEndpoint property. Type: string, SecureString or AzureKeyVaultSecretReference. :type connection_string: object + :param account_key: The Azure key vault secret reference of accountKey in + connection string. + :type account_key: + ~azure.mgmt.datafactory.models.AzureKeyVaultSecretReference :param sas_uri: SAS URI of the Azure Blob Storage resource. It is mutually - exclusive with connectionString, serviceEndpoint property. - :type sas_uri: ~azure.mgmt.datafactory.models.SecretBase + exclusive with connectionString, serviceEndpoint property. Type: string, + SecureString or AzureKeyVaultSecretReference. + :type sas_uri: object + :param sas_token: The Azure key vault secret reference of sasToken in sas + uri. + :type sas_token: + ~azure.mgmt.datafactory.models.AzureKeyVaultSecretReference :param service_endpoint: Blob service endpoint of the Azure Blob Storage resource. It is mutually exclusive with connectionString, sasUri property. :type service_endpoint: str @@ -71,7 +80,9 @@ class AzureBlobStorageLinkedService(LinkedService): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'type': {'key': 'type', 'type': 'str'}, 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, - 'sas_uri': {'key': 'typeProperties.sasUri', 'type': 'SecretBase'}, + 'account_key': {'key': 'typeProperties.accountKey', 'type': 'AzureKeyVaultSecretReference'}, + 'sas_uri': {'key': 'typeProperties.sasUri', 'type': 'object'}, + 'sas_token': {'key': 'typeProperties.sasToken', 'type': 'AzureKeyVaultSecretReference'}, 'service_endpoint': {'key': 'typeProperties.serviceEndpoint', 'type': 'str'}, 'service_principal_id': {'key': 'typeProperties.servicePrincipalId', 'type': 'object'}, 'service_principal_key': {'key': 'typeProperties.servicePrincipalKey', 'type': 'SecretBase'}, @@ -79,10 +90,12 @@ class AzureBlobStorageLinkedService(LinkedService): 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'str'}, } - def __init__(self, *, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, connection_string=None, sas_uri=None, service_endpoint: str=None, service_principal_id=None, service_principal_key=None, tenant=None, encrypted_credential: str=None, **kwargs) -> None: + def __init__(self, *, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, connection_string=None, account_key=None, sas_uri=None, sas_token=None, service_endpoint: str=None, service_principal_id=None, service_principal_key=None, tenant=None, encrypted_credential: str=None, **kwargs) -> None: super(AzureBlobStorageLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) self.connection_string = connection_string + self.account_key = account_key self.sas_uri = sas_uri + self.sas_token = sas_token self.service_endpoint = service_endpoint self.service_principal_id = service_principal_id self.service_principal_key = service_principal_key diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_linked_service.py index aa8f74d8ecda..64a072f1f38b 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_linked_service.py @@ -36,6 +36,10 @@ class AzureMySqlLinkedService(LinkedService): :param connection_string: Required. The connection string. Type: string, SecureString or AzureKeyVaultSecretReference. :type connection_string: object + :param password: The Azure key vault secret reference of password in + connection string. + :type password: + ~azure.mgmt.datafactory.models.AzureKeyVaultSecretReference :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -55,11 +59,13 @@ class AzureMySqlLinkedService(LinkedService): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'type': {'key': 'type', 'type': 'str'}, 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, + 'password': {'key': 'typeProperties.password', 'type': 'AzureKeyVaultSecretReference'}, 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } def __init__(self, **kwargs): super(AzureMySqlLinkedService, self).__init__(**kwargs) self.connection_string = kwargs.get('connection_string', None) + self.password = kwargs.get('password', None) self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'AzureMySql' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_linked_service_py3.py index 928a621ff74b..dcf4861da573 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_linked_service_py3.py @@ -36,6 +36,10 @@ class AzureMySqlLinkedService(LinkedService): :param connection_string: Required. The connection string. Type: string, SecureString or AzureKeyVaultSecretReference. :type connection_string: object + :param password: The Azure key vault secret reference of password in + connection string. + :type password: + ~azure.mgmt.datafactory.models.AzureKeyVaultSecretReference :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -55,11 +59,13 @@ class AzureMySqlLinkedService(LinkedService): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'type': {'key': 'type', 'type': 'str'}, 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, + 'password': {'key': 'typeProperties.password', 'type': 'AzureKeyVaultSecretReference'}, 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, *, connection_string, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, encrypted_credential=None, **kwargs) -> None: + def __init__(self, *, connection_string, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, password=None, encrypted_credential=None, **kwargs) -> None: super(AzureMySqlLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) self.connection_string = connection_string + self.password = password self.encrypted_credential = encrypted_credential self.type = 'AzureMySql' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_linked_service.py index 168ba5f589d7..89c9b29cdcde 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_linked_service.py @@ -36,6 +36,10 @@ class AzurePostgreSqlLinkedService(LinkedService): :param connection_string: An ODBC connection string. Type: string, SecureString or AzureKeyVaultSecretReference. :type connection_string: object + :param password: The Azure key vault secret reference of password in + connection string. + :type password: + ~azure.mgmt.datafactory.models.AzureKeyVaultSecretReference :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -54,11 +58,13 @@ class AzurePostgreSqlLinkedService(LinkedService): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'type': {'key': 'type', 'type': 'str'}, 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, + 'password': {'key': 'typeProperties.password', 'type': 'AzureKeyVaultSecretReference'}, 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } def __init__(self, **kwargs): super(AzurePostgreSqlLinkedService, self).__init__(**kwargs) self.connection_string = kwargs.get('connection_string', None) + self.password = kwargs.get('password', None) self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'AzurePostgreSql' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_linked_service_py3.py index e64ab78696d4..e885498530ed 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_linked_service_py3.py @@ -36,6 +36,10 @@ class AzurePostgreSqlLinkedService(LinkedService): :param connection_string: An ODBC connection string. Type: string, SecureString or AzureKeyVaultSecretReference. :type connection_string: object + :param password: The Azure key vault secret reference of password in + connection string. + :type password: + ~azure.mgmt.datafactory.models.AzureKeyVaultSecretReference :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -54,11 +58,13 @@ class AzurePostgreSqlLinkedService(LinkedService): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'type': {'key': 'type', 'type': 'str'}, 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, + 'password': {'key': 'typeProperties.password', 'type': 'AzureKeyVaultSecretReference'}, 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, *, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, connection_string=None, encrypted_credential=None, **kwargs) -> None: + def __init__(self, *, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, connection_string=None, password=None, encrypted_credential=None, **kwargs) -> None: super(AzurePostgreSqlLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) self.connection_string = connection_string + self.password = password self.encrypted_credential = encrypted_credential self.type = 'AzurePostgreSql' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_database_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_database_linked_service.py index c1ecc189d194..68ad549ed733 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_database_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_database_linked_service.py @@ -36,6 +36,10 @@ class AzureSqlDatabaseLinkedService(LinkedService): :param connection_string: Required. The connection string. Type: string, SecureString or AzureKeyVaultSecretReference. :type connection_string: object + :param password: The Azure key vault secret reference of password in + connection string. + :type password: + ~azure.mgmt.datafactory.models.AzureKeyVaultSecretReference :param service_principal_id: The ID of the service principal used to authenticate against Azure SQL Database. Type: string (or Expression with resultType string). @@ -65,6 +69,7 @@ class AzureSqlDatabaseLinkedService(LinkedService): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'type': {'key': 'type', 'type': 'str'}, 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, + 'password': {'key': 'typeProperties.password', 'type': 'AzureKeyVaultSecretReference'}, 'service_principal_id': {'key': 'typeProperties.servicePrincipalId', 'type': 'object'}, 'service_principal_key': {'key': 'typeProperties.servicePrincipalKey', 'type': 'SecretBase'}, 'tenant': {'key': 'typeProperties.tenant', 'type': 'object'}, @@ -74,6 +79,7 @@ class AzureSqlDatabaseLinkedService(LinkedService): def __init__(self, **kwargs): super(AzureSqlDatabaseLinkedService, self).__init__(**kwargs) self.connection_string = kwargs.get('connection_string', None) + self.password = kwargs.get('password', None) self.service_principal_id = kwargs.get('service_principal_id', None) self.service_principal_key = kwargs.get('service_principal_key', None) self.tenant = kwargs.get('tenant', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_database_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_database_linked_service_py3.py index 80286301c6a8..afd58ae43354 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_database_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_database_linked_service_py3.py @@ -36,6 +36,10 @@ class AzureSqlDatabaseLinkedService(LinkedService): :param connection_string: Required. The connection string. Type: string, SecureString or AzureKeyVaultSecretReference. :type connection_string: object + :param password: The Azure key vault secret reference of password in + connection string. + :type password: + ~azure.mgmt.datafactory.models.AzureKeyVaultSecretReference :param service_principal_id: The ID of the service principal used to authenticate against Azure SQL Database. Type: string (or Expression with resultType string). @@ -65,15 +69,17 @@ class AzureSqlDatabaseLinkedService(LinkedService): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'type': {'key': 'type', 'type': 'str'}, 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, + 'password': {'key': 'typeProperties.password', 'type': 'AzureKeyVaultSecretReference'}, 'service_principal_id': {'key': 'typeProperties.servicePrincipalId', 'type': 'object'}, 'service_principal_key': {'key': 'typeProperties.servicePrincipalKey', 'type': 'SecretBase'}, 'tenant': {'key': 'typeProperties.tenant', 'type': 'object'}, 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, *, connection_string, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, service_principal_id=None, service_principal_key=None, tenant=None, encrypted_credential=None, **kwargs) -> None: + def __init__(self, *, connection_string, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, password=None, service_principal_id=None, service_principal_key=None, tenant=None, encrypted_credential=None, **kwargs) -> None: super(AzureSqlDatabaseLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) self.connection_string = connection_string + self.password = password self.service_principal_id = service_principal_id self.service_principal_key = service_principal_key self.tenant = tenant diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_dw_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_dw_linked_service.py index ef56c4142321..d4aa961cb424 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_dw_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_dw_linked_service.py @@ -37,6 +37,10 @@ class AzureSqlDWLinkedService(LinkedService): SecureString or AzureKeyVaultSecretReference. Type: string, SecureString or AzureKeyVaultSecretReference. :type connection_string: object + :param password: The Azure key vault secret reference of password in + connection string. + :type password: + ~azure.mgmt.datafactory.models.AzureKeyVaultSecretReference :param service_principal_id: The ID of the service principal used to authenticate against Azure SQL Data Warehouse. Type: string (or Expression with resultType string). @@ -66,6 +70,7 @@ class AzureSqlDWLinkedService(LinkedService): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'type': {'key': 'type', 'type': 'str'}, 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, + 'password': {'key': 'typeProperties.password', 'type': 'AzureKeyVaultSecretReference'}, 'service_principal_id': {'key': 'typeProperties.servicePrincipalId', 'type': 'object'}, 'service_principal_key': {'key': 'typeProperties.servicePrincipalKey', 'type': 'SecretBase'}, 'tenant': {'key': 'typeProperties.tenant', 'type': 'object'}, @@ -75,6 +80,7 @@ class AzureSqlDWLinkedService(LinkedService): def __init__(self, **kwargs): super(AzureSqlDWLinkedService, self).__init__(**kwargs) self.connection_string = kwargs.get('connection_string', None) + self.password = kwargs.get('password', None) self.service_principal_id = kwargs.get('service_principal_id', None) self.service_principal_key = kwargs.get('service_principal_key', None) self.tenant = kwargs.get('tenant', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_dw_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_dw_linked_service_py3.py index 2f8d316510aa..a78551dff273 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_dw_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_dw_linked_service_py3.py @@ -37,6 +37,10 @@ class AzureSqlDWLinkedService(LinkedService): SecureString or AzureKeyVaultSecretReference. Type: string, SecureString or AzureKeyVaultSecretReference. :type connection_string: object + :param password: The Azure key vault secret reference of password in + connection string. + :type password: + ~azure.mgmt.datafactory.models.AzureKeyVaultSecretReference :param service_principal_id: The ID of the service principal used to authenticate against Azure SQL Data Warehouse. Type: string (or Expression with resultType string). @@ -66,15 +70,17 @@ class AzureSqlDWLinkedService(LinkedService): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'type': {'key': 'type', 'type': 'str'}, 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, + 'password': {'key': 'typeProperties.password', 'type': 'AzureKeyVaultSecretReference'}, 'service_principal_id': {'key': 'typeProperties.servicePrincipalId', 'type': 'object'}, 'service_principal_key': {'key': 'typeProperties.servicePrincipalKey', 'type': 'SecretBase'}, 'tenant': {'key': 'typeProperties.tenant', 'type': 'object'}, 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, *, connection_string, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, service_principal_id=None, service_principal_key=None, tenant=None, encrypted_credential=None, **kwargs) -> None: + def __init__(self, *, connection_string, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, password=None, service_principal_id=None, service_principal_key=None, tenant=None, encrypted_credential=None, **kwargs) -> None: super(AzureSqlDWLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) self.connection_string = connection_string + self.password = password self.service_principal_id = service_principal_id self.service_principal_key = service_principal_key self.tenant = tenant diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_storage_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_storage_linked_service.py index d59dc43d5260..711b09a80004 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_storage_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_storage_linked_service.py @@ -37,9 +37,18 @@ class AzureStorageLinkedService(LinkedService): with sasUri property. Type: string, SecureString or AzureKeyVaultSecretReference. :type connection_string: object + :param account_key: The Azure key vault secret reference of accountKey in + connection string. + :type account_key: + ~azure.mgmt.datafactory.models.AzureKeyVaultSecretReference :param sas_uri: SAS URI of the Azure Storage resource. It is mutually - exclusive with connectionString property. - :type sas_uri: ~azure.mgmt.datafactory.models.SecretBase + exclusive with connectionString property. Type: string, SecureString or + AzureKeyVaultSecretReference. + :type sas_uri: object + :param sas_token: The Azure key vault secret reference of sasToken in sas + uri. + :type sas_token: + ~azure.mgmt.datafactory.models.AzureKeyVaultSecretReference :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -58,13 +67,17 @@ class AzureStorageLinkedService(LinkedService): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'type': {'key': 'type', 'type': 'str'}, 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, - 'sas_uri': {'key': 'typeProperties.sasUri', 'type': 'SecretBase'}, + 'account_key': {'key': 'typeProperties.accountKey', 'type': 'AzureKeyVaultSecretReference'}, + 'sas_uri': {'key': 'typeProperties.sasUri', 'type': 'object'}, + 'sas_token': {'key': 'typeProperties.sasToken', 'type': 'AzureKeyVaultSecretReference'}, 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'str'}, } def __init__(self, **kwargs): super(AzureStorageLinkedService, self).__init__(**kwargs) self.connection_string = kwargs.get('connection_string', None) + self.account_key = kwargs.get('account_key', None) self.sas_uri = kwargs.get('sas_uri', None) + self.sas_token = kwargs.get('sas_token', None) self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'AzureStorage' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_storage_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_storage_linked_service_py3.py index e508de34ee21..428fb82e871a 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_storage_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_storage_linked_service_py3.py @@ -37,9 +37,18 @@ class AzureStorageLinkedService(LinkedService): with sasUri property. Type: string, SecureString or AzureKeyVaultSecretReference. :type connection_string: object + :param account_key: The Azure key vault secret reference of accountKey in + connection string. + :type account_key: + ~azure.mgmt.datafactory.models.AzureKeyVaultSecretReference :param sas_uri: SAS URI of the Azure Storage resource. It is mutually - exclusive with connectionString property. - :type sas_uri: ~azure.mgmt.datafactory.models.SecretBase + exclusive with connectionString property. Type: string, SecureString or + AzureKeyVaultSecretReference. + :type sas_uri: object + :param sas_token: The Azure key vault secret reference of sasToken in sas + uri. + :type sas_token: + ~azure.mgmt.datafactory.models.AzureKeyVaultSecretReference :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -58,13 +67,17 @@ class AzureStorageLinkedService(LinkedService): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'type': {'key': 'type', 'type': 'str'}, 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, - 'sas_uri': {'key': 'typeProperties.sasUri', 'type': 'SecretBase'}, + 'account_key': {'key': 'typeProperties.accountKey', 'type': 'AzureKeyVaultSecretReference'}, + 'sas_uri': {'key': 'typeProperties.sasUri', 'type': 'object'}, + 'sas_token': {'key': 'typeProperties.sasToken', 'type': 'AzureKeyVaultSecretReference'}, 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'str'}, } - def __init__(self, *, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, connection_string=None, sas_uri=None, encrypted_credential: str=None, **kwargs) -> None: + def __init__(self, *, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, connection_string=None, account_key=None, sas_uri=None, sas_token=None, encrypted_credential: str=None, **kwargs) -> None: super(AzureStorageLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) self.connection_string = connection_string + self.account_key = account_key self.sas_uri = sas_uri + self.sas_token = sas_token self.encrypted_credential = encrypted_credential self.type = 'AzureStorage' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_table_storage_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_table_storage_linked_service.py index 8fdd6e339123..152fae6368a6 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_table_storage_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_table_storage_linked_service.py @@ -37,9 +37,18 @@ class AzureTableStorageLinkedService(LinkedService): with sasUri property. Type: string, SecureString or AzureKeyVaultSecretReference. :type connection_string: object + :param account_key: The Azure key vault secret reference of accountKey in + connection string. + :type account_key: + ~azure.mgmt.datafactory.models.AzureKeyVaultSecretReference :param sas_uri: SAS URI of the Azure Storage resource. It is mutually - exclusive with connectionString property. - :type sas_uri: ~azure.mgmt.datafactory.models.SecretBase + exclusive with connectionString property. Type: string, SecureString or + AzureKeyVaultSecretReference. + :type sas_uri: object + :param sas_token: The Azure key vault secret reference of sasToken in sas + uri. + :type sas_token: + ~azure.mgmt.datafactory.models.AzureKeyVaultSecretReference :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -58,13 +67,17 @@ class AzureTableStorageLinkedService(LinkedService): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'type': {'key': 'type', 'type': 'str'}, 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, - 'sas_uri': {'key': 'typeProperties.sasUri', 'type': 'SecretBase'}, + 'account_key': {'key': 'typeProperties.accountKey', 'type': 'AzureKeyVaultSecretReference'}, + 'sas_uri': {'key': 'typeProperties.sasUri', 'type': 'object'}, + 'sas_token': {'key': 'typeProperties.sasToken', 'type': 'AzureKeyVaultSecretReference'}, 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'str'}, } def __init__(self, **kwargs): super(AzureTableStorageLinkedService, self).__init__(**kwargs) self.connection_string = kwargs.get('connection_string', None) + self.account_key = kwargs.get('account_key', None) self.sas_uri = kwargs.get('sas_uri', None) + self.sas_token = kwargs.get('sas_token', None) self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'AzureTableStorage' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_table_storage_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_table_storage_linked_service_py3.py index 2a1f3df9222d..533ad3509483 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_table_storage_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_table_storage_linked_service_py3.py @@ -37,9 +37,18 @@ class AzureTableStorageLinkedService(LinkedService): with sasUri property. Type: string, SecureString or AzureKeyVaultSecretReference. :type connection_string: object + :param account_key: The Azure key vault secret reference of accountKey in + connection string. + :type account_key: + ~azure.mgmt.datafactory.models.AzureKeyVaultSecretReference :param sas_uri: SAS URI of the Azure Storage resource. It is mutually - exclusive with connectionString property. - :type sas_uri: ~azure.mgmt.datafactory.models.SecretBase + exclusive with connectionString property. Type: string, SecureString or + AzureKeyVaultSecretReference. + :type sas_uri: object + :param sas_token: The Azure key vault secret reference of sasToken in sas + uri. + :type sas_token: + ~azure.mgmt.datafactory.models.AzureKeyVaultSecretReference :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -58,13 +67,17 @@ class AzureTableStorageLinkedService(LinkedService): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'type': {'key': 'type', 'type': 'str'}, 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, - 'sas_uri': {'key': 'typeProperties.sasUri', 'type': 'SecretBase'}, + 'account_key': {'key': 'typeProperties.accountKey', 'type': 'AzureKeyVaultSecretReference'}, + 'sas_uri': {'key': 'typeProperties.sasUri', 'type': 'object'}, + 'sas_token': {'key': 'typeProperties.sasToken', 'type': 'AzureKeyVaultSecretReference'}, 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'str'}, } - def __init__(self, *, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, connection_string=None, sas_uri=None, encrypted_credential: str=None, **kwargs) -> None: + def __init__(self, *, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, connection_string=None, account_key=None, sas_uri=None, sas_token=None, encrypted_credential: str=None, **kwargs) -> None: super(AzureTableStorageLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) self.connection_string = connection_string + self.account_key = account_key self.sas_uri = sas_uri + self.sas_token = sas_token self.encrypted_credential = encrypted_credential self.type = 'AzureTableStorage' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cosmos_db_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cosmos_db_linked_service.py index 98d6e9a81eea..ed9136eee5fe 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cosmos_db_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cosmos_db_linked_service.py @@ -36,6 +36,10 @@ class CosmosDbLinkedService(LinkedService): :param connection_string: Required. The connection string. Type: string, SecureString or AzureKeyVaultSecretReference. :type connection_string: object + :param account_key: The Azure key vault secret reference of accountKey in + connection string. + :type account_key: + ~azure.mgmt.datafactory.models.AzureKeyVaultSecretReference :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -55,11 +59,13 @@ class CosmosDbLinkedService(LinkedService): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'type': {'key': 'type', 'type': 'str'}, 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, + 'account_key': {'key': 'typeProperties.accountKey', 'type': 'AzureKeyVaultSecretReference'}, 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } def __init__(self, **kwargs): super(CosmosDbLinkedService, self).__init__(**kwargs) self.connection_string = kwargs.get('connection_string', None) + self.account_key = kwargs.get('account_key', None) self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'CosmosDb' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cosmos_db_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cosmos_db_linked_service_py3.py index a338f3dc093b..3b951a68a65a 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cosmos_db_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cosmos_db_linked_service_py3.py @@ -36,6 +36,10 @@ class CosmosDbLinkedService(LinkedService): :param connection_string: Required. The connection string. Type: string, SecureString or AzureKeyVaultSecretReference. :type connection_string: object + :param account_key: The Azure key vault secret reference of accountKey in + connection string. + :type account_key: + ~azure.mgmt.datafactory.models.AzureKeyVaultSecretReference :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -55,11 +59,13 @@ class CosmosDbLinkedService(LinkedService): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'type': {'key': 'type', 'type': 'str'}, 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, + 'account_key': {'key': 'typeProperties.accountKey', 'type': 'AzureKeyVaultSecretReference'}, 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, *, connection_string, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, encrypted_credential=None, **kwargs) -> None: + def __init__(self, *, connection_string, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, account_key=None, encrypted_credential=None, **kwargs) -> None: super(CosmosDbLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) self.connection_string = connection_string + self.account_key = account_key self.encrypted_credential = encrypted_credential self.type = 'CosmosDb' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_linked_service.py index 7c8aceb401dd..f5c02a071718 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_linked_service.py @@ -36,6 +36,10 @@ class CouchbaseLinkedService(LinkedService): :param connection_string: An ODBC connection string. Type: string, SecureString or AzureKeyVaultSecretReference. :type connection_string: object + :param cred_string: The Azure key vault secret reference of credString in + connection string. + :type cred_string: + ~azure.mgmt.datafactory.models.AzureKeyVaultSecretReference :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -54,11 +58,13 @@ class CouchbaseLinkedService(LinkedService): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'type': {'key': 'type', 'type': 'str'}, 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, + 'cred_string': {'key': 'typeProperties.credString', 'type': 'AzureKeyVaultSecretReference'}, 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } def __init__(self, **kwargs): super(CouchbaseLinkedService, self).__init__(**kwargs) self.connection_string = kwargs.get('connection_string', None) + self.cred_string = kwargs.get('cred_string', None) self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'Couchbase' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_linked_service_py3.py index 7f94dca390eb..1507d6ab7b32 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_linked_service_py3.py @@ -36,6 +36,10 @@ class CouchbaseLinkedService(LinkedService): :param connection_string: An ODBC connection string. Type: string, SecureString or AzureKeyVaultSecretReference. :type connection_string: object + :param cred_string: The Azure key vault secret reference of credString in + connection string. + :type cred_string: + ~azure.mgmt.datafactory.models.AzureKeyVaultSecretReference :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -54,11 +58,13 @@ class CouchbaseLinkedService(LinkedService): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'type': {'key': 'type', 'type': 'str'}, 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, + 'cred_string': {'key': 'typeProperties.credString', 'type': 'AzureKeyVaultSecretReference'}, 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, *, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, connection_string=None, encrypted_credential=None, **kwargs) -> None: + def __init__(self, *, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, connection_string=None, cred_string=None, encrypted_credential=None, **kwargs) -> None: super(CouchbaseLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) self.connection_string = connection_string + self.cred_string = cred_string self.encrypted_credential = encrypted_credential self.type = 'Couchbase' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_linked_service.py index 0ece6ef6ad05..52ad5888f5f2 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_linked_service.py @@ -36,6 +36,9 @@ class DrillLinkedService(LinkedService): :param connection_string: An ODBC connection string. Type: string, SecureString or AzureKeyVaultSecretReference. :type connection_string: object + :param pwd: The Azure key vault secret reference of password in connection + string. + :type pwd: ~azure.mgmt.datafactory.models.AzureKeyVaultSecretReference :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -54,11 +57,13 @@ class DrillLinkedService(LinkedService): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'type': {'key': 'type', 'type': 'str'}, 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, + 'pwd': {'key': 'typeProperties.pwd', 'type': 'AzureKeyVaultSecretReference'}, 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } def __init__(self, **kwargs): super(DrillLinkedService, self).__init__(**kwargs) self.connection_string = kwargs.get('connection_string', None) + self.pwd = kwargs.get('pwd', None) self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'Drill' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_linked_service_py3.py index 7e6a7a64bde2..b556d7e92be3 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_linked_service_py3.py @@ -36,6 +36,9 @@ class DrillLinkedService(LinkedService): :param connection_string: An ODBC connection string. Type: string, SecureString or AzureKeyVaultSecretReference. :type connection_string: object + :param pwd: The Azure key vault secret reference of password in connection + string. + :type pwd: ~azure.mgmt.datafactory.models.AzureKeyVaultSecretReference :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -54,11 +57,13 @@ class DrillLinkedService(LinkedService): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'type': {'key': 'type', 'type': 'str'}, 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, + 'pwd': {'key': 'typeProperties.pwd', 'type': 'AzureKeyVaultSecretReference'}, 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, *, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, connection_string=None, encrypted_credential=None, **kwargs) -> None: + def __init__(self, *, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, connection_string=None, pwd=None, encrypted_credential=None, **kwargs) -> None: super(DrillLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) self.connection_string = connection_string + self.pwd = pwd self.encrypted_credential = encrypted_credential self.type = 'Drill' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_linked_service.py index d7507545f9e5..d3de7ccab502 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_linked_service.py @@ -36,6 +36,9 @@ class GreenplumLinkedService(LinkedService): :param connection_string: An ODBC connection string. Type: string, SecureString or AzureKeyVaultSecretReference. :type connection_string: object + :param pwd: The Azure key vault secret reference of password in connection + string. + :type pwd: ~azure.mgmt.datafactory.models.AzureKeyVaultSecretReference :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -54,11 +57,13 @@ class GreenplumLinkedService(LinkedService): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'type': {'key': 'type', 'type': 'str'}, 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, + 'pwd': {'key': 'typeProperties.pwd', 'type': 'AzureKeyVaultSecretReference'}, 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } def __init__(self, **kwargs): super(GreenplumLinkedService, self).__init__(**kwargs) self.connection_string = kwargs.get('connection_string', None) + self.pwd = kwargs.get('pwd', None) self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'Greenplum' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_linked_service_py3.py index ea15aa2690dd..886d38718ecd 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_linked_service_py3.py @@ -36,6 +36,9 @@ class GreenplumLinkedService(LinkedService): :param connection_string: An ODBC connection string. Type: string, SecureString or AzureKeyVaultSecretReference. :type connection_string: object + :param pwd: The Azure key vault secret reference of password in connection + string. + :type pwd: ~azure.mgmt.datafactory.models.AzureKeyVaultSecretReference :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -54,11 +57,13 @@ class GreenplumLinkedService(LinkedService): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'type': {'key': 'type', 'type': 'str'}, 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, + 'pwd': {'key': 'typeProperties.pwd', 'type': 'AzureKeyVaultSecretReference'}, 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, *, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, connection_string=None, encrypted_credential=None, **kwargs) -> None: + def __init__(self, *, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, connection_string=None, pwd=None, encrypted_credential=None, **kwargs) -> None: super(GreenplumLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) self.connection_string = connection_string + self.pwd = pwd self.encrypted_credential = encrypted_credential self.type = 'Greenplum' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_linked_service.py index 853df4dbd376..0a98a04138dc 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_linked_service.py @@ -36,6 +36,9 @@ class MariaDBLinkedService(LinkedService): :param connection_string: An ODBC connection string. Type: string, SecureString or AzureKeyVaultSecretReference. :type connection_string: object + :param pwd: The Azure key vault secret reference of password in connection + string. + :type pwd: ~azure.mgmt.datafactory.models.AzureKeyVaultSecretReference :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -54,11 +57,13 @@ class MariaDBLinkedService(LinkedService): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'type': {'key': 'type', 'type': 'str'}, 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, + 'pwd': {'key': 'typeProperties.pwd', 'type': 'AzureKeyVaultSecretReference'}, 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } def __init__(self, **kwargs): super(MariaDBLinkedService, self).__init__(**kwargs) self.connection_string = kwargs.get('connection_string', None) + self.pwd = kwargs.get('pwd', None) self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'MariaDB' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_linked_service_py3.py index b0bbd9c4f733..ef1114660ad7 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_linked_service_py3.py @@ -36,6 +36,9 @@ class MariaDBLinkedService(LinkedService): :param connection_string: An ODBC connection string. Type: string, SecureString or AzureKeyVaultSecretReference. :type connection_string: object + :param pwd: The Azure key vault secret reference of password in connection + string. + :type pwd: ~azure.mgmt.datafactory.models.AzureKeyVaultSecretReference :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -54,11 +57,13 @@ class MariaDBLinkedService(LinkedService): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'type': {'key': 'type', 'type': 'str'}, 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, + 'pwd': {'key': 'typeProperties.pwd', 'type': 'AzureKeyVaultSecretReference'}, 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, *, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, connection_string=None, encrypted_credential=None, **kwargs) -> None: + def __init__(self, *, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, connection_string=None, pwd=None, encrypted_credential=None, **kwargs) -> None: super(MariaDBLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) self.connection_string = connection_string + self.pwd = pwd self.encrypted_credential = encrypted_credential self.type = 'MariaDB' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/my_sql_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/my_sql_linked_service.py index b643e1c7266b..542fb13b7a37 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/my_sql_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/my_sql_linked_service.py @@ -35,6 +35,10 @@ class MySqlLinkedService(LinkedService): :type type: str :param connection_string: Required. The connection string. :type connection_string: ~azure.mgmt.datafactory.models.SecretBase + :param password: The Azure key vault secret reference of password in + connection string. + :type password: + ~azure.mgmt.datafactory.models.AzureKeyVaultSecretReference :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -54,11 +58,13 @@ class MySqlLinkedService(LinkedService): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'type': {'key': 'type', 'type': 'str'}, 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'SecretBase'}, + 'password': {'key': 'typeProperties.password', 'type': 'AzureKeyVaultSecretReference'}, 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } def __init__(self, **kwargs): super(MySqlLinkedService, self).__init__(**kwargs) self.connection_string = kwargs.get('connection_string', None) + self.password = kwargs.get('password', None) self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'MySql' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/my_sql_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/my_sql_linked_service_py3.py index 38cf9b34198a..cd87d5e7e3b5 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/my_sql_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/my_sql_linked_service_py3.py @@ -35,6 +35,10 @@ class MySqlLinkedService(LinkedService): :type type: str :param connection_string: Required. The connection string. :type connection_string: ~azure.mgmt.datafactory.models.SecretBase + :param password: The Azure key vault secret reference of password in + connection string. + :type password: + ~azure.mgmt.datafactory.models.AzureKeyVaultSecretReference :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -54,11 +58,13 @@ class MySqlLinkedService(LinkedService): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'type': {'key': 'type', 'type': 'str'}, 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'SecretBase'}, + 'password': {'key': 'typeProperties.password', 'type': 'AzureKeyVaultSecretReference'}, 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, *, connection_string, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, encrypted_credential=None, **kwargs) -> None: + def __init__(self, *, connection_string, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, password=None, encrypted_credential=None, **kwargs) -> None: super(MySqlLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) self.connection_string = connection_string + self.password = password self.encrypted_credential = encrypted_credential self.type = 'MySql' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/netezza_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/netezza_linked_service.py index 1d7b72403b54..319a68efddc5 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/netezza_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/netezza_linked_service.py @@ -36,6 +36,9 @@ class NetezzaLinkedService(LinkedService): :param connection_string: An ODBC connection string. Type: string, SecureString or AzureKeyVaultSecretReference. :type connection_string: object + :param pwd: The Azure key vault secret reference of password in connection + string. + :type pwd: ~azure.mgmt.datafactory.models.AzureKeyVaultSecretReference :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -54,11 +57,13 @@ class NetezzaLinkedService(LinkedService): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'type': {'key': 'type', 'type': 'str'}, 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, + 'pwd': {'key': 'typeProperties.pwd', 'type': 'AzureKeyVaultSecretReference'}, 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } def __init__(self, **kwargs): super(NetezzaLinkedService, self).__init__(**kwargs) self.connection_string = kwargs.get('connection_string', None) + self.pwd = kwargs.get('pwd', None) self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'Netezza' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/netezza_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/netezza_linked_service_py3.py index b6576954e26b..6c3b607d60dc 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/netezza_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/netezza_linked_service_py3.py @@ -36,6 +36,9 @@ class NetezzaLinkedService(LinkedService): :param connection_string: An ODBC connection string. Type: string, SecureString or AzureKeyVaultSecretReference. :type connection_string: object + :param pwd: The Azure key vault secret reference of password in connection + string. + :type pwd: ~azure.mgmt.datafactory.models.AzureKeyVaultSecretReference :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -54,11 +57,13 @@ class NetezzaLinkedService(LinkedService): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'type': {'key': 'type', 'type': 'str'}, 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, + 'pwd': {'key': 'typeProperties.pwd', 'type': 'AzureKeyVaultSecretReference'}, 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, *, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, connection_string=None, encrypted_credential=None, **kwargs) -> None: + def __init__(self, *, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, connection_string=None, pwd=None, encrypted_credential=None, **kwargs) -> None: super(NetezzaLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) self.connection_string = connection_string + self.pwd = pwd self.encrypted_credential = encrypted_credential self.type = 'Netezza' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_linked_service.py index 2aca7d0a7215..5485151adb1f 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_linked_service.py @@ -36,6 +36,10 @@ class OracleLinkedService(LinkedService): :param connection_string: Required. The connection string. Type: string, SecureString or AzureKeyVaultSecretReference. :type connection_string: object + :param password: The Azure key vault secret reference of password in + connection string. + :type password: + ~azure.mgmt.datafactory.models.AzureKeyVaultSecretReference :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -55,11 +59,13 @@ class OracleLinkedService(LinkedService): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'type': {'key': 'type', 'type': 'str'}, 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, + 'password': {'key': 'typeProperties.password', 'type': 'AzureKeyVaultSecretReference'}, 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } def __init__(self, **kwargs): super(OracleLinkedService, self).__init__(**kwargs) self.connection_string = kwargs.get('connection_string', None) + self.password = kwargs.get('password', None) self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'Oracle' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_linked_service_py3.py index bc4765e0b2e2..80b0ed1176ff 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_linked_service_py3.py @@ -36,6 +36,10 @@ class OracleLinkedService(LinkedService): :param connection_string: Required. The connection string. Type: string, SecureString or AzureKeyVaultSecretReference. :type connection_string: object + :param password: The Azure key vault secret reference of password in + connection string. + :type password: + ~azure.mgmt.datafactory.models.AzureKeyVaultSecretReference :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -55,11 +59,13 @@ class OracleLinkedService(LinkedService): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'type': {'key': 'type', 'type': 'str'}, 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, + 'password': {'key': 'typeProperties.password', 'type': 'AzureKeyVaultSecretReference'}, 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, *, connection_string, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, encrypted_credential=None, **kwargs) -> None: + def __init__(self, *, connection_string, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, password=None, encrypted_credential=None, **kwargs) -> None: super(OracleLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) self.connection_string = connection_string + self.password = password self.encrypted_credential = encrypted_credential self.type = 'Oracle' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/postgre_sql_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/postgre_sql_linked_service.py index 6a572c24c1cd..af16c6c89cd2 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/postgre_sql_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/postgre_sql_linked_service.py @@ -35,6 +35,10 @@ class PostgreSqlLinkedService(LinkedService): :type type: str :param connection_string: Required. The connection string. :type connection_string: ~azure.mgmt.datafactory.models.SecretBase + :param password: The Azure key vault secret reference of password in + connection string. + :type password: + ~azure.mgmt.datafactory.models.AzureKeyVaultSecretReference :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -54,11 +58,13 @@ class PostgreSqlLinkedService(LinkedService): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'type': {'key': 'type', 'type': 'str'}, 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'SecretBase'}, + 'password': {'key': 'typeProperties.password', 'type': 'AzureKeyVaultSecretReference'}, 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } def __init__(self, **kwargs): super(PostgreSqlLinkedService, self).__init__(**kwargs) self.connection_string = kwargs.get('connection_string', None) + self.password = kwargs.get('password', None) self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'PostgreSql' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/postgre_sql_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/postgre_sql_linked_service_py3.py index ae16efbbfdad..5e7e674a2447 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/postgre_sql_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/postgre_sql_linked_service_py3.py @@ -35,6 +35,10 @@ class PostgreSqlLinkedService(LinkedService): :type type: str :param connection_string: Required. The connection string. :type connection_string: ~azure.mgmt.datafactory.models.SecretBase + :param password: The Azure key vault secret reference of password in + connection string. + :type password: + ~azure.mgmt.datafactory.models.AzureKeyVaultSecretReference :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -54,11 +58,13 @@ class PostgreSqlLinkedService(LinkedService): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'type': {'key': 'type', 'type': 'str'}, 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'SecretBase'}, + 'password': {'key': 'typeProperties.password', 'type': 'AzureKeyVaultSecretReference'}, 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, *, connection_string, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, encrypted_credential=None, **kwargs) -> None: + def __init__(self, *, connection_string, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, password=None, encrypted_credential=None, **kwargs) -> None: super(PostgreSqlLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) self.connection_string = connection_string + self.password = password self.encrypted_credential = encrypted_credential self.type = 'PostgreSql' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/vertica_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/vertica_linked_service.py index d872a0d547cc..fafba164a752 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/vertica_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/vertica_linked_service.py @@ -36,6 +36,9 @@ class VerticaLinkedService(LinkedService): :param connection_string: An ODBC connection string. Type: string, SecureString or AzureKeyVaultSecretReference. :type connection_string: object + :param pwd: The Azure key vault secret reference of password in connection + string. + :type pwd: ~azure.mgmt.datafactory.models.AzureKeyVaultSecretReference :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -54,11 +57,13 @@ class VerticaLinkedService(LinkedService): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'type': {'key': 'type', 'type': 'str'}, 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, + 'pwd': {'key': 'typeProperties.pwd', 'type': 'AzureKeyVaultSecretReference'}, 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } def __init__(self, **kwargs): super(VerticaLinkedService, self).__init__(**kwargs) self.connection_string = kwargs.get('connection_string', None) + self.pwd = kwargs.get('pwd', None) self.encrypted_credential = kwargs.get('encrypted_credential', None) self.type = 'Vertica' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/vertica_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/vertica_linked_service_py3.py index 1f8d8d570205..77caf915eaab 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/vertica_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/vertica_linked_service_py3.py @@ -36,6 +36,9 @@ class VerticaLinkedService(LinkedService): :param connection_string: An ODBC connection string. Type: string, SecureString or AzureKeyVaultSecretReference. :type connection_string: object + :param pwd: The Azure key vault secret reference of password in connection + string. + :type pwd: ~azure.mgmt.datafactory.models.AzureKeyVaultSecretReference :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -54,11 +57,13 @@ class VerticaLinkedService(LinkedService): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'type': {'key': 'type', 'type': 'str'}, 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, + 'pwd': {'key': 'typeProperties.pwd', 'type': 'AzureKeyVaultSecretReference'}, 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, *, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, connection_string=None, encrypted_credential=None, **kwargs) -> None: + def __init__(self, *, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, connection_string=None, pwd=None, encrypted_credential=None, **kwargs) -> None: super(VerticaLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) self.connection_string = connection_string + self.pwd = pwd self.encrypted_credential = encrypted_credential self.type = 'Vertica' From f5fe8454bd11d2789b804dffd569d0a40160902f Mon Sep 17 00:00:00 2001 From: Azure SDK for Python bot Date: Tue, 13 Nov 2018 22:06:45 -0800 Subject: [PATCH 30/54] Generated from 136aff8dae7f767f7391d0b20a2c071aef29f77f (#3759) DataFactory: add table name in dataset --- .../datafactory/models/amazon_mws_object_dataset.py | 5 +++++ .../models/amazon_mws_object_dataset_py3.py | 7 ++++++- .../models/azure_postgre_sql_table_dataset.py | 5 +++++ .../models/azure_postgre_sql_table_dataset_py3.py | 7 ++++++- .../mgmt/datafactory/models/concur_object_dataset.py | 5 +++++ .../datafactory/models/concur_object_dataset_py3.py | 7 ++++++- .../mgmt/datafactory/models/couchbase_table_dataset.py | 5 +++++ .../datafactory/models/couchbase_table_dataset_py3.py | 7 ++++++- .../mgmt/datafactory/models/drill_table_dataset.py | 5 +++++ .../mgmt/datafactory/models/drill_table_dataset_py3.py | 7 ++++++- .../mgmt/datafactory/models/eloqua_object_dataset.py | 5 +++++ .../datafactory/models/eloqua_object_dataset_py3.py | 7 ++++++- .../models/google_big_query_object_dataset.py | 5 +++++ .../models/google_big_query_object_dataset_py3.py | 7 ++++++- .../mgmt/datafactory/models/greenplum_table_dataset.py | 5 +++++ .../datafactory/models/greenplum_table_dataset_py3.py | 7 ++++++- .../mgmt/datafactory/models/hbase_object_dataset.py | 5 +++++ .../datafactory/models/hbase_object_dataset_py3.py | 7 ++++++- .../mgmt/datafactory/models/hive_object_dataset.py | 5 +++++ .../mgmt/datafactory/models/hive_object_dataset_py3.py | 7 ++++++- .../mgmt/datafactory/models/hubspot_object_dataset.py | 5 +++++ .../datafactory/models/hubspot_object_dataset_py3.py | 7 ++++++- .../mgmt/datafactory/models/impala_object_dataset.py | 5 +++++ .../datafactory/models/impala_object_dataset_py3.py | 7 ++++++- .../mgmt/datafactory/models/jira_object_dataset.py | 5 +++++ .../mgmt/datafactory/models/jira_object_dataset_py3.py | 7 ++++++- .../mgmt/datafactory/models/magento_object_dataset.py | 5 +++++ .../datafactory/models/magento_object_dataset_py3.py | 7 ++++++- .../mgmt/datafactory/models/maria_db_table_dataset.py | 5 +++++ .../datafactory/models/maria_db_table_dataset_py3.py | 7 ++++++- .../mgmt/datafactory/models/marketo_object_dataset.py | 5 +++++ .../datafactory/models/marketo_object_dataset_py3.py | 7 ++++++- .../mgmt/datafactory/models/netezza_table_dataset.py | 5 +++++ .../datafactory/models/netezza_table_dataset_py3.py | 7 ++++++- .../mgmt/datafactory/models/paypal_object_dataset.py | 5 +++++ .../datafactory/models/paypal_object_dataset_py3.py | 7 ++++++- .../mgmt/datafactory/models/phoenix_object_dataset.py | 5 +++++ .../datafactory/models/phoenix_object_dataset_py3.py | 7 ++++++- .../mgmt/datafactory/models/presto_object_dataset.py | 5 +++++ .../datafactory/models/presto_object_dataset_py3.py | 7 ++++++- .../datafactory/models/quick_books_object_dataset.py | 5 +++++ .../models/quick_books_object_dataset_py3.py | 7 ++++++- .../datafactory/models/responsys_object_dataset.py | 5 +++++ .../datafactory/models/responsys_object_dataset_py3.py | 7 ++++++- .../salesforce_marketing_cloud_object_dataset.py | 5 +++++ .../salesforce_marketing_cloud_object_dataset_py3.py | 7 ++++++- .../datafactory/models/service_now_object_dataset.py | 5 +++++ .../models/service_now_object_dataset_py3.py | 7 ++++++- .../mgmt/datafactory/models/shopify_object_dataset.py | 5 +++++ .../datafactory/models/shopify_object_dataset_py3.py | 7 ++++++- .../mgmt/datafactory/models/spark_object_dataset.py | 5 +++++ .../datafactory/models/spark_object_dataset_py3.py | 7 ++++++- .../mgmt/datafactory/models/square_object_dataset.py | 5 +++++ .../datafactory/models/square_object_dataset_py3.py | 7 ++++++- .../datafactory/models/stored_procedure_parameter.py | 8 +------- .../models/stored_procedure_parameter_py3.py | 10 ++-------- .../mgmt/datafactory/models/vertica_table_dataset.py | 5 +++++ .../datafactory/models/vertica_table_dataset_py3.py | 7 ++++++- .../mgmt/datafactory/models/xero_object_dataset.py | 5 +++++ .../mgmt/datafactory/models/xero_object_dataset_py3.py | 7 ++++++- .../mgmt/datafactory/models/zoho_object_dataset.py | 5 +++++ .../mgmt/datafactory/models/zoho_object_dataset_py3.py | 7 ++++++- 62 files changed, 333 insertions(+), 45 deletions(-) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_object_dataset.py index 07eaa902b276..4b6782cee208 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_object_dataset.py @@ -39,6 +39,9 @@ class AmazonMWSObjectDataset(Dataset): :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str + :param table_name: The table name. Type: string (or Expression with + resultType string). + :type table_name: object """ _validation = { @@ -55,8 +58,10 @@ class AmazonMWSObjectDataset(Dataset): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, + 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } def __init__(self, **kwargs): super(AmazonMWSObjectDataset, self).__init__(**kwargs) + self.table_name = kwargs.get('table_name', None) self.type = 'AmazonMWSObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_object_dataset_py3.py index 3180f8952185..abe8569df119 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_object_dataset_py3.py @@ -39,6 +39,9 @@ class AmazonMWSObjectDataset(Dataset): :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str + :param table_name: The table name. Type: string (or Expression with + resultType string). + :type table_name: object """ _validation = { @@ -55,8 +58,10 @@ class AmazonMWSObjectDataset(Dataset): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, + 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: super(AmazonMWSObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + self.table_name = table_name self.type = 'AmazonMWSObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_table_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_table_dataset.py index 33e74887c012..02ce3e1e20f9 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_table_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_table_dataset.py @@ -39,6 +39,9 @@ class AzurePostgreSqlTableDataset(Dataset): :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str + :param table_name: The table name. Type: string (or Expression with + resultType string). + :type table_name: object """ _validation = { @@ -55,8 +58,10 @@ class AzurePostgreSqlTableDataset(Dataset): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, + 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } def __init__(self, **kwargs): super(AzurePostgreSqlTableDataset, self).__init__(**kwargs) + self.table_name = kwargs.get('table_name', None) self.type = 'AzurePostgreSqlTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_table_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_table_dataset_py3.py index 0a6d551868f7..88b251d4cb59 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_table_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_table_dataset_py3.py @@ -39,6 +39,9 @@ class AzurePostgreSqlTableDataset(Dataset): :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str + :param table_name: The table name. Type: string (or Expression with + resultType string). + :type table_name: object """ _validation = { @@ -55,8 +58,10 @@ class AzurePostgreSqlTableDataset(Dataset): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, + 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: super(AzurePostgreSqlTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + self.table_name = table_name self.type = 'AzurePostgreSqlTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_object_dataset.py index f0bf7c705098..2f3829b16b20 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_object_dataset.py @@ -39,6 +39,9 @@ class ConcurObjectDataset(Dataset): :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str + :param table_name: The table name. Type: string (or Expression with + resultType string). + :type table_name: object """ _validation = { @@ -55,8 +58,10 @@ class ConcurObjectDataset(Dataset): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, + 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } def __init__(self, **kwargs): super(ConcurObjectDataset, self).__init__(**kwargs) + self.table_name = kwargs.get('table_name', None) self.type = 'ConcurObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_object_dataset_py3.py index 6ca127fab236..0b7016c75604 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_object_dataset_py3.py @@ -39,6 +39,9 @@ class ConcurObjectDataset(Dataset): :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str + :param table_name: The table name. Type: string (or Expression with + resultType string). + :type table_name: object """ _validation = { @@ -55,8 +58,10 @@ class ConcurObjectDataset(Dataset): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, + 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: super(ConcurObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + self.table_name = table_name self.type = 'ConcurObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_table_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_table_dataset.py index 88053e34fe9d..cb87312e1377 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_table_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_table_dataset.py @@ -39,6 +39,9 @@ class CouchbaseTableDataset(Dataset): :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str + :param table_name: The table name. Type: string (or Expression with + resultType string). + :type table_name: object """ _validation = { @@ -55,8 +58,10 @@ class CouchbaseTableDataset(Dataset): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, + 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } def __init__(self, **kwargs): super(CouchbaseTableDataset, self).__init__(**kwargs) + self.table_name = kwargs.get('table_name', None) self.type = 'CouchbaseTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_table_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_table_dataset_py3.py index 00bf158a6f84..1fa41220cfc1 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_table_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_table_dataset_py3.py @@ -39,6 +39,9 @@ class CouchbaseTableDataset(Dataset): :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str + :param table_name: The table name. Type: string (or Expression with + resultType string). + :type table_name: object """ _validation = { @@ -55,8 +58,10 @@ class CouchbaseTableDataset(Dataset): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, + 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: super(CouchbaseTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + self.table_name = table_name self.type = 'CouchbaseTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_table_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_table_dataset.py index 59802dcfcd77..d392331627ad 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_table_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_table_dataset.py @@ -39,6 +39,9 @@ class DrillTableDataset(Dataset): :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str + :param table_name: The table name. Type: string (or Expression with + resultType string). + :type table_name: object """ _validation = { @@ -55,8 +58,10 @@ class DrillTableDataset(Dataset): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, + 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } def __init__(self, **kwargs): super(DrillTableDataset, self).__init__(**kwargs) + self.table_name = kwargs.get('table_name', None) self.type = 'DrillTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_table_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_table_dataset_py3.py index be4da0ea660d..bec695cdeaee 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_table_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_table_dataset_py3.py @@ -39,6 +39,9 @@ class DrillTableDataset(Dataset): :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str + :param table_name: The table name. Type: string (or Expression with + resultType string). + :type table_name: object """ _validation = { @@ -55,8 +58,10 @@ class DrillTableDataset(Dataset): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, + 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: super(DrillTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + self.table_name = table_name self.type = 'DrillTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/eloqua_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/eloqua_object_dataset.py index 5ff8abf3bf32..d077cd28c5b4 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/eloqua_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/eloqua_object_dataset.py @@ -39,6 +39,9 @@ class EloquaObjectDataset(Dataset): :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str + :param table_name: The table name. Type: string (or Expression with + resultType string). + :type table_name: object """ _validation = { @@ -55,8 +58,10 @@ class EloquaObjectDataset(Dataset): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, + 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } def __init__(self, **kwargs): super(EloquaObjectDataset, self).__init__(**kwargs) + self.table_name = kwargs.get('table_name', None) self.type = 'EloquaObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/eloqua_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/eloqua_object_dataset_py3.py index bd4e833d1eea..1220f816cbf4 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/eloqua_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/eloqua_object_dataset_py3.py @@ -39,6 +39,9 @@ class EloquaObjectDataset(Dataset): :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str + :param table_name: The table name. Type: string (or Expression with + resultType string). + :type table_name: object """ _validation = { @@ -55,8 +58,10 @@ class EloquaObjectDataset(Dataset): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, + 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: super(EloquaObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + self.table_name = table_name self.type = 'EloquaObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_big_query_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_big_query_object_dataset.py index 2a739203a74b..6404a7ae5110 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_big_query_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_big_query_object_dataset.py @@ -39,6 +39,9 @@ class GoogleBigQueryObjectDataset(Dataset): :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str + :param table_name: The table name. Type: string (or Expression with + resultType string). + :type table_name: object """ _validation = { @@ -55,8 +58,10 @@ class GoogleBigQueryObjectDataset(Dataset): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, + 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } def __init__(self, **kwargs): super(GoogleBigQueryObjectDataset, self).__init__(**kwargs) + self.table_name = kwargs.get('table_name', None) self.type = 'GoogleBigQueryObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_big_query_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_big_query_object_dataset_py3.py index 57423a6bc421..30ca4d4efe46 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_big_query_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_big_query_object_dataset_py3.py @@ -39,6 +39,9 @@ class GoogleBigQueryObjectDataset(Dataset): :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str + :param table_name: The table name. Type: string (or Expression with + resultType string). + :type table_name: object """ _validation = { @@ -55,8 +58,10 @@ class GoogleBigQueryObjectDataset(Dataset): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, + 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: super(GoogleBigQueryObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + self.table_name = table_name self.type = 'GoogleBigQueryObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_table_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_table_dataset.py index 7a5bbdb32063..338042ec56d2 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_table_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_table_dataset.py @@ -39,6 +39,9 @@ class GreenplumTableDataset(Dataset): :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str + :param table_name: The table name. Type: string (or Expression with + resultType string). + :type table_name: object """ _validation = { @@ -55,8 +58,10 @@ class GreenplumTableDataset(Dataset): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, + 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } def __init__(self, **kwargs): super(GreenplumTableDataset, self).__init__(**kwargs) + self.table_name = kwargs.get('table_name', None) self.type = 'GreenplumTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_table_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_table_dataset_py3.py index a693ddcd4e41..1b189fbfdb4b 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_table_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_table_dataset_py3.py @@ -39,6 +39,9 @@ class GreenplumTableDataset(Dataset): :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str + :param table_name: The table name. Type: string (or Expression with + resultType string). + :type table_name: object """ _validation = { @@ -55,8 +58,10 @@ class GreenplumTableDataset(Dataset): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, + 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: super(GreenplumTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + self.table_name = table_name self.type = 'GreenplumTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hbase_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hbase_object_dataset.py index 8500e647d220..f0f7b5735861 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hbase_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hbase_object_dataset.py @@ -39,6 +39,9 @@ class HBaseObjectDataset(Dataset): :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str + :param table_name: The table name. Type: string (or Expression with + resultType string). + :type table_name: object """ _validation = { @@ -55,8 +58,10 @@ class HBaseObjectDataset(Dataset): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, + 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } def __init__(self, **kwargs): super(HBaseObjectDataset, self).__init__(**kwargs) + self.table_name = kwargs.get('table_name', None) self.type = 'HBaseObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hbase_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hbase_object_dataset_py3.py index d52489d391c0..d16c7bfba5d1 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hbase_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hbase_object_dataset_py3.py @@ -39,6 +39,9 @@ class HBaseObjectDataset(Dataset): :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str + :param table_name: The table name. Type: string (or Expression with + resultType string). + :type table_name: object """ _validation = { @@ -55,8 +58,10 @@ class HBaseObjectDataset(Dataset): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, + 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: super(HBaseObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + self.table_name = table_name self.type = 'HBaseObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hive_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hive_object_dataset.py index 848031ff4115..064772f3e32f 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hive_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hive_object_dataset.py @@ -39,6 +39,9 @@ class HiveObjectDataset(Dataset): :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str + :param table_name: The table name. Type: string (or Expression with + resultType string). + :type table_name: object """ _validation = { @@ -55,8 +58,10 @@ class HiveObjectDataset(Dataset): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, + 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } def __init__(self, **kwargs): super(HiveObjectDataset, self).__init__(**kwargs) + self.table_name = kwargs.get('table_name', None) self.type = 'HiveObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hive_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hive_object_dataset_py3.py index 944cab17444e..b36690c692a8 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hive_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hive_object_dataset_py3.py @@ -39,6 +39,9 @@ class HiveObjectDataset(Dataset): :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str + :param table_name: The table name. Type: string (or Expression with + resultType string). + :type table_name: object """ _validation = { @@ -55,8 +58,10 @@ class HiveObjectDataset(Dataset): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, + 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: super(HiveObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + self.table_name = table_name self.type = 'HiveObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_object_dataset.py index d3a53c067ede..a48c4ea67a92 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_object_dataset.py @@ -39,6 +39,9 @@ class HubspotObjectDataset(Dataset): :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str + :param table_name: The table name. Type: string (or Expression with + resultType string). + :type table_name: object """ _validation = { @@ -55,8 +58,10 @@ class HubspotObjectDataset(Dataset): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, + 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } def __init__(self, **kwargs): super(HubspotObjectDataset, self).__init__(**kwargs) + self.table_name = kwargs.get('table_name', None) self.type = 'HubspotObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_object_dataset_py3.py index c5efaf9cad56..996b626ed551 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_object_dataset_py3.py @@ -39,6 +39,9 @@ class HubspotObjectDataset(Dataset): :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str + :param table_name: The table name. Type: string (or Expression with + resultType string). + :type table_name: object """ _validation = { @@ -55,8 +58,10 @@ class HubspotObjectDataset(Dataset): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, + 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: super(HubspotObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + self.table_name = table_name self.type = 'HubspotObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/impala_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/impala_object_dataset.py index 62e709ebebc3..13a1a548fd59 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/impala_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/impala_object_dataset.py @@ -39,6 +39,9 @@ class ImpalaObjectDataset(Dataset): :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str + :param table_name: The table name. Type: string (or Expression with + resultType string). + :type table_name: object """ _validation = { @@ -55,8 +58,10 @@ class ImpalaObjectDataset(Dataset): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, + 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } def __init__(self, **kwargs): super(ImpalaObjectDataset, self).__init__(**kwargs) + self.table_name = kwargs.get('table_name', None) self.type = 'ImpalaObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/impala_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/impala_object_dataset_py3.py index a17b4a08fb26..c0dc4d490e78 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/impala_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/impala_object_dataset_py3.py @@ -39,6 +39,9 @@ class ImpalaObjectDataset(Dataset): :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str + :param table_name: The table name. Type: string (or Expression with + resultType string). + :type table_name: object """ _validation = { @@ -55,8 +58,10 @@ class ImpalaObjectDataset(Dataset): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, + 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: super(ImpalaObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + self.table_name = table_name self.type = 'ImpalaObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_object_dataset.py index 464a640719c4..614c2fe6a5fb 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_object_dataset.py @@ -39,6 +39,9 @@ class JiraObjectDataset(Dataset): :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str + :param table_name: The table name. Type: string (or Expression with + resultType string). + :type table_name: object """ _validation = { @@ -55,8 +58,10 @@ class JiraObjectDataset(Dataset): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, + 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } def __init__(self, **kwargs): super(JiraObjectDataset, self).__init__(**kwargs) + self.table_name = kwargs.get('table_name', None) self.type = 'JiraObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_object_dataset_py3.py index 955eabebb92c..9a2a445743da 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_object_dataset_py3.py @@ -39,6 +39,9 @@ class JiraObjectDataset(Dataset): :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str + :param table_name: The table name. Type: string (or Expression with + resultType string). + :type table_name: object """ _validation = { @@ -55,8 +58,10 @@ class JiraObjectDataset(Dataset): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, + 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: super(JiraObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + self.table_name = table_name self.type = 'JiraObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/magento_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/magento_object_dataset.py index 913ccbb68326..d0c83370bd47 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/magento_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/magento_object_dataset.py @@ -39,6 +39,9 @@ class MagentoObjectDataset(Dataset): :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str + :param table_name: The table name. Type: string (or Expression with + resultType string). + :type table_name: object """ _validation = { @@ -55,8 +58,10 @@ class MagentoObjectDataset(Dataset): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, + 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } def __init__(self, **kwargs): super(MagentoObjectDataset, self).__init__(**kwargs) + self.table_name = kwargs.get('table_name', None) self.type = 'MagentoObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/magento_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/magento_object_dataset_py3.py index 08639eae595c..8d95ca580679 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/magento_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/magento_object_dataset_py3.py @@ -39,6 +39,9 @@ class MagentoObjectDataset(Dataset): :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str + :param table_name: The table name. Type: string (or Expression with + resultType string). + :type table_name: object """ _validation = { @@ -55,8 +58,10 @@ class MagentoObjectDataset(Dataset): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, + 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: super(MagentoObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + self.table_name = table_name self.type = 'MagentoObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_table_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_table_dataset.py index 5ef7db26acdd..58a2b3f4da0e 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_table_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_table_dataset.py @@ -39,6 +39,9 @@ class MariaDBTableDataset(Dataset): :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str + :param table_name: The table name. Type: string (or Expression with + resultType string). + :type table_name: object """ _validation = { @@ -55,8 +58,10 @@ class MariaDBTableDataset(Dataset): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, + 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } def __init__(self, **kwargs): super(MariaDBTableDataset, self).__init__(**kwargs) + self.table_name = kwargs.get('table_name', None) self.type = 'MariaDBTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_table_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_table_dataset_py3.py index 6b9b1149e4f3..af0a17cac20f 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_table_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_table_dataset_py3.py @@ -39,6 +39,9 @@ class MariaDBTableDataset(Dataset): :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str + :param table_name: The table name. Type: string (or Expression with + resultType string). + :type table_name: object """ _validation = { @@ -55,8 +58,10 @@ class MariaDBTableDataset(Dataset): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, + 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: super(MariaDBTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + self.table_name = table_name self.type = 'MariaDBTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/marketo_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/marketo_object_dataset.py index ee2b58603215..dba47a291f96 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/marketo_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/marketo_object_dataset.py @@ -39,6 +39,9 @@ class MarketoObjectDataset(Dataset): :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str + :param table_name: The table name. Type: string (or Expression with + resultType string). + :type table_name: object """ _validation = { @@ -55,8 +58,10 @@ class MarketoObjectDataset(Dataset): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, + 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } def __init__(self, **kwargs): super(MarketoObjectDataset, self).__init__(**kwargs) + self.table_name = kwargs.get('table_name', None) self.type = 'MarketoObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/marketo_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/marketo_object_dataset_py3.py index 9218191b80c1..1322fee85893 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/marketo_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/marketo_object_dataset_py3.py @@ -39,6 +39,9 @@ class MarketoObjectDataset(Dataset): :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str + :param table_name: The table name. Type: string (or Expression with + resultType string). + :type table_name: object """ _validation = { @@ -55,8 +58,10 @@ class MarketoObjectDataset(Dataset): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, + 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: super(MarketoObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + self.table_name = table_name self.type = 'MarketoObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/netezza_table_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/netezza_table_dataset.py index ddbfe1b223b8..8780abac2f6c 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/netezza_table_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/netezza_table_dataset.py @@ -39,6 +39,9 @@ class NetezzaTableDataset(Dataset): :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str + :param table_name: The table name. Type: string (or Expression with + resultType string). + :type table_name: object """ _validation = { @@ -55,8 +58,10 @@ class NetezzaTableDataset(Dataset): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, + 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } def __init__(self, **kwargs): super(NetezzaTableDataset, self).__init__(**kwargs) + self.table_name = kwargs.get('table_name', None) self.type = 'NetezzaTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/netezza_table_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/netezza_table_dataset_py3.py index 4540fae23f87..1bc693017215 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/netezza_table_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/netezza_table_dataset_py3.py @@ -39,6 +39,9 @@ class NetezzaTableDataset(Dataset): :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str + :param table_name: The table name. Type: string (or Expression with + resultType string). + :type table_name: object """ _validation = { @@ -55,8 +58,10 @@ class NetezzaTableDataset(Dataset): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, + 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: super(NetezzaTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + self.table_name = table_name self.type = 'NetezzaTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_object_dataset.py index 0fd71aabe0c7..980638e5defa 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_object_dataset.py @@ -39,6 +39,9 @@ class PaypalObjectDataset(Dataset): :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str + :param table_name: The table name. Type: string (or Expression with + resultType string). + :type table_name: object """ _validation = { @@ -55,8 +58,10 @@ class PaypalObjectDataset(Dataset): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, + 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } def __init__(self, **kwargs): super(PaypalObjectDataset, self).__init__(**kwargs) + self.table_name = kwargs.get('table_name', None) self.type = 'PaypalObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_object_dataset_py3.py index f44966955b76..beb80be03787 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_object_dataset_py3.py @@ -39,6 +39,9 @@ class PaypalObjectDataset(Dataset): :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str + :param table_name: The table name. Type: string (or Expression with + resultType string). + :type table_name: object """ _validation = { @@ -55,8 +58,10 @@ class PaypalObjectDataset(Dataset): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, + 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: super(PaypalObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + self.table_name = table_name self.type = 'PaypalObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/phoenix_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/phoenix_object_dataset.py index 4573ed10a086..44b271aeb1dd 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/phoenix_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/phoenix_object_dataset.py @@ -39,6 +39,9 @@ class PhoenixObjectDataset(Dataset): :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str + :param table_name: The table name. Type: string (or Expression with + resultType string). + :type table_name: object """ _validation = { @@ -55,8 +58,10 @@ class PhoenixObjectDataset(Dataset): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, + 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } def __init__(self, **kwargs): super(PhoenixObjectDataset, self).__init__(**kwargs) + self.table_name = kwargs.get('table_name', None) self.type = 'PhoenixObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/phoenix_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/phoenix_object_dataset_py3.py index b10bf84d2f52..1f9f6e66dec2 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/phoenix_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/phoenix_object_dataset_py3.py @@ -39,6 +39,9 @@ class PhoenixObjectDataset(Dataset): :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str + :param table_name: The table name. Type: string (or Expression with + resultType string). + :type table_name: object """ _validation = { @@ -55,8 +58,10 @@ class PhoenixObjectDataset(Dataset): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, + 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: super(PhoenixObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + self.table_name = table_name self.type = 'PhoenixObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/presto_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/presto_object_dataset.py index e2215ca65bbe..f94090dbd359 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/presto_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/presto_object_dataset.py @@ -39,6 +39,9 @@ class PrestoObjectDataset(Dataset): :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str + :param table_name: The table name. Type: string (or Expression with + resultType string). + :type table_name: object """ _validation = { @@ -55,8 +58,10 @@ class PrestoObjectDataset(Dataset): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, + 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } def __init__(self, **kwargs): super(PrestoObjectDataset, self).__init__(**kwargs) + self.table_name = kwargs.get('table_name', None) self.type = 'PrestoObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/presto_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/presto_object_dataset_py3.py index 1cb012b082e0..c65b6d6d6520 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/presto_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/presto_object_dataset_py3.py @@ -39,6 +39,9 @@ class PrestoObjectDataset(Dataset): :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str + :param table_name: The table name. Type: string (or Expression with + resultType string). + :type table_name: object """ _validation = { @@ -55,8 +58,10 @@ class PrestoObjectDataset(Dataset): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, + 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: super(PrestoObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + self.table_name = table_name self.type = 'PrestoObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_object_dataset.py index 7b9a358c5adf..1653ee62d028 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_object_dataset.py @@ -39,6 +39,9 @@ class QuickBooksObjectDataset(Dataset): :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str + :param table_name: The table name. Type: string (or Expression with + resultType string). + :type table_name: object """ _validation = { @@ -55,8 +58,10 @@ class QuickBooksObjectDataset(Dataset): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, + 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } def __init__(self, **kwargs): super(QuickBooksObjectDataset, self).__init__(**kwargs) + self.table_name = kwargs.get('table_name', None) self.type = 'QuickBooksObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_object_dataset_py3.py index 63ea660c5330..9273418585b8 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_object_dataset_py3.py @@ -39,6 +39,9 @@ class QuickBooksObjectDataset(Dataset): :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str + :param table_name: The table name. Type: string (or Expression with + resultType string). + :type table_name: object """ _validation = { @@ -55,8 +58,10 @@ class QuickBooksObjectDataset(Dataset): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, + 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: super(QuickBooksObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + self.table_name = table_name self.type = 'QuickBooksObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_object_dataset.py index 84bc7035f7f4..78b2e6859f7e 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_object_dataset.py @@ -39,6 +39,9 @@ class ResponsysObjectDataset(Dataset): :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str + :param table_name: The table name. Type: string (or Expression with + resultType string). + :type table_name: object """ _validation = { @@ -55,8 +58,10 @@ class ResponsysObjectDataset(Dataset): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, + 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } def __init__(self, **kwargs): super(ResponsysObjectDataset, self).__init__(**kwargs) + self.table_name = kwargs.get('table_name', None) self.type = 'ResponsysObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_object_dataset_py3.py index 682db13e5c4a..59d3e2d001d4 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_object_dataset_py3.py @@ -39,6 +39,9 @@ class ResponsysObjectDataset(Dataset): :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str + :param table_name: The table name. Type: string (or Expression with + resultType string). + :type table_name: object """ _validation = { @@ -55,8 +58,10 @@ class ResponsysObjectDataset(Dataset): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, + 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: super(ResponsysObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + self.table_name = table_name self.type = 'ResponsysObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_marketing_cloud_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_marketing_cloud_object_dataset.py index 32b02df94ea6..3dfba59c2dab 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_marketing_cloud_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_marketing_cloud_object_dataset.py @@ -39,6 +39,9 @@ class SalesforceMarketingCloudObjectDataset(Dataset): :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str + :param table_name: The table name. Type: string (or Expression with + resultType string). + :type table_name: object """ _validation = { @@ -55,8 +58,10 @@ class SalesforceMarketingCloudObjectDataset(Dataset): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, + 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } def __init__(self, **kwargs): super(SalesforceMarketingCloudObjectDataset, self).__init__(**kwargs) + self.table_name = kwargs.get('table_name', None) self.type = 'SalesforceMarketingCloudObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_marketing_cloud_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_marketing_cloud_object_dataset_py3.py index 6e260fac8a66..31490518a869 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_marketing_cloud_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_marketing_cloud_object_dataset_py3.py @@ -39,6 +39,9 @@ class SalesforceMarketingCloudObjectDataset(Dataset): :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str + :param table_name: The table name. Type: string (or Expression with + resultType string). + :type table_name: object """ _validation = { @@ -55,8 +58,10 @@ class SalesforceMarketingCloudObjectDataset(Dataset): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, + 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: super(SalesforceMarketingCloudObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + self.table_name = table_name self.type = 'SalesforceMarketingCloudObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_object_dataset.py index 382d005ad5c5..96b86b1f8f9c 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_object_dataset.py @@ -39,6 +39,9 @@ class ServiceNowObjectDataset(Dataset): :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str + :param table_name: The table name. Type: string (or Expression with + resultType string). + :type table_name: object """ _validation = { @@ -55,8 +58,10 @@ class ServiceNowObjectDataset(Dataset): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, + 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } def __init__(self, **kwargs): super(ServiceNowObjectDataset, self).__init__(**kwargs) + self.table_name = kwargs.get('table_name', None) self.type = 'ServiceNowObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_object_dataset_py3.py index 82d0177ff286..54c7f0225a0f 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_object_dataset_py3.py @@ -39,6 +39,9 @@ class ServiceNowObjectDataset(Dataset): :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str + :param table_name: The table name. Type: string (or Expression with + resultType string). + :type table_name: object """ _validation = { @@ -55,8 +58,10 @@ class ServiceNowObjectDataset(Dataset): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, + 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: super(ServiceNowObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + self.table_name = table_name self.type = 'ServiceNowObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_object_dataset.py index a70be7708fb2..690ef424e55c 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_object_dataset.py @@ -39,6 +39,9 @@ class ShopifyObjectDataset(Dataset): :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str + :param table_name: The table name. Type: string (or Expression with + resultType string). + :type table_name: object """ _validation = { @@ -55,8 +58,10 @@ class ShopifyObjectDataset(Dataset): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, + 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } def __init__(self, **kwargs): super(ShopifyObjectDataset, self).__init__(**kwargs) + self.table_name = kwargs.get('table_name', None) self.type = 'ShopifyObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_object_dataset_py3.py index 548d10f44028..5d9304404b35 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_object_dataset_py3.py @@ -39,6 +39,9 @@ class ShopifyObjectDataset(Dataset): :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str + :param table_name: The table name. Type: string (or Expression with + resultType string). + :type table_name: object """ _validation = { @@ -55,8 +58,10 @@ class ShopifyObjectDataset(Dataset): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, + 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: super(ShopifyObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + self.table_name = table_name self.type = 'ShopifyObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/spark_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/spark_object_dataset.py index 52d31be1f06e..a33390e391fe 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/spark_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/spark_object_dataset.py @@ -39,6 +39,9 @@ class SparkObjectDataset(Dataset): :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str + :param table_name: The table name. Type: string (or Expression with + resultType string). + :type table_name: object """ _validation = { @@ -55,8 +58,10 @@ class SparkObjectDataset(Dataset): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, + 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } def __init__(self, **kwargs): super(SparkObjectDataset, self).__init__(**kwargs) + self.table_name = kwargs.get('table_name', None) self.type = 'SparkObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/spark_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/spark_object_dataset_py3.py index ceb7634d9237..de4736d96ead 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/spark_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/spark_object_dataset_py3.py @@ -39,6 +39,9 @@ class SparkObjectDataset(Dataset): :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str + :param table_name: The table name. Type: string (or Expression with + resultType string). + :type table_name: object """ _validation = { @@ -55,8 +58,10 @@ class SparkObjectDataset(Dataset): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, + 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: super(SparkObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + self.table_name = table_name self.type = 'SparkObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_object_dataset.py index 77d8b7f3f3c3..3bc47707d770 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_object_dataset.py @@ -39,6 +39,9 @@ class SquareObjectDataset(Dataset): :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str + :param table_name: The table name. Type: string (or Expression with + resultType string). + :type table_name: object """ _validation = { @@ -55,8 +58,10 @@ class SquareObjectDataset(Dataset): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, + 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } def __init__(self, **kwargs): super(SquareObjectDataset, self).__init__(**kwargs) + self.table_name = kwargs.get('table_name', None) self.type = 'SquareObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_object_dataset_py3.py index d7ea66c2d81a..282ff4922514 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_object_dataset_py3.py @@ -39,6 +39,9 @@ class SquareObjectDataset(Dataset): :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str + :param table_name: The table name. Type: string (or Expression with + resultType string). + :type table_name: object """ _validation = { @@ -55,8 +58,10 @@ class SquareObjectDataset(Dataset): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, + 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: super(SquareObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + self.table_name = table_name self.type = 'SquareObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/stored_procedure_parameter.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/stored_procedure_parameter.py index 63d2c3b80aa5..748cf7cba53c 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/stored_procedure_parameter.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/stored_procedure_parameter.py @@ -15,9 +15,7 @@ class StoredProcedureParameter(Model): """SQL stored procedure parameter. - All required parameters must be populated in order to send to Azure. - - :param value: Required. Stored procedure parameter value. Type: string (or + :param value: Stored procedure parameter value. Type: string (or Expression with resultType string). :type value: object :param type: Stored procedure parameter type. Possible values include: @@ -26,10 +24,6 @@ class StoredProcedureParameter(Model): ~azure.mgmt.datafactory.models.StoredProcedureParameterType """ - _validation = { - 'value': {'required': True}, - } - _attribute_map = { 'value': {'key': 'value', 'type': 'object'}, 'type': {'key': 'type', 'type': 'str'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/stored_procedure_parameter_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/stored_procedure_parameter_py3.py index f5c55777e0c3..bd967ce52876 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/stored_procedure_parameter_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/stored_procedure_parameter_py3.py @@ -15,9 +15,7 @@ class StoredProcedureParameter(Model): """SQL stored procedure parameter. - All required parameters must be populated in order to send to Azure. - - :param value: Required. Stored procedure parameter value. Type: string (or + :param value: Stored procedure parameter value. Type: string (or Expression with resultType string). :type value: object :param type: Stored procedure parameter type. Possible values include: @@ -26,16 +24,12 @@ class StoredProcedureParameter(Model): ~azure.mgmt.datafactory.models.StoredProcedureParameterType """ - _validation = { - 'value': {'required': True}, - } - _attribute_map = { 'value': {'key': 'value', 'type': 'object'}, 'type': {'key': 'type', 'type': 'str'}, } - def __init__(self, *, value, type=None, **kwargs) -> None: + def __init__(self, *, value=None, type=None, **kwargs) -> None: super(StoredProcedureParameter, self).__init__(**kwargs) self.value = value self.type = type diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/vertica_table_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/vertica_table_dataset.py index 755a0225e1f9..8ef0d1a07d99 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/vertica_table_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/vertica_table_dataset.py @@ -39,6 +39,9 @@ class VerticaTableDataset(Dataset): :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str + :param table_name: The table name. Type: string (or Expression with + resultType string). + :type table_name: object """ _validation = { @@ -55,8 +58,10 @@ class VerticaTableDataset(Dataset): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, + 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } def __init__(self, **kwargs): super(VerticaTableDataset, self).__init__(**kwargs) + self.table_name = kwargs.get('table_name', None) self.type = 'VerticaTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/vertica_table_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/vertica_table_dataset_py3.py index cd6f7186855c..82be33c77831 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/vertica_table_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/vertica_table_dataset_py3.py @@ -39,6 +39,9 @@ class VerticaTableDataset(Dataset): :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str + :param table_name: The table name. Type: string (or Expression with + resultType string). + :type table_name: object """ _validation = { @@ -55,8 +58,10 @@ class VerticaTableDataset(Dataset): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, + 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: super(VerticaTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + self.table_name = table_name self.type = 'VerticaTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_object_dataset.py index f3efb679137d..58f2ab42fe1f 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_object_dataset.py @@ -39,6 +39,9 @@ class XeroObjectDataset(Dataset): :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str + :param table_name: The table name. Type: string (or Expression with + resultType string). + :type table_name: object """ _validation = { @@ -55,8 +58,10 @@ class XeroObjectDataset(Dataset): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, + 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } def __init__(self, **kwargs): super(XeroObjectDataset, self).__init__(**kwargs) + self.table_name = kwargs.get('table_name', None) self.type = 'XeroObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_object_dataset_py3.py index 3c5c970f2310..2b1f6d82814a 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_object_dataset_py3.py @@ -39,6 +39,9 @@ class XeroObjectDataset(Dataset): :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str + :param table_name: The table name. Type: string (or Expression with + resultType string). + :type table_name: object """ _validation = { @@ -55,8 +58,10 @@ class XeroObjectDataset(Dataset): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, + 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: super(XeroObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + self.table_name = table_name self.type = 'XeroObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/zoho_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/zoho_object_dataset.py index fb988d8e5e04..e2df390d74ea 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/zoho_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/zoho_object_dataset.py @@ -39,6 +39,9 @@ class ZohoObjectDataset(Dataset): :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str + :param table_name: The table name. Type: string (or Expression with + resultType string). + :type table_name: object """ _validation = { @@ -55,8 +58,10 @@ class ZohoObjectDataset(Dataset): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, + 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } def __init__(self, **kwargs): super(ZohoObjectDataset, self).__init__(**kwargs) + self.table_name = kwargs.get('table_name', None) self.type = 'ZohoObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/zoho_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/zoho_object_dataset_py3.py index 32cac965c535..465e922ecd34 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/zoho_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/zoho_object_dataset_py3.py @@ -39,6 +39,9 @@ class ZohoObjectDataset(Dataset): :type folder: ~azure.mgmt.datafactory.models.DatasetFolder :param type: Required. Constant filled by server. :type type: str + :param table_name: The table name. Type: string (or Expression with + resultType string). + :type table_name: object """ _validation = { @@ -55,8 +58,10 @@ class ZohoObjectDataset(Dataset): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'type': {'key': 'type', 'type': 'str'}, + 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: super(ZohoObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + self.table_name = table_name self.type = 'ZohoObject' From 5933a201f2c92b474c542bd8b4046dcbeecf7e54 Mon Sep 17 00:00:00 2001 From: Azure SDK for Python bot Date: Tue, 20 Nov 2018 13:27:26 -0800 Subject: [PATCH 31/54] [AutoPR datafactory/resource-manager] Adding get Dataplane readonly token API (#3387) * Generated from 19c65d65e8c41ccca7d29b996cbece26c8d428b7 Updated the casing and the example * Generated from 066ef3258cc2941ad470d91dea4add2b2c5665b1 Updating examples * Packaging update of azure-mgmt-datafactory * Generated from 066ef3258cc2941ad470d91dea4add2b2c5665b1 Updating examples --- .../azure/mgmt/datafactory/models/__init__.py | 6 ++ .../models/access_policy_response.py | 36 ++++++++++ .../models/access_policy_response_py3.py | 36 ++++++++++ .../datafactory/models/user_access_policy.py | 51 ++++++++++++++ .../models/user_access_policy_py3.py | 51 ++++++++++++++ .../operations/factories_operations.py | 68 +++++++++++++++++++ 6 files changed, 248 insertions(+) create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/access_policy_response.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/access_policy_response_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/user_access_policy.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/user_access_policy_py3.py diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py index 4970f33f96e1..d185d578748b 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py @@ -50,6 +50,8 @@ from .factory_repo_update_py3 import FactoryRepoUpdate from .git_hub_access_token_request_py3 import GitHubAccessTokenRequest from .git_hub_access_token_response_py3 import GitHubAccessTokenResponse + from .user_access_policy_py3 import UserAccessPolicy + from .access_policy_response_py3 import AccessPolicyResponse from .pipeline_reference_py3 import PipelineReference from .trigger_pipeline_reference_py3 import TriggerPipelineReference from .factory_update_parameters_py3 import FactoryUpdateParameters @@ -403,6 +405,8 @@ from .factory_repo_update import FactoryRepoUpdate from .git_hub_access_token_request import GitHubAccessTokenRequest from .git_hub_access_token_response import GitHubAccessTokenResponse + from .user_access_policy import UserAccessPolicy + from .access_policy_response import AccessPolicyResponse from .pipeline_reference import PipelineReference from .trigger_pipeline_reference import TriggerPipelineReference from .factory_update_parameters import FactoryUpdateParameters @@ -826,6 +830,8 @@ 'FactoryRepoUpdate', 'GitHubAccessTokenRequest', 'GitHubAccessTokenResponse', + 'UserAccessPolicy', + 'AccessPolicyResponse', 'PipelineReference', 'TriggerPipelineReference', 'FactoryUpdateParameters', diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/access_policy_response.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/access_policy_response.py new file mode 100644 index 000000000000..033d0fd9591f --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/access_policy_response.py @@ -0,0 +1,36 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class AccessPolicyResponse(Model): + """Get Data Plane read only token response definition. + + :param policy: The user access policy. + :type policy: ~azure.mgmt.datafactory.models.UserAccessPolicy + :param access_token: Data Plane read only access token. + :type access_token: str + :param data_plane_url: Data Plane service base URL. + :type data_plane_url: str + """ + + _attribute_map = { + 'policy': {'key': 'policy', 'type': 'UserAccessPolicy'}, + 'access_token': {'key': 'accessToken', 'type': 'str'}, + 'data_plane_url': {'key': 'dataPlaneUrl', 'type': 'str'}, + } + + def __init__(self, **kwargs): + super(AccessPolicyResponse, self).__init__(**kwargs) + self.policy = kwargs.get('policy', None) + self.access_token = kwargs.get('access_token', None) + self.data_plane_url = kwargs.get('data_plane_url', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/access_policy_response_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/access_policy_response_py3.py new file mode 100644 index 000000000000..2932f547ff26 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/access_policy_response_py3.py @@ -0,0 +1,36 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class AccessPolicyResponse(Model): + """Get Data Plane read only token response definition. + + :param policy: The user access policy. + :type policy: ~azure.mgmt.datafactory.models.UserAccessPolicy + :param access_token: Data Plane read only access token. + :type access_token: str + :param data_plane_url: Data Plane service base URL. + :type data_plane_url: str + """ + + _attribute_map = { + 'policy': {'key': 'policy', 'type': 'UserAccessPolicy'}, + 'access_token': {'key': 'accessToken', 'type': 'str'}, + 'data_plane_url': {'key': 'dataPlaneUrl', 'type': 'str'}, + } + + def __init__(self, *, policy=None, access_token: str=None, data_plane_url: str=None, **kwargs) -> None: + super(AccessPolicyResponse, self).__init__(**kwargs) + self.policy = policy + self.access_token = access_token + self.data_plane_url = data_plane_url diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/user_access_policy.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/user_access_policy.py new file mode 100644 index 000000000000..b51e313b6f0c --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/user_access_policy.py @@ -0,0 +1,51 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class UserAccessPolicy(Model): + """Get Data Plane read only token request definition. + + :param permissions: The string with permissions for Data Plane access. + Currently only 'r' is supported which grants read only access. + :type permissions: str + :param access_resource_path: The resource path to get access relative to + factory. Currently only empty string is supported which corresponds to the + factory resource. + :type access_resource_path: str + :param profile_name: The name of the profile. Currently only the default + is supported. The default value is DefaultProfile. + :type profile_name: str + :param start_time: Start time for the token. If not specified the current + time will be used. + :type start_time: str + :param expire_time: Expiration time for the token. Maximum duration for + the token is eight hours and by default the token will expire in eight + hours. + :type expire_time: str + """ + + _attribute_map = { + 'permissions': {'key': 'permissions', 'type': 'str'}, + 'access_resource_path': {'key': 'accessResourcePath', 'type': 'str'}, + 'profile_name': {'key': 'profileName', 'type': 'str'}, + 'start_time': {'key': 'startTime', 'type': 'str'}, + 'expire_time': {'key': 'expireTime', 'type': 'str'}, + } + + def __init__(self, **kwargs): + super(UserAccessPolicy, self).__init__(**kwargs) + self.permissions = kwargs.get('permissions', None) + self.access_resource_path = kwargs.get('access_resource_path', None) + self.profile_name = kwargs.get('profile_name', None) + self.start_time = kwargs.get('start_time', None) + self.expire_time = kwargs.get('expire_time', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/user_access_policy_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/user_access_policy_py3.py new file mode 100644 index 000000000000..26e2a7639a09 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/user_access_policy_py3.py @@ -0,0 +1,51 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class UserAccessPolicy(Model): + """Get Data Plane read only token request definition. + + :param permissions: The string with permissions for Data Plane access. + Currently only 'r' is supported which grants read only access. + :type permissions: str + :param access_resource_path: The resource path to get access relative to + factory. Currently only empty string is supported which corresponds to the + factory resource. + :type access_resource_path: str + :param profile_name: The name of the profile. Currently only the default + is supported. The default value is DefaultProfile. + :type profile_name: str + :param start_time: Start time for the token. If not specified the current + time will be used. + :type start_time: str + :param expire_time: Expiration time for the token. Maximum duration for + the token is eight hours and by default the token will expire in eight + hours. + :type expire_time: str + """ + + _attribute_map = { + 'permissions': {'key': 'permissions', 'type': 'str'}, + 'access_resource_path': {'key': 'accessResourcePath', 'type': 'str'}, + 'profile_name': {'key': 'profileName', 'type': 'str'}, + 'start_time': {'key': 'startTime', 'type': 'str'}, + 'expire_time': {'key': 'expireTime', 'type': 'str'}, + } + + def __init__(self, *, permissions: str=None, access_resource_path: str=None, profile_name: str=None, start_time: str=None, expire_time: str=None, **kwargs) -> None: + super(UserAccessPolicy, self).__init__(**kwargs) + self.permissions = permissions + self.access_resource_path = access_resource_path + self.profile_name = profile_name + self.start_time = start_time + self.expire_time = expire_time diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/factories_operations.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/factories_operations.py index 975421b1e222..b06c12f3e8c5 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/factories_operations.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/factories_operations.py @@ -574,3 +574,71 @@ def get_git_hub_access_token( return deserialized get_git_hub_access_token.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/getGitHubAccessToken'} + + def get_data_plane_access( + self, resource_group_name, factory_name, policy, custom_headers=None, raw=False, **operation_config): + """Get Data Plane access. + + :param resource_group_name: The resource group name. + :type resource_group_name: str + :param factory_name: The factory name. + :type factory_name: str + :param policy: Data Plane user access policy definition. + :type policy: ~azure.mgmt.datafactory.models.UserAccessPolicy + :param dict custom_headers: headers that will be added to the request + :param bool raw: returns the direct response alongside the + deserialized response + :param operation_config: :ref:`Operation configuration + overrides`. + :return: AccessPolicyResponse or ClientRawResponse if raw=true + :rtype: ~azure.mgmt.datafactory.models.AccessPolicyResponse or + ~msrest.pipeline.ClientRawResponse + :raises: :class:`CloudError` + """ + # Construct URL + url = self.get_data_plane_access.metadata['url'] + path_format_arguments = { + 'subscriptionId': self._serialize.url("self.config.subscription_id", self.config.subscription_id, 'str'), + 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str', max_length=90, min_length=1, pattern=r'^[-\w\._\(\)]+$'), + 'factoryName': self._serialize.url("factory_name", factory_name, 'str', max_length=63, min_length=3, pattern=r'^[A-Za-z0-9]+(?:-[A-Za-z0-9]+)*$') + } + url = self._client.format_url(url, **path_format_arguments) + + # Construct parameters + query_parameters = {} + query_parameters['api-version'] = self._serialize.query("self.api_version", self.api_version, 'str') + + # Construct headers + header_parameters = {} + header_parameters['Accept'] = 'application/json' + header_parameters['Content-Type'] = 'application/json; charset=utf-8' + if self.config.generate_client_request_id: + header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) + if custom_headers: + header_parameters.update(custom_headers) + if self.config.accept_language is not None: + header_parameters['accept-language'] = self._serialize.header("self.config.accept_language", self.config.accept_language, 'str') + + # Construct body + body_content = self._serialize.body(policy, 'UserAccessPolicy') + + # Construct and send request + request = self._client.post(url, query_parameters, header_parameters, body_content) + response = self._client.send(request, stream=False, **operation_config) + + if response.status_code not in [200]: + exp = CloudError(response) + exp.request_id = response.headers.get('x-ms-request-id') + raise exp + + deserialized = None + + if response.status_code == 200: + deserialized = self._deserialize('AccessPolicyResponse', response) + + if raw: + client_raw_response = ClientRawResponse(deserialized, response) + return client_raw_response + + return deserialized + get_data_plane_access.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/getDataPlaneAccess'} From 66b88fdcf7fb6ba9623c4e91b91dcc6fda424103 Mon Sep 17 00:00:00 2001 From: Azure SDK for Python bot Date: Mon, 26 Nov 2018 10:14:35 -0800 Subject: [PATCH 32/54] Generated from 247c6a22105cd7c58a189359cf973edcb0ff05bd (#3855) typo: datafactory/resource-manager/Microsoft.DataFactory/entityTypes/LinkedService - retrive -> retrieve - seperate -> separate - Serivce- > Service --- .../mgmt/datafactory/models/amazon_mws_linked_service.py | 4 ++-- .../mgmt/datafactory/models/amazon_mws_linked_service_py3.py | 4 ++-- .../azure/mgmt/datafactory/models/concur_linked_service.py | 2 +- .../mgmt/datafactory/models/concur_linked_service_py3.py | 2 +- .../azure/mgmt/datafactory/models/hubspot_linked_service.py | 2 +- .../mgmt/datafactory/models/hubspot_linked_service_py3.py | 2 +- .../azure/mgmt/datafactory/models/jira_linked_service.py | 2 +- .../azure/mgmt/datafactory/models/jira_linked_service_py3.py | 2 +- .../azure/mgmt/datafactory/models/paypal_linked_service.py | 2 +- .../mgmt/datafactory/models/paypal_linked_service_py3.py | 2 +- .../azure/mgmt/datafactory/models/shopify_linked_service.py | 2 +- .../mgmt/datafactory/models/shopify_linked_service_py3.py | 2 +- .../azure/mgmt/datafactory/models/square_linked_service.py | 2 +- .../mgmt/datafactory/models/square_linked_service_py3.py | 2 +- .../azure/mgmt/datafactory/models/xero_linked_service.py | 2 +- .../azure/mgmt/datafactory/models/xero_linked_service_py3.py | 2 +- 16 files changed, 18 insertions(+), 18 deletions(-) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_linked_service.py index f83ccbfa9568..4531b28777c6 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_linked_service.py @@ -37,8 +37,8 @@ class AmazonMWSLinkedService(LinkedService): mws.amazonservices.com) :type endpoint: object :param marketplace_id: Required. The Amazon Marketplace ID you want to - retrieve data from. To retrive data from multiple Marketplace IDs, - seperate them with a comma (,). (i.e. A2EUQ1WTGCTBG2) + retrieve data from. To retrieve data from multiple Marketplace IDs, + separate them with a comma (,). (i.e. A2EUQ1WTGCTBG2) :type marketplace_id: object :param seller_id: Required. The Amazon seller ID. :type seller_id: object diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_linked_service_py3.py index e1a5ba11f560..421c20dc2d4a 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_linked_service_py3.py @@ -37,8 +37,8 @@ class AmazonMWSLinkedService(LinkedService): mws.amazonservices.com) :type endpoint: object :param marketplace_id: Required. The Amazon Marketplace ID you want to - retrieve data from. To retrive data from multiple Marketplace IDs, - seperate them with a comma (,). (i.e. A2EUQ1WTGCTBG2) + retrieve data from. To retrieve data from multiple Marketplace IDs, + separate them with a comma (,). (i.e. A2EUQ1WTGCTBG2) :type marketplace_id: object :param seller_id: Required. The Amazon seller ID. :type seller_id: object diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_linked_service.py index f8bc6443d656..7b85f1400ff6 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_linked_service.py @@ -13,7 +13,7 @@ class ConcurLinkedService(LinkedService): - """Concur Serivce linked service. + """Concur Service linked service. All required parameters must be populated in order to send to Azure. diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_linked_service_py3.py index 790e6bc76b76..6e17a2c9cc8e 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_linked_service_py3.py @@ -13,7 +13,7 @@ class ConcurLinkedService(LinkedService): - """Concur Serivce linked service. + """Concur Service linked service. All required parameters must be populated in order to send to Azure. diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_linked_service.py index bd446c19c6da..08af04633c12 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_linked_service.py @@ -13,7 +13,7 @@ class HubspotLinkedService(LinkedService): - """Hubspot Serivce linked service. + """Hubspot Service linked service. All required parameters must be populated in order to send to Azure. diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_linked_service_py3.py index 5c92b45e0252..93f66cd8e17b 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_linked_service_py3.py @@ -13,7 +13,7 @@ class HubspotLinkedService(LinkedService): - """Hubspot Serivce linked service. + """Hubspot Service linked service. All required parameters must be populated in order to send to Azure. diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_linked_service.py index 4ea04a954079..d8b9a62fc878 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_linked_service.py @@ -13,7 +13,7 @@ class JiraLinkedService(LinkedService): - """Jira Serivce linked service. + """Jira Service linked service. All required parameters must be populated in order to send to Azure. diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_linked_service_py3.py index 327ce4d0793e..69606ee7cfcf 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_linked_service_py3.py @@ -13,7 +13,7 @@ class JiraLinkedService(LinkedService): - """Jira Serivce linked service. + """Jira Service linked service. All required parameters must be populated in order to send to Azure. diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_linked_service.py index fa7ba90e1e77..190fc45985d3 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_linked_service.py @@ -13,7 +13,7 @@ class PaypalLinkedService(LinkedService): - """Paypal Serivce linked service. + """Paypal Service linked service. All required parameters must be populated in order to send to Azure. diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_linked_service_py3.py index 41fdb52595ab..832b0dff257b 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_linked_service_py3.py @@ -13,7 +13,7 @@ class PaypalLinkedService(LinkedService): - """Paypal Serivce linked service. + """Paypal Service linked service. All required parameters must be populated in order to send to Azure. diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_linked_service.py index 8d38e5d64a17..b57922620ef8 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_linked_service.py @@ -13,7 +13,7 @@ class ShopifyLinkedService(LinkedService): - """Shopify Serivce linked service. + """Shopify Service linked service. All required parameters must be populated in order to send to Azure. diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_linked_service_py3.py index 8d7d659bebe1..714de7f0ddf6 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_linked_service_py3.py @@ -13,7 +13,7 @@ class ShopifyLinkedService(LinkedService): - """Shopify Serivce linked service. + """Shopify Service linked service. All required parameters must be populated in order to send to Azure. diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_linked_service.py index 2b175ead12df..4e9df2b68e62 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_linked_service.py @@ -13,7 +13,7 @@ class SquareLinkedService(LinkedService): - """Square Serivce linked service. + """Square Service linked service. All required parameters must be populated in order to send to Azure. diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_linked_service_py3.py index ffbe49d74ee8..0b9218efba97 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_linked_service_py3.py @@ -13,7 +13,7 @@ class SquareLinkedService(LinkedService): - """Square Serivce linked service. + """Square Service linked service. All required parameters must be populated in order to send to Azure. diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_linked_service.py index 84129d6e38d1..e9daa4ff7d2a 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_linked_service.py @@ -13,7 +13,7 @@ class XeroLinkedService(LinkedService): - """Xero Serivce linked service. + """Xero Service linked service. All required parameters must be populated in order to send to Azure. diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_linked_service_py3.py index 7a8a08d0bb22..eb665519f4ea 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_linked_service_py3.py @@ -13,7 +13,7 @@ class XeroLinkedService(LinkedService): - """Xero Serivce linked service. + """Xero Service linked service. All required parameters must be populated in order to send to Azure. From 8fde68547addd6304ad869d5851ccdefd11b69f1 Mon Sep 17 00:00:00 2001 From: Azure SDK for Python bot Date: Tue, 27 Nov 2018 10:47:51 -0800 Subject: [PATCH 33/54] Generated from daf7fe30a7e5857636a834d880f9b6345b41bfac (#3878) typo: datafactory/resource-manager/Microsoft.DataF actory - availably -> availability - Serivce -> Service - specificed -> specified - occurence -> occurrence - everytime -> every time - Double word "the" - Rrepository -> Repository - proprety -> property --- .../azure/mgmt/datafactory/models/blob_events_trigger.py | 2 +- .../azure/mgmt/datafactory/models/blob_events_trigger_py3.py | 2 +- .../azure/mgmt/datafactory/models/blob_trigger.py | 2 +- .../azure/mgmt/datafactory/models/blob_trigger_py3.py | 2 +- .../azure/mgmt/datafactory/models/concur_object_dataset.py | 2 +- .../mgmt/datafactory/models/concur_object_dataset_py3.py | 2 +- .../azure/mgmt/datafactory/models/concur_source.py | 2 +- .../azure/mgmt/datafactory/models/concur_source_py3.py | 2 +- .../azure/mgmt/datafactory/models/copy_activity.py | 2 +- .../azure/mgmt/datafactory/models/copy_activity_py3.py | 2 +- .../mgmt/datafactory/models/factory_git_hub_configuration.py | 2 +- .../datafactory/models/factory_git_hub_configuration_py3.py | 2 +- .../mgmt/datafactory/models/factory_repo_configuration.py | 2 +- .../mgmt/datafactory/models/factory_repo_configuration_py3.py | 2 +- .../mgmt/datafactory/models/factory_vsts_configuration.py | 2 +- .../mgmt/datafactory/models/factory_vsts_configuration_py3.py | 2 +- .../azure/mgmt/datafactory/models/hubspot_object_dataset.py | 2 +- .../mgmt/datafactory/models/hubspot_object_dataset_py3.py | 2 +- .../azure/mgmt/datafactory/models/hubspot_source.py | 2 +- .../azure/mgmt/datafactory/models/hubspot_source_py3.py | 2 +- .../azure/mgmt/datafactory/models/jira_object_dataset.py | 2 +- .../azure/mgmt/datafactory/models/jira_object_dataset_py3.py | 2 +- .../azure/mgmt/datafactory/models/jira_source.py | 2 +- .../azure/mgmt/datafactory/models/jira_source_py3.py | 2 +- .../azure/mgmt/datafactory/models/paypal_object_dataset.py | 2 +- .../mgmt/datafactory/models/paypal_object_dataset_py3.py | 2 +- .../azure/mgmt/datafactory/models/paypal_source.py | 2 +- .../azure/mgmt/datafactory/models/paypal_source_py3.py | 2 +- .../mgmt/datafactory/models/recurrence_schedule_occurrence.py | 2 +- .../datafactory/models/recurrence_schedule_occurrence_py3.py | 2 +- .../azure/mgmt/datafactory/models/shopify_object_dataset.py | 2 +- .../mgmt/datafactory/models/shopify_object_dataset_py3.py | 2 +- .../azure/mgmt/datafactory/models/shopify_source.py | 2 +- .../azure/mgmt/datafactory/models/shopify_source_py3.py | 2 +- .../azure/mgmt/datafactory/models/square_object_dataset.py | 2 +- .../mgmt/datafactory/models/square_object_dataset_py3.py | 2 +- .../azure/mgmt/datafactory/models/square_source.py | 2 +- .../azure/mgmt/datafactory/models/square_source_py3.py | 2 +- .../azure/mgmt/datafactory/models/user_property.py | 4 ++-- .../azure/mgmt/datafactory/models/user_property_py3.py | 4 ++-- .../azure/mgmt/datafactory/models/xero_object_dataset.py | 2 +- .../azure/mgmt/datafactory/models/xero_object_dataset_py3.py | 2 +- .../azure/mgmt/datafactory/models/xero_source.py | 2 +- .../azure/mgmt/datafactory/models/xero_source_py3.py | 2 +- .../datafactory/operations/integration_runtimes_operations.py | 3 ++- .../azure/mgmt/datafactory/operations/pipelines_operations.py | 4 ++-- 46 files changed, 50 insertions(+), 49 deletions(-) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/blob_events_trigger.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/blob_events_trigger.py index cb8647fc1577..681cc44d278b 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/blob_events_trigger.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/blob_events_trigger.py @@ -13,7 +13,7 @@ class BlobEventsTrigger(MultiplePipelineTrigger): - """Trigger that runs everytime a Blob event occurs. + """Trigger that runs every time a Blob event occurs. Variables are only populated by the server, and will be ignored when sending a request. diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/blob_events_trigger_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/blob_events_trigger_py3.py index 150f9eed2012..08d9c542f4af 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/blob_events_trigger_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/blob_events_trigger_py3.py @@ -13,7 +13,7 @@ class BlobEventsTrigger(MultiplePipelineTrigger): - """Trigger that runs everytime a Blob event occurs. + """Trigger that runs every time a Blob event occurs. Variables are only populated by the server, and will be ignored when sending a request. diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/blob_trigger.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/blob_trigger.py index 07504d83c933..6abdece68966 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/blob_trigger.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/blob_trigger.py @@ -13,7 +13,7 @@ class BlobTrigger(MultiplePipelineTrigger): - """Trigger that runs everytime the selected Blob container changes. + """Trigger that runs every time the selected Blob container changes. Variables are only populated by the server, and will be ignored when sending a request. diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/blob_trigger_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/blob_trigger_py3.py index cf53b39ba72f..2c80ac605368 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/blob_trigger_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/blob_trigger_py3.py @@ -13,7 +13,7 @@ class BlobTrigger(MultiplePipelineTrigger): - """Trigger that runs everytime the selected Blob container changes. + """Trigger that runs every time the selected Blob container changes. Variables are only populated by the server, and will be ignored when sending a request. diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_object_dataset.py index 2f3829b16b20..1aab27f4f8f3 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_object_dataset.py @@ -13,7 +13,7 @@ class ConcurObjectDataset(Dataset): - """Concur Serivce dataset. + """Concur Service dataset. All required parameters must be populated in order to send to Azure. diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_object_dataset_py3.py index 0b7016c75604..9ae3833082c6 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_object_dataset_py3.py @@ -13,7 +13,7 @@ class ConcurObjectDataset(Dataset): - """Concur Serivce dataset. + """Concur Service dataset. All required parameters must be populated in order to send to Azure. diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_source.py index 7320de6aa008..f8053415520c 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_source.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_source.py @@ -13,7 +13,7 @@ class ConcurSource(CopySource): - """A copy activity Concur Serivce source. + """A copy activity Concur Service source. All required parameters must be populated in order to send to Azure. diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_source_py3.py index 5205786a4585..db9104869417 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_source_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_source_py3.py @@ -13,7 +13,7 @@ class ConcurSource(CopySource): - """A copy activity Concur Serivce source. + """A copy activity Concur Service source. All required parameters must be populated in order to send to Azure. diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_activity.py index 997c3820cb00..9182efe2469a 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_activity.py @@ -39,7 +39,7 @@ class CopyActivity(ExecutionActivity): :type source: ~azure.mgmt.datafactory.models.CopySource :param sink: Required. Copy activity sink. :type sink: ~azure.mgmt.datafactory.models.CopySink - :param translator: Copy activity translator. If not specificed, tabular + :param translator: Copy activity translator. If not specified, tabular translator is used. :type translator: ~azure.mgmt.datafactory.models.CopyTranslator :param enable_staging: Specifies whether to copy data via an interim diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_activity_py3.py index 8a740db78983..fd663bd71dc6 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_activity_py3.py @@ -39,7 +39,7 @@ class CopyActivity(ExecutionActivity): :type source: ~azure.mgmt.datafactory.models.CopySource :param sink: Required. Copy activity sink. :type sink: ~azure.mgmt.datafactory.models.CopySink - :param translator: Copy activity translator. If not specificed, tabular + :param translator: Copy activity translator. If not specified, tabular translator is used. :type translator: ~azure.mgmt.datafactory.models.CopyTranslator :param enable_staging: Specifies whether to copy data via an interim diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_git_hub_configuration.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_git_hub_configuration.py index 2607d41d510b..02cec39d8313 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_git_hub_configuration.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_git_hub_configuration.py @@ -19,7 +19,7 @@ class FactoryGitHubConfiguration(FactoryRepoConfiguration): :param account_name: Required. Account name. :type account_name: str - :param repository_name: Required. Rrepository name. + :param repository_name: Required. Repository name. :type repository_name: str :param collaboration_branch: Required. Collaboration branch. :type collaboration_branch: str diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_git_hub_configuration_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_git_hub_configuration_py3.py index 8948a4e4420b..23c5dbf21f0c 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_git_hub_configuration_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_git_hub_configuration_py3.py @@ -19,7 +19,7 @@ class FactoryGitHubConfiguration(FactoryRepoConfiguration): :param account_name: Required. Account name. :type account_name: str - :param repository_name: Required. Rrepository name. + :param repository_name: Required. Repository name. :type repository_name: str :param collaboration_branch: Required. Collaboration branch. :type collaboration_branch: str diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_repo_configuration.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_repo_configuration.py index 753bd2bdb039..7c20db016c71 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_repo_configuration.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_repo_configuration.py @@ -22,7 +22,7 @@ class FactoryRepoConfiguration(Model): :param account_name: Required. Account name. :type account_name: str - :param repository_name: Required. Rrepository name. + :param repository_name: Required. Repository name. :type repository_name: str :param collaboration_branch: Required. Collaboration branch. :type collaboration_branch: str diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_repo_configuration_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_repo_configuration_py3.py index c724cb64a18d..eefed7978850 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_repo_configuration_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_repo_configuration_py3.py @@ -22,7 +22,7 @@ class FactoryRepoConfiguration(Model): :param account_name: Required. Account name. :type account_name: str - :param repository_name: Required. Rrepository name. + :param repository_name: Required. Repository name. :type repository_name: str :param collaboration_branch: Required. Collaboration branch. :type collaboration_branch: str diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_vsts_configuration.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_vsts_configuration.py index 2d8d6d34861a..6d07c68d23e3 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_vsts_configuration.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_vsts_configuration.py @@ -19,7 +19,7 @@ class FactoryVSTSConfiguration(FactoryRepoConfiguration): :param account_name: Required. Account name. :type account_name: str - :param repository_name: Required. Rrepository name. + :param repository_name: Required. Repository name. :type repository_name: str :param collaboration_branch: Required. Collaboration branch. :type collaboration_branch: str diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_vsts_configuration_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_vsts_configuration_py3.py index afa78e1b1f24..4f13c0959d63 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_vsts_configuration_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/factory_vsts_configuration_py3.py @@ -19,7 +19,7 @@ class FactoryVSTSConfiguration(FactoryRepoConfiguration): :param account_name: Required. Account name. :type account_name: str - :param repository_name: Required. Rrepository name. + :param repository_name: Required. Repository name. :type repository_name: str :param collaboration_branch: Required. Collaboration branch. :type collaboration_branch: str diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_object_dataset.py index a48c4ea67a92..7a54a96e855a 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_object_dataset.py @@ -13,7 +13,7 @@ class HubspotObjectDataset(Dataset): - """Hubspot Serivce dataset. + """Hubspot Service dataset. All required parameters must be populated in order to send to Azure. diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_object_dataset_py3.py index 996b626ed551..c7886d783d95 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_object_dataset_py3.py @@ -13,7 +13,7 @@ class HubspotObjectDataset(Dataset): - """Hubspot Serivce dataset. + """Hubspot Service dataset. All required parameters must be populated in order to send to Azure. diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_source.py index 267a6ff3797c..bca6b525860c 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_source.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_source.py @@ -13,7 +13,7 @@ class HubspotSource(CopySource): - """A copy activity Hubspot Serivce source. + """A copy activity Hubspot Service source. All required parameters must be populated in order to send to Azure. diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_source_py3.py index 4162be4216a6..cfc2d2d815b5 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_source_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_source_py3.py @@ -13,7 +13,7 @@ class HubspotSource(CopySource): - """A copy activity Hubspot Serivce source. + """A copy activity Hubspot Service source. All required parameters must be populated in order to send to Azure. diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_object_dataset.py index 614c2fe6a5fb..e197d0d9eb97 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_object_dataset.py @@ -13,7 +13,7 @@ class JiraObjectDataset(Dataset): - """Jira Serivce dataset. + """Jira Service dataset. All required parameters must be populated in order to send to Azure. diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_object_dataset_py3.py index 9a2a445743da..022e88aa0575 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_object_dataset_py3.py @@ -13,7 +13,7 @@ class JiraObjectDataset(Dataset): - """Jira Serivce dataset. + """Jira Service dataset. All required parameters must be populated in order to send to Azure. diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_source.py index 993d5ecb1d11..7bb6a8649b8f 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_source.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_source.py @@ -13,7 +13,7 @@ class JiraSource(CopySource): - """A copy activity Jira Serivce source. + """A copy activity Jira Service source. All required parameters must be populated in order to send to Azure. diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_source_py3.py index 5835c7eceef4..1a19ed99c55a 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_source_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_source_py3.py @@ -13,7 +13,7 @@ class JiraSource(CopySource): - """A copy activity Jira Serivce source. + """A copy activity Jira Service source. All required parameters must be populated in order to send to Azure. diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_object_dataset.py index 980638e5defa..69ac88775d42 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_object_dataset.py @@ -13,7 +13,7 @@ class PaypalObjectDataset(Dataset): - """Paypal Serivce dataset. + """Paypal Service dataset. All required parameters must be populated in order to send to Azure. diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_object_dataset_py3.py index beb80be03787..080e482cf978 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_object_dataset_py3.py @@ -13,7 +13,7 @@ class PaypalObjectDataset(Dataset): - """Paypal Serivce dataset. + """Paypal Service dataset. All required parameters must be populated in order to send to Azure. diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_source.py index 61f89f28bc59..5bb73029d10c 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_source.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_source.py @@ -13,7 +13,7 @@ class PaypalSource(CopySource): - """A copy activity Paypal Serivce source. + """A copy activity Paypal Service source. All required parameters must be populated in order to send to Azure. diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_source_py3.py index 347b0e3e550e..6a9dcce16a2d 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_source_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_source_py3.py @@ -13,7 +13,7 @@ class PaypalSource(CopySource): - """A copy activity Paypal Serivce source. + """A copy activity Paypal Service source. All required parameters must be populated in order to send to Azure. diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/recurrence_schedule_occurrence.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/recurrence_schedule_occurrence.py index f42136545ac2..bbbe1fa28f17 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/recurrence_schedule_occurrence.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/recurrence_schedule_occurrence.py @@ -13,7 +13,7 @@ class RecurrenceScheduleOccurrence(Model): - """The recurrence schedule occurence. + """The recurrence schedule occurrence. :param additional_properties: Unmatched properties from the message are deserialized this collection diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/recurrence_schedule_occurrence_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/recurrence_schedule_occurrence_py3.py index e3173e47e8e4..10aea1f00163 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/recurrence_schedule_occurrence_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/recurrence_schedule_occurrence_py3.py @@ -13,7 +13,7 @@ class RecurrenceScheduleOccurrence(Model): - """The recurrence schedule occurence. + """The recurrence schedule occurrence. :param additional_properties: Unmatched properties from the message are deserialized this collection diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_object_dataset.py index 690ef424e55c..73569a397edc 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_object_dataset.py @@ -13,7 +13,7 @@ class ShopifyObjectDataset(Dataset): - """Shopify Serivce dataset. + """Shopify Service dataset. All required parameters must be populated in order to send to Azure. diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_object_dataset_py3.py index 5d9304404b35..ed18a9f34769 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_object_dataset_py3.py @@ -13,7 +13,7 @@ class ShopifyObjectDataset(Dataset): - """Shopify Serivce dataset. + """Shopify Service dataset. All required parameters must be populated in order to send to Azure. diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_source.py index cc5202bdc6bb..3006ede4633d 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_source.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_source.py @@ -13,7 +13,7 @@ class ShopifySource(CopySource): - """A copy activity Shopify Serivce source. + """A copy activity Shopify Service source. All required parameters must be populated in order to send to Azure. diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_source_py3.py index 953d9246d9a2..ec17bdce3e35 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_source_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_source_py3.py @@ -13,7 +13,7 @@ class ShopifySource(CopySource): - """A copy activity Shopify Serivce source. + """A copy activity Shopify Service source. All required parameters must be populated in order to send to Azure. diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_object_dataset.py index 3bc47707d770..515bab8bbff0 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_object_dataset.py @@ -13,7 +13,7 @@ class SquareObjectDataset(Dataset): - """Square Serivce dataset. + """Square Service dataset. All required parameters must be populated in order to send to Azure. diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_object_dataset_py3.py index 282ff4922514..730f11108723 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_object_dataset_py3.py @@ -13,7 +13,7 @@ class SquareObjectDataset(Dataset): - """Square Serivce dataset. + """Square Service dataset. All required parameters must be populated in order to send to Azure. diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_source.py index 98765da2fea8..919abc0b19fa 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_source.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_source.py @@ -13,7 +13,7 @@ class SquareSource(CopySource): - """A copy activity Square Serivce source. + """A copy activity Square Service source. All required parameters must be populated in order to send to Azure. diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_source_py3.py index a87b924949de..f7ba625398af 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_source_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_source_py3.py @@ -13,7 +13,7 @@ class SquareSource(CopySource): - """A copy activity Square Serivce source. + """A copy activity Square Service source. All required parameters must be populated in order to send to Azure. diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/user_property.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/user_property.py index b29aa61a0e6e..30692d2960ec 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/user_property.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/user_property.py @@ -17,9 +17,9 @@ class UserProperty(Model): All required parameters must be populated in order to send to Azure. - :param name: Required. User proprety name. + :param name: Required. User property name. :type name: str - :param value: Required. User proprety value. Type: string (or Expression + :param value: Required. User property value. Type: string (or Expression with resultType string). :type value: object """ diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/user_property_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/user_property_py3.py index 52691ac1d4e4..7b4f3beb0195 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/user_property_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/user_property_py3.py @@ -17,9 +17,9 @@ class UserProperty(Model): All required parameters must be populated in order to send to Azure. - :param name: Required. User proprety name. + :param name: Required. User property name. :type name: str - :param value: Required. User proprety value. Type: string (or Expression + :param value: Required. User property value. Type: string (or Expression with resultType string). :type value: object """ diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_object_dataset.py index 58f2ab42fe1f..8623618a99d8 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_object_dataset.py @@ -13,7 +13,7 @@ class XeroObjectDataset(Dataset): - """Xero Serivce dataset. + """Xero Service dataset. All required parameters must be populated in order to send to Azure. diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_object_dataset_py3.py index 2b1f6d82814a..b23685dda4ca 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_object_dataset_py3.py @@ -13,7 +13,7 @@ class XeroObjectDataset(Dataset): - """Xero Serivce dataset. + """Xero Service dataset. All required parameters must be populated in order to send to Azure. diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_source.py index c9780ecf4d67..4695780bf41b 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_source.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_source.py @@ -13,7 +13,7 @@ class XeroSource(CopySource): - """A copy activity Xero Serivce source. + """A copy activity Xero Service source. All required parameters must be populated in order to send to Azure. diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_source_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_source_py3.py index e8c8e954eda6..8de950856bae 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_source_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_source_py3.py @@ -13,7 +13,7 @@ class XeroSource(CopySource): - """A copy activity Xero Serivce source. + """A copy activity Xero Service source. All required parameters must be populated in order to send to Azure. diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/integration_runtimes_operations.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/integration_runtimes_operations.py index b61c8c4e5715..0a64be3b1441 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/integration_runtimes_operations.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/integration_runtimes_operations.py @@ -983,7 +983,8 @@ def get_monitoring_data( def upgrade( self, resource_group_name, factory_name, integration_runtime_name, custom_headers=None, raw=False, **operation_config): - """Upgrade self-hosted integration runtime to latest version if availably. + """Upgrade self-hosted integration runtime to latest version if + availability. :param resource_group_name: The resource group name. :type resource_group_name: str diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/pipelines_operations.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/pipelines_operations.py index b00d9a12e180..8a01ce6a8408 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/pipelines_operations.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/pipelines_operations.py @@ -322,8 +322,8 @@ def create_run( :param pipeline_name: The pipeline name. :type pipeline_name: str :param reference_pipeline_run_id: The pipeline run identifier. If run - ID is specified the parameters of the the specified run will be used - to create a new run. + ID is specified the parameters of the specified run will be used to + create a new run. :type reference_pipeline_run_id: str :param parameters: Parameters of the pipeline run. These parameters will be used only if the runId is not specified. From b335e0b78f031e685b586673dd76a1cc3e26d820 Mon Sep 17 00:00:00 2001 From: Azure SDK for Python bot Date: Tue, 4 Dec 2018 12:55:52 -0800 Subject: [PATCH 34/54] Generated from 83b2841a6cd518502da5ea60b22e5dad7557a3b2 (#3931) typo: Microsoft.DataFactory - azure-mgmt-dafactory -> azure-mgmt-datafactory - arbitary -> arbitrary - requries -> requires - Fitler -> Filter --- .../azure/mgmt/datafactory/data_factory_management_client.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/data_factory_management_client.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/data_factory_management_client.py index 864f9f3d0d34..d2919f5bb6ea 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/data_factory_management_client.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/data_factory_management_client.py @@ -53,7 +53,7 @@ def __init__( super(DataFactoryManagementClientConfiguration, self).__init__(base_url) - self.add_user_agent('azure-mgmt-dafactory/{}'.format(VERSION)) + self.add_user_agent('azure-mgmt-datafactory/{}'.format(VERSION)) self.add_user_agent('Azure-SDK-For-Python') self.credentials = credentials From df51f3bef8cfe1187cbe2394783945744e26b784 Mon Sep 17 00:00:00 2001 From: Azure SDK for Python bot Date: Thu, 6 Dec 2018 00:48:41 -0800 Subject: [PATCH 35/54] [AutoPR datafactory/resource-manager] Refactor delete activity payload (#3983) * Generated from 6effad585c2bd2f1de30e2c6e5467750384a3d0b Refactor delete activity content * Generated from fd13d7e45b8d65c0f81299d8ef0e5f977d1f149a Update --- .../azure/mgmt/datafactory/models/__init__.py | 3 ++ .../datafactory/models/delete_activity.py | 24 ++++++++-- .../datafactory/models/delete_activity_py3.py | 26 +++++++++-- .../models/log_storage_settings.py | 46 +++++++++++++++++++ .../models/log_storage_settings_py3.py | 46 +++++++++++++++++++ 5 files changed, 138 insertions(+), 7 deletions(-) create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/log_storage_settings.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/log_storage_settings_py3.py diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py index d185d578748b..8ef32dd63526 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py @@ -297,6 +297,7 @@ from .azure_table_source_py3 import AzureTableSource from .copy_source_py3 import CopySource from .lookup_activity_py3 import LookupActivity + from .log_storage_settings_py3 import LogStorageSettings from .delete_activity_py3 import DeleteActivity from .sql_server_stored_procedure_activity_py3 import SqlServerStoredProcedureActivity from .custom_activity_reference_object_py3 import CustomActivityReferenceObject @@ -652,6 +653,7 @@ from .azure_table_source import AzureTableSource from .copy_source import CopySource from .lookup_activity import LookupActivity + from .log_storage_settings import LogStorageSettings from .delete_activity import DeleteActivity from .sql_server_stored_procedure_activity import SqlServerStoredProcedureActivity from .custom_activity_reference_object import CustomActivityReferenceObject @@ -1077,6 +1079,7 @@ 'AzureTableSource', 'CopySource', 'LookupActivity', + 'LogStorageSettings', 'DeleteActivity', 'SqlServerStoredProcedureActivity', 'CustomActivityReferenceObject', diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/delete_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/delete_activity.py index 62b88a50185f..34ba33a414d5 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/delete_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/delete_activity.py @@ -35,10 +35,21 @@ class DeleteActivity(ExecutionActivity): ~azure.mgmt.datafactory.models.LinkedServiceReference :param policy: Activity policy. :type policy: ~azure.mgmt.datafactory.models.ActivityPolicy - :param recursive: If true, files under the folder path will be deleted - recursively. Default is true. Type: boolean (or Expression with resultType - boolean). + :param recursive: If true, files or sub-folders under current folder path + will be deleted recursively. Default is false. Type: boolean (or + Expression with resultType boolean). :type recursive: object + :param max_concurrent_connections: The max concurrent connections to + connect data source at the same time. + :type max_concurrent_connections: int + :param enable_logging: Whether to record detailed logs of delete-activity + execution. Default value is false. Type: boolean (or Expression with + resultType boolean). + :type enable_logging: object + :param log_storage_settings: Log storage settings customer need to provide + when enableLogging is true. + :type log_storage_settings: + ~azure.mgmt.datafactory.models.LogStorageSettings :param dataset: Required. Delete activity dataset reference. :type dataset: ~azure.mgmt.datafactory.models.DatasetReference """ @@ -46,6 +57,7 @@ class DeleteActivity(ExecutionActivity): _validation = { 'name': {'required': True}, 'type': {'required': True}, + 'max_concurrent_connections': {'minimum': 1}, 'dataset': {'required': True}, } @@ -59,11 +71,17 @@ class DeleteActivity(ExecutionActivity): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, 'recursive': {'key': 'typeProperties.recursive', 'type': 'object'}, + 'max_concurrent_connections': {'key': 'typeProperties.maxConcurrentConnections', 'type': 'int'}, + 'enable_logging': {'key': 'typeProperties.enableLogging', 'type': 'object'}, + 'log_storage_settings': {'key': 'typeProperties.logStorageSettings', 'type': 'LogStorageSettings'}, 'dataset': {'key': 'typeProperties.dataset', 'type': 'DatasetReference'}, } def __init__(self, **kwargs): super(DeleteActivity, self).__init__(**kwargs) self.recursive = kwargs.get('recursive', None) + self.max_concurrent_connections = kwargs.get('max_concurrent_connections', None) + self.enable_logging = kwargs.get('enable_logging', None) + self.log_storage_settings = kwargs.get('log_storage_settings', None) self.dataset = kwargs.get('dataset', None) self.type = 'Delete' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/delete_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/delete_activity_py3.py index f05ce0832ca8..5107d9a3381a 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/delete_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/delete_activity_py3.py @@ -35,10 +35,21 @@ class DeleteActivity(ExecutionActivity): ~azure.mgmt.datafactory.models.LinkedServiceReference :param policy: Activity policy. :type policy: ~azure.mgmt.datafactory.models.ActivityPolicy - :param recursive: If true, files under the folder path will be deleted - recursively. Default is true. Type: boolean (or Expression with resultType - boolean). + :param recursive: If true, files or sub-folders under current folder path + will be deleted recursively. Default is false. Type: boolean (or + Expression with resultType boolean). :type recursive: object + :param max_concurrent_connections: The max concurrent connections to + connect data source at the same time. + :type max_concurrent_connections: int + :param enable_logging: Whether to record detailed logs of delete-activity + execution. Default value is false. Type: boolean (or Expression with + resultType boolean). + :type enable_logging: object + :param log_storage_settings: Log storage settings customer need to provide + when enableLogging is true. + :type log_storage_settings: + ~azure.mgmt.datafactory.models.LogStorageSettings :param dataset: Required. Delete activity dataset reference. :type dataset: ~azure.mgmt.datafactory.models.DatasetReference """ @@ -46,6 +57,7 @@ class DeleteActivity(ExecutionActivity): _validation = { 'name': {'required': True}, 'type': {'required': True}, + 'max_concurrent_connections': {'minimum': 1}, 'dataset': {'required': True}, } @@ -59,11 +71,17 @@ class DeleteActivity(ExecutionActivity): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, 'recursive': {'key': 'typeProperties.recursive', 'type': 'object'}, + 'max_concurrent_connections': {'key': 'typeProperties.maxConcurrentConnections', 'type': 'int'}, + 'enable_logging': {'key': 'typeProperties.enableLogging', 'type': 'object'}, + 'log_storage_settings': {'key': 'typeProperties.logStorageSettings', 'type': 'LogStorageSettings'}, 'dataset': {'key': 'typeProperties.dataset', 'type': 'DatasetReference'}, } - def __init__(self, *, name: str, dataset, additional_properties=None, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, recursive=None, **kwargs) -> None: + def __init__(self, *, name: str, dataset, additional_properties=None, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, recursive=None, max_concurrent_connections: int=None, enable_logging=None, log_storage_settings=None, **kwargs) -> None: super(DeleteActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) self.recursive = recursive + self.max_concurrent_connections = max_concurrent_connections + self.enable_logging = enable_logging + self.log_storage_settings = log_storage_settings self.dataset = dataset self.type = 'Delete' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/log_storage_settings.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/log_storage_settings.py new file mode 100644 index 000000000000..81b4e7ca619e --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/log_storage_settings.py @@ -0,0 +1,46 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class LogStorageSettings(Model): + """Log storage settings. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param linked_service_name: Required. Log storage linked service + reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param path: The path to storage for storing detailed logs of activity + execution. Type: string (or Expression with resultType string). + :type path: object + """ + + _validation = { + 'linked_service_name': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'path': {'key': 'path', 'type': 'object'}, + } + + def __init__(self, **kwargs): + super(LogStorageSettings, self).__init__(**kwargs) + self.additional_properties = kwargs.get('additional_properties', None) + self.linked_service_name = kwargs.get('linked_service_name', None) + self.path = kwargs.get('path', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/log_storage_settings_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/log_storage_settings_py3.py new file mode 100644 index 000000000000..4850b7adacdf --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/log_storage_settings_py3.py @@ -0,0 +1,46 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class LogStorageSettings(Model): + """Log storage settings. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param linked_service_name: Required. Log storage linked service + reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param path: The path to storage for storing detailed logs of activity + execution. Type: string (or Expression with resultType string). + :type path: object + """ + + _validation = { + 'linked_service_name': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'path': {'key': 'path', 'type': 'object'}, + } + + def __init__(self, *, linked_service_name, additional_properties=None, path=None, **kwargs) -> None: + super(LogStorageSettings, self).__init__(**kwargs) + self.additional_properties = additional_properties + self.linked_service_name = linked_service_name + self.path = path From 6a480c46c9ac5ff92bccb05700532b0586965b4b Mon Sep 17 00:00:00 2001 From: Azure SDK for Python bot Date: Thu, 6 Dec 2018 06:51:37 -0800 Subject: [PATCH 36/54] [AutoPR datafactory/resource-manager] Add support for HDIinsight cluster with Enterprise Serurity Package (#4002) * Generated from 7b751f105846a18d948d19efc742af6f514303de Add support for HDIinsight cluster with Enterprise Serurity Package * Generated from 7b751f105846a18d948d19efc742af6f514303de Add support for HDIinsight cluster with Enterprise Serurity Package --- .../datafactory/models/hd_insight_hive_activity.py | 9 +++++++++ .../models/hd_insight_hive_activity_py3.py | 11 ++++++++++- .../datafactory/models/hd_insight_linked_service.py | 5 +++++ .../models/hd_insight_linked_service_py3.py | 7 ++++++- 4 files changed, 30 insertions(+), 2 deletions(-) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_hive_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_hive_activity.py index 4944c4ceff75..d927f1ed0581 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_hive_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_hive_activity.py @@ -52,6 +52,11 @@ class HDInsightHiveActivity(ExecutionActivity): ~azure.mgmt.datafactory.models.LinkedServiceReference :param defines: Allows user to specify defines for Hive job request. :type defines: dict[str, object] + :param variables: User specified arguments under hivevar namespace. + :type variables: list[object] + :param query_timeout: Query timeout value (in minutes). Effective when + the HDInsight culster is with ESP (Enterprise Security Package) + :type query_timeout: int """ _validation = { @@ -74,6 +79,8 @@ class HDInsightHiveActivity(ExecutionActivity): 'script_path': {'key': 'typeProperties.scriptPath', 'type': 'object'}, 'script_linked_service': {'key': 'typeProperties.scriptLinkedService', 'type': 'LinkedServiceReference'}, 'defines': {'key': 'typeProperties.defines', 'type': '{object}'}, + 'variables': {'key': 'typeProperties.variables', 'type': '[object]'}, + 'query_timeout': {'key': 'typeProperties.queryTimeout', 'type': 'int'}, } def __init__(self, **kwargs): @@ -84,4 +91,6 @@ def __init__(self, **kwargs): self.script_path = kwargs.get('script_path', None) self.script_linked_service = kwargs.get('script_linked_service', None) self.defines = kwargs.get('defines', None) + self.variables = kwargs.get('variables', None) + self.query_timeout = kwargs.get('query_timeout', None) self.type = 'HDInsightHive' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_hive_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_hive_activity_py3.py index 748c1d5fc9dc..f62211526203 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_hive_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_hive_activity_py3.py @@ -52,6 +52,11 @@ class HDInsightHiveActivity(ExecutionActivity): ~azure.mgmt.datafactory.models.LinkedServiceReference :param defines: Allows user to specify defines for Hive job request. :type defines: dict[str, object] + :param variables: User specified arguments under hivevar namespace. + :type variables: list[object] + :param query_timeout: Query timeout value (in minutes). Effective when + the HDInsight culster is with ESP (Enterprise Security Package) + :type query_timeout: int """ _validation = { @@ -74,9 +79,11 @@ class HDInsightHiveActivity(ExecutionActivity): 'script_path': {'key': 'typeProperties.scriptPath', 'type': 'object'}, 'script_linked_service': {'key': 'typeProperties.scriptLinkedService', 'type': 'LinkedServiceReference'}, 'defines': {'key': 'typeProperties.defines', 'type': '{object}'}, + 'variables': {'key': 'typeProperties.variables', 'type': '[object]'}, + 'query_timeout': {'key': 'typeProperties.queryTimeout', 'type': 'int'}, } - def __init__(self, *, name: str, additional_properties=None, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, storage_linked_services=None, arguments=None, get_debug_info=None, script_path=None, script_linked_service=None, defines=None, **kwargs) -> None: + def __init__(self, *, name: str, additional_properties=None, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, storage_linked_services=None, arguments=None, get_debug_info=None, script_path=None, script_linked_service=None, defines=None, variables=None, query_timeout: int=None, **kwargs) -> None: super(HDInsightHiveActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) self.storage_linked_services = storage_linked_services self.arguments = arguments @@ -84,4 +91,6 @@ def __init__(self, *, name: str, additional_properties=None, description: str=No self.script_path = script_path self.script_linked_service = script_linked_service self.defines = defines + self.variables = variables + self.query_timeout = query_timeout self.type = 'HDInsightHive' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_linked_service.py index db7e14837305..b18a138a855e 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_linked_service.py @@ -52,6 +52,9 @@ class HDInsightLinkedService(LinkedService): authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). :type encrypted_credential: object + :param is_esp_enabled: Specify if the HDInsight is created with ESP + (Enterprise Security Package). Type: Boolean. + :type is_esp_enabled: object """ _validation = { @@ -72,6 +75,7 @@ class HDInsightLinkedService(LinkedService): 'linked_service_name': {'key': 'typeProperties.linkedServiceName', 'type': 'LinkedServiceReference'}, 'hcatalog_linked_service_name': {'key': 'typeProperties.hcatalogLinkedServiceName', 'type': 'LinkedServiceReference'}, 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + 'is_esp_enabled': {'key': 'typeProperties.isEspEnabled', 'type': 'object'}, } def __init__(self, **kwargs): @@ -82,4 +86,5 @@ def __init__(self, **kwargs): self.linked_service_name = kwargs.get('linked_service_name', None) self.hcatalog_linked_service_name = kwargs.get('hcatalog_linked_service_name', None) self.encrypted_credential = kwargs.get('encrypted_credential', None) + self.is_esp_enabled = kwargs.get('is_esp_enabled', None) self.type = 'HDInsight' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_linked_service_py3.py index e7e60b5c8333..769cf031a403 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_linked_service_py3.py @@ -52,6 +52,9 @@ class HDInsightLinkedService(LinkedService): authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). :type encrypted_credential: object + :param is_esp_enabled: Specify if the HDInsight is created with ESP + (Enterprise Security Package). Type: Boolean. + :type is_esp_enabled: object """ _validation = { @@ -72,9 +75,10 @@ class HDInsightLinkedService(LinkedService): 'linked_service_name': {'key': 'typeProperties.linkedServiceName', 'type': 'LinkedServiceReference'}, 'hcatalog_linked_service_name': {'key': 'typeProperties.hcatalogLinkedServiceName', 'type': 'LinkedServiceReference'}, 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + 'is_esp_enabled': {'key': 'typeProperties.isEspEnabled', 'type': 'object'}, } - def __init__(self, *, cluster_uri, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, user_name=None, password=None, linked_service_name=None, hcatalog_linked_service_name=None, encrypted_credential=None, **kwargs) -> None: + def __init__(self, *, cluster_uri, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, user_name=None, password=None, linked_service_name=None, hcatalog_linked_service_name=None, encrypted_credential=None, is_esp_enabled=None, **kwargs) -> None: super(HDInsightLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) self.cluster_uri = cluster_uri self.user_name = user_name @@ -82,4 +86,5 @@ def __init__(self, *, cluster_uri, additional_properties=None, connect_via=None, self.linked_service_name = linked_service_name self.hcatalog_linked_service_name = hcatalog_linked_service_name self.encrypted_credential = encrypted_credential + self.is_esp_enabled = is_esp_enabled self.type = 'HDInsight' From 6c4199726e8fcd57d92bc6abb3fb1d17ff2159c9 Mon Sep 17 00:00:00 2001 From: Azure SDK for Python bot Date: Mon, 10 Dec 2018 19:55:18 -0800 Subject: [PATCH 37/54] [AutoPR datafactory/resource-manager] Adding refresh and get IntegrationRuntimeObjectMetadata API (#3962) * Generated from aded924bcbec0926d1f273fdc99b3a8d2d51246b Adding refresh and get IntegrationRuntimeObjectMetadata API * Generated from 38fb817c3d29a206d1371a3d4cea5002cc960535 Change ssis to upper case * Generated from 43c2dab88ba9dd4abfe793fddfe0c1156518c4a9 Fix model validation issue --- .../data_factory_management_client.py | 5 + .../azure/mgmt/datafactory/models/__init__.py | 14 ++ .../data_factory_management_client_enums.py | 8 + .../get_ssis_object_metadata_request.py | 28 +++ .../get_ssis_object_metadata_request_py3.py | 28 +++ .../models/ssis_object_metadata.py | 46 ++++ .../ssis_object_metadata_list_response.py | 33 +++ .../ssis_object_metadata_list_response_py3.py | 33 +++ .../models/ssis_object_metadata_py3.py | 46 ++++ .../ssis_object_metadata_status_response.py | 40 ++++ ...sis_object_metadata_status_response_py3.py | 40 ++++ .../mgmt/datafactory/operations/__init__.py | 2 + ...tion_runtime_object_metadata_operations.py | 218 ++++++++++++++++++ 13 files changed, 541 insertions(+) create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/get_ssis_object_metadata_request.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/get_ssis_object_metadata_request_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/ssis_object_metadata.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/ssis_object_metadata_list_response.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/ssis_object_metadata_list_response_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/ssis_object_metadata_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/ssis_object_metadata_status_response.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/ssis_object_metadata_status_response_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/operations/integration_runtime_object_metadata_operations.py diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/data_factory_management_client.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/data_factory_management_client.py index d2919f5bb6ea..bfb43b111214 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/data_factory_management_client.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/data_factory_management_client.py @@ -16,6 +16,7 @@ from .operations.operations import Operations from .operations.factories_operations import FactoriesOperations from .operations.integration_runtimes_operations import IntegrationRuntimesOperations +from .operations.integration_runtime_object_metadata_operations import IntegrationRuntimeObjectMetadataOperations from .operations.integration_runtime_nodes_operations import IntegrationRuntimeNodesOperations from .operations.linked_services_operations import LinkedServicesOperations from .operations.datasets_operations import DatasetsOperations @@ -72,6 +73,8 @@ class DataFactoryManagementClient(SDKClient): :vartype factories: azure.mgmt.datafactory.operations.FactoriesOperations :ivar integration_runtimes: IntegrationRuntimes operations :vartype integration_runtimes: azure.mgmt.datafactory.operations.IntegrationRuntimesOperations + :ivar integration_runtime_object_metadata: IntegrationRuntimeObjectMetadata operations + :vartype integration_runtime_object_metadata: azure.mgmt.datafactory.operations.IntegrationRuntimeObjectMetadataOperations :ivar integration_runtime_nodes: IntegrationRuntimeNodes operations :vartype integration_runtime_nodes: azure.mgmt.datafactory.operations.IntegrationRuntimeNodesOperations :ivar linked_services: LinkedServices operations @@ -116,6 +119,8 @@ def __init__( self._client, self.config, self._serialize, self._deserialize) self.integration_runtimes = IntegrationRuntimesOperations( self._client, self.config, self._serialize, self._deserialize) + self.integration_runtime_object_metadata = IntegrationRuntimeObjectMetadataOperations( + self._client, self.config, self._serialize, self._deserialize) self.integration_runtime_nodes = IntegrationRuntimeNodesOperations( self._client, self.config, self._serialize, self._deserialize) self.linked_services = LinkedServicesOperations( diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py index 8ef32dd63526..72e1c1a551b6 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py @@ -76,6 +76,8 @@ from .operation_metric_specification_py3 import OperationMetricSpecification from .operation_service_specification_py3 import OperationServiceSpecification from .operation_py3 import Operation + from .get_ssis_object_metadata_request_py3 import GetSsisObjectMetadataRequest + from .ssis_object_metadata_status_response_py3 import SsisObjectMetadataStatusResponse from .self_dependency_tumbling_window_trigger_reference_py3 import SelfDependencyTumblingWindowTriggerReference from .trigger_reference_py3 import TriggerReference from .tumbling_window_trigger_dependency_reference_py3 import TumblingWindowTriggerDependencyReference @@ -360,6 +362,8 @@ from .integration_runtime_compute_properties_py3 import IntegrationRuntimeComputeProperties from .managed_integration_runtime_py3 import ManagedIntegrationRuntime from .integration_runtime_node_ip_address_py3 import IntegrationRuntimeNodeIpAddress + from .ssis_object_metadata_py3 import SsisObjectMetadata + from .ssis_object_metadata_list_response_py3 import SsisObjectMetadataListResponse from .integration_runtime_node_monitoring_data_py3 import IntegrationRuntimeNodeMonitoringData from .integration_runtime_monitoring_data_py3 import IntegrationRuntimeMonitoringData from .integration_runtime_auth_keys_py3 import IntegrationRuntimeAuthKeys @@ -432,6 +436,8 @@ from .operation_metric_specification import OperationMetricSpecification from .operation_service_specification import OperationServiceSpecification from .operation import Operation + from .get_ssis_object_metadata_request import GetSsisObjectMetadataRequest + from .ssis_object_metadata_status_response import SsisObjectMetadataStatusResponse from .self_dependency_tumbling_window_trigger_reference import SelfDependencyTumblingWindowTriggerReference from .trigger_reference import TriggerReference from .tumbling_window_trigger_dependency_reference import TumblingWindowTriggerDependencyReference @@ -716,6 +722,8 @@ from .integration_runtime_compute_properties import IntegrationRuntimeComputeProperties from .managed_integration_runtime import ManagedIntegrationRuntime from .integration_runtime_node_ip_address import IntegrationRuntimeNodeIpAddress + from .ssis_object_metadata import SsisObjectMetadata + from .ssis_object_metadata_list_response import SsisObjectMetadataListResponse from .integration_runtime_node_monitoring_data import IntegrationRuntimeNodeMonitoringData from .integration_runtime_monitoring_data import IntegrationRuntimeMonitoringData from .integration_runtime_auth_keys import IntegrationRuntimeAuthKeys @@ -788,6 +796,7 @@ IntegrationRuntimeSsisCatalogPricingTier, IntegrationRuntimeLicenseType, IntegrationRuntimeEdition, + SsisObjectMetadataType, IntegrationRuntimeAuthKeyName, ) @@ -858,6 +867,8 @@ 'OperationMetricSpecification', 'OperationServiceSpecification', 'Operation', + 'GetSsisObjectMetadataRequest', + 'SsisObjectMetadataStatusResponse', 'SelfDependencyTumblingWindowTriggerReference', 'TriggerReference', 'TumblingWindowTriggerDependencyReference', @@ -1142,6 +1153,8 @@ 'IntegrationRuntimeComputeProperties', 'ManagedIntegrationRuntime', 'IntegrationRuntimeNodeIpAddress', + 'SsisObjectMetadata', + 'SsisObjectMetadataListResponse', 'IntegrationRuntimeNodeMonitoringData', 'IntegrationRuntimeMonitoringData', 'IntegrationRuntimeAuthKeys', @@ -1213,5 +1226,6 @@ 'IntegrationRuntimeSsisCatalogPricingTier', 'IntegrationRuntimeLicenseType', 'IntegrationRuntimeEdition', + 'SsisObjectMetadataType', 'IntegrationRuntimeAuthKeyName', ] diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_factory_management_client_enums.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_factory_management_client_enums.py index aa76fe05ddc0..1e4c04780d32 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_factory_management_client_enums.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_factory_management_client_enums.py @@ -449,6 +449,14 @@ class IntegrationRuntimeEdition(str, Enum): enterprise = "Enterprise" +class SsisObjectMetadataType(str, Enum): + + folder = "Folder" + project = "Project" + package = "Package" + environment = "Environment" + + class IntegrationRuntimeAuthKeyName(str, Enum): auth_key1 = "authKey1" diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/get_ssis_object_metadata_request.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/get_ssis_object_metadata_request.py new file mode 100644 index 000000000000..1be4a2afece0 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/get_ssis_object_metadata_request.py @@ -0,0 +1,28 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class GetSsisObjectMetadataRequest(Model): + """The request payload of get SSIS object metadata. + + :param metadata_path: Metadata path. + :type metadata_path: str + """ + + _attribute_map = { + 'metadata_path': {'key': 'metadataPath', 'type': 'str'}, + } + + def __init__(self, **kwargs): + super(GetSsisObjectMetadataRequest, self).__init__(**kwargs) + self.metadata_path = kwargs.get('metadata_path', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/get_ssis_object_metadata_request_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/get_ssis_object_metadata_request_py3.py new file mode 100644 index 000000000000..310cd9783d81 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/get_ssis_object_metadata_request_py3.py @@ -0,0 +1,28 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class GetSsisObjectMetadataRequest(Model): + """The request payload of get SSIS object metadata. + + :param metadata_path: Metadata path. + :type metadata_path: str + """ + + _attribute_map = { + 'metadata_path': {'key': 'metadataPath', 'type': 'str'}, + } + + def __init__(self, *, metadata_path: str=None, **kwargs) -> None: + super(GetSsisObjectMetadataRequest, self).__init__(**kwargs) + self.metadata_path = metadata_path diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/ssis_object_metadata.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/ssis_object_metadata.py new file mode 100644 index 000000000000..ed7940124645 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/ssis_object_metadata.py @@ -0,0 +1,46 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class SsisObjectMetadata(Model): + """SSIS object metadata. + + All required parameters must be populated in order to send to Azure. + + :param id: Metadata id. + :type id: long + :param name: Metadata name. + :type name: str + :param description: Metadata description. + :type description: str + :param type: Required. Constant filled by server. + :type type: str + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'id': {'key': 'id', 'type': 'long'}, + 'name': {'key': 'name', 'type': 'str'}, + 'description': {'key': 'description', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, **kwargs): + super(SsisObjectMetadata, self).__init__(**kwargs) + self.id = kwargs.get('id', None) + self.name = kwargs.get('name', None) + self.description = kwargs.get('description', None) + self.type = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/ssis_object_metadata_list_response.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/ssis_object_metadata_list_response.py new file mode 100644 index 000000000000..a029c9f7ebc4 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/ssis_object_metadata_list_response.py @@ -0,0 +1,33 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class SsisObjectMetadataListResponse(Model): + """A list of SSIS object metadata. + + :param value: List of SSIS object metadata. + :type value: list[~azure.mgmt.datafactory.models.SsisObjectMetadata] + :param next_link: The link to the next page of results, if any remaining + results exist. + :type next_link: str + """ + + _attribute_map = { + 'value': {'key': 'value', 'type': '[SsisObjectMetadata]'}, + 'next_link': {'key': 'nextLink', 'type': 'str'}, + } + + def __init__(self, **kwargs): + super(SsisObjectMetadataListResponse, self).__init__(**kwargs) + self.value = kwargs.get('value', None) + self.next_link = kwargs.get('next_link', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/ssis_object_metadata_list_response_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/ssis_object_metadata_list_response_py3.py new file mode 100644 index 000000000000..79931e1ceaf7 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/ssis_object_metadata_list_response_py3.py @@ -0,0 +1,33 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class SsisObjectMetadataListResponse(Model): + """A list of SSIS object metadata. + + :param value: List of SSIS object metadata. + :type value: list[~azure.mgmt.datafactory.models.SsisObjectMetadata] + :param next_link: The link to the next page of results, if any remaining + results exist. + :type next_link: str + """ + + _attribute_map = { + 'value': {'key': 'value', 'type': '[SsisObjectMetadata]'}, + 'next_link': {'key': 'nextLink', 'type': 'str'}, + } + + def __init__(self, *, value=None, next_link: str=None, **kwargs) -> None: + super(SsisObjectMetadataListResponse, self).__init__(**kwargs) + self.value = value + self.next_link = next_link diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/ssis_object_metadata_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/ssis_object_metadata_py3.py new file mode 100644 index 000000000000..b7373e36523c --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/ssis_object_metadata_py3.py @@ -0,0 +1,46 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class SsisObjectMetadata(Model): + """SSIS object metadata. + + All required parameters must be populated in order to send to Azure. + + :param id: Metadata id. + :type id: long + :param name: Metadata name. + :type name: str + :param description: Metadata description. + :type description: str + :param type: Required. Constant filled by server. + :type type: str + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'id': {'key': 'id', 'type': 'long'}, + 'name': {'key': 'name', 'type': 'str'}, + 'description': {'key': 'description', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__(self, *, id: int=None, name: str=None, description: str=None, **kwargs) -> None: + super(SsisObjectMetadata, self).__init__(**kwargs) + self.id = id + self.name = name + self.description = description + self.type = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/ssis_object_metadata_status_response.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/ssis_object_metadata_status_response.py new file mode 100644 index 000000000000..9b782613ee08 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/ssis_object_metadata_status_response.py @@ -0,0 +1,40 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class SsisObjectMetadataStatusResponse(Model): + """The status of the operation. + + :param status: The status of the operation. + :type status: str + :param name: The operation name. + :type name: str + :param properties: The operation properties. + :type properties: str + :param error: The operation error message. + :type error: str + """ + + _attribute_map = { + 'status': {'key': 'status', 'type': 'str'}, + 'name': {'key': 'name', 'type': 'str'}, + 'properties': {'key': 'properties', 'type': 'str'}, + 'error': {'key': 'error', 'type': 'str'}, + } + + def __init__(self, **kwargs): + super(SsisObjectMetadataStatusResponse, self).__init__(**kwargs) + self.status = kwargs.get('status', None) + self.name = kwargs.get('name', None) + self.properties = kwargs.get('properties', None) + self.error = kwargs.get('error', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/ssis_object_metadata_status_response_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/ssis_object_metadata_status_response_py3.py new file mode 100644 index 000000000000..a4b82b8f6bcd --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/ssis_object_metadata_status_response_py3.py @@ -0,0 +1,40 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class SsisObjectMetadataStatusResponse(Model): + """The status of the operation. + + :param status: The status of the operation. + :type status: str + :param name: The operation name. + :type name: str + :param properties: The operation properties. + :type properties: str + :param error: The operation error message. + :type error: str + """ + + _attribute_map = { + 'status': {'key': 'status', 'type': 'str'}, + 'name': {'key': 'name', 'type': 'str'}, + 'properties': {'key': 'properties', 'type': 'str'}, + 'error': {'key': 'error', 'type': 'str'}, + } + + def __init__(self, *, status: str=None, name: str=None, properties: str=None, error: str=None, **kwargs) -> None: + super(SsisObjectMetadataStatusResponse, self).__init__(**kwargs) + self.status = status + self.name = name + self.properties = properties + self.error = error diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/__init__.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/__init__.py index a987d7ddfec7..c0dccd86b56a 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/__init__.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/__init__.py @@ -12,6 +12,7 @@ from .operations import Operations from .factories_operations import FactoriesOperations from .integration_runtimes_operations import IntegrationRuntimesOperations +from .integration_runtime_object_metadata_operations import IntegrationRuntimeObjectMetadataOperations from .integration_runtime_nodes_operations import IntegrationRuntimeNodesOperations from .linked_services_operations import LinkedServicesOperations from .datasets_operations import DatasetsOperations @@ -26,6 +27,7 @@ 'Operations', 'FactoriesOperations', 'IntegrationRuntimesOperations', + 'IntegrationRuntimeObjectMetadataOperations', 'IntegrationRuntimeNodesOperations', 'LinkedServicesOperations', 'DatasetsOperations', diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/integration_runtime_object_metadata_operations.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/integration_runtime_object_metadata_operations.py new file mode 100644 index 000000000000..230f12d023c3 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/integration_runtime_object_metadata_operations.py @@ -0,0 +1,218 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +import uuid +from msrest.pipeline import ClientRawResponse +from msrestazure.azure_exceptions import CloudError +from msrest.polling import LROPoller, NoPolling +from msrestazure.polling.arm_polling import ARMPolling + +from .. import models + + +class IntegrationRuntimeObjectMetadataOperations(object): + """IntegrationRuntimeObjectMetadataOperations operations. + + :param client: Client for service requests. + :param config: Configuration of service client. + :param serializer: An object model serializer. + :param deserializer: An object model deserializer. + :ivar api_version: The API version. Constant value: "2018-06-01". + """ + + models = models + + def __init__(self, client, config, serializer, deserializer): + + self._client = client + self._serialize = serializer + self._deserialize = deserializer + self.api_version = "2018-06-01" + + self.config = config + + + def _refresh_initial( + self, resource_group_name, factory_name, integration_runtime_name, custom_headers=None, raw=False, **operation_config): + # Construct URL + url = self.refresh.metadata['url'] + path_format_arguments = { + 'subscriptionId': self._serialize.url("self.config.subscription_id", self.config.subscription_id, 'str'), + 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str', max_length=90, min_length=1, pattern=r'^[-\w\._\(\)]+$'), + 'factoryName': self._serialize.url("factory_name", factory_name, 'str', max_length=63, min_length=3, pattern=r'^[A-Za-z0-9]+(?:-[A-Za-z0-9]+)*$'), + 'integrationRuntimeName': self._serialize.url("integration_runtime_name", integration_runtime_name, 'str', max_length=63, min_length=3, pattern=r'^[A-Za-z0-9]+(?:-[A-Za-z0-9]+)*$') + } + url = self._client.format_url(url, **path_format_arguments) + + # Construct parameters + query_parameters = {} + query_parameters['api-version'] = self._serialize.query("self.api_version", self.api_version, 'str') + + # Construct headers + header_parameters = {} + header_parameters['Accept'] = 'application/json' + if self.config.generate_client_request_id: + header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) + if custom_headers: + header_parameters.update(custom_headers) + if self.config.accept_language is not None: + header_parameters['accept-language'] = self._serialize.header("self.config.accept_language", self.config.accept_language, 'str') + + # Construct and send request + request = self._client.post(url, query_parameters, header_parameters) + response = self._client.send(request, stream=False, **operation_config) + + if response.status_code not in [200, 202]: + exp = CloudError(response) + exp.request_id = response.headers.get('x-ms-request-id') + raise exp + + deserialized = None + + if response.status_code == 200: + deserialized = self._deserialize('SsisObjectMetadataStatusResponse', response) + + if raw: + client_raw_response = ClientRawResponse(deserialized, response) + return client_raw_response + + return deserialized + + def refresh( + self, resource_group_name, factory_name, integration_runtime_name, custom_headers=None, raw=False, polling=True, **operation_config): + """Refresh a SSIS integration runtime object metadata. + + :param resource_group_name: The resource group name. + :type resource_group_name: str + :param factory_name: The factory name. + :type factory_name: str + :param integration_runtime_name: The integration runtime name. + :type integration_runtime_name: str + :param dict custom_headers: headers that will be added to the request + :param bool raw: The poller return type is ClientRawResponse, the + direct response alongside the deserialized response + :param polling: True for ARMPolling, False for no polling, or a + polling object for personal polling strategy + :return: An instance of LROPoller that returns + SsisObjectMetadataStatusResponse or + ClientRawResponse if raw==True + :rtype: + ~msrestazure.azure_operation.AzureOperationPoller[~azure.mgmt.datafactory.models.SsisObjectMetadataStatusResponse] + or + ~msrestazure.azure_operation.AzureOperationPoller[~msrest.pipeline.ClientRawResponse[~azure.mgmt.datafactory.models.SsisObjectMetadataStatusResponse]] + :raises: :class:`CloudError` + """ + raw_result = self._refresh_initial( + resource_group_name=resource_group_name, + factory_name=factory_name, + integration_runtime_name=integration_runtime_name, + custom_headers=custom_headers, + raw=True, + **operation_config + ) + + def get_long_running_output(response): + deserialized = self._deserialize('SsisObjectMetadataStatusResponse', response) + + if raw: + client_raw_response = ClientRawResponse(deserialized, response) + return client_raw_response + + return deserialized + + lro_delay = operation_config.get( + 'long_running_operation_timeout', + self.config.long_running_operation_timeout) + if polling is True: polling_method = ARMPolling(lro_delay, **operation_config) + elif polling is False: polling_method = NoPolling() + else: polling_method = polling + return LROPoller(self._client, raw_result, get_long_running_output, polling_method) + refresh.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/integrationRuntimes/{integrationRuntimeName}/refreshObjectMetadata'} + + def get( + self, resource_group_name, factory_name, integration_runtime_name, metadata_path=None, custom_headers=None, raw=False, **operation_config): + """Get a SSIS integration runtime object metadata by specified path. The + return is pageable metadata list. + + :param resource_group_name: The resource group name. + :type resource_group_name: str + :param factory_name: The factory name. + :type factory_name: str + :param integration_runtime_name: The integration runtime name. + :type integration_runtime_name: str + :param metadata_path: Metadata path. + :type metadata_path: str + :param dict custom_headers: headers that will be added to the request + :param bool raw: returns the direct response alongside the + deserialized response + :param operation_config: :ref:`Operation configuration + overrides`. + :return: SsisObjectMetadataListResponse or ClientRawResponse if + raw=true + :rtype: ~azure.mgmt.datafactory.models.SsisObjectMetadataListResponse + or ~msrest.pipeline.ClientRawResponse + :raises: :class:`CloudError` + """ + get_metadata_request = None + if metadata_path is not None: + get_metadata_request = models.GetSsisObjectMetadataRequest(metadata_path=metadata_path) + + # Construct URL + url = self.get.metadata['url'] + path_format_arguments = { + 'subscriptionId': self._serialize.url("self.config.subscription_id", self.config.subscription_id, 'str'), + 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str', max_length=90, min_length=1, pattern=r'^[-\w\._\(\)]+$'), + 'factoryName': self._serialize.url("factory_name", factory_name, 'str', max_length=63, min_length=3, pattern=r'^[A-Za-z0-9]+(?:-[A-Za-z0-9]+)*$'), + 'integrationRuntimeName': self._serialize.url("integration_runtime_name", integration_runtime_name, 'str', max_length=63, min_length=3, pattern=r'^[A-Za-z0-9]+(?:-[A-Za-z0-9]+)*$') + } + url = self._client.format_url(url, **path_format_arguments) + + # Construct parameters + query_parameters = {} + query_parameters['api-version'] = self._serialize.query("self.api_version", self.api_version, 'str') + + # Construct headers + header_parameters = {} + header_parameters['Accept'] = 'application/json' + header_parameters['Content-Type'] = 'application/json; charset=utf-8' + if self.config.generate_client_request_id: + header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) + if custom_headers: + header_parameters.update(custom_headers) + if self.config.accept_language is not None: + header_parameters['accept-language'] = self._serialize.header("self.config.accept_language", self.config.accept_language, 'str') + + # Construct body + if get_metadata_request is not None: + body_content = self._serialize.body(get_metadata_request, 'GetSsisObjectMetadataRequest') + else: + body_content = None + + # Construct and send request + request = self._client.post(url, query_parameters, header_parameters, body_content) + response = self._client.send(request, stream=False, **operation_config) + + if response.status_code not in [200]: + exp = CloudError(response) + exp.request_id = response.headers.get('x-ms-request-id') + raise exp + + deserialized = None + + if response.status_code == 200: + deserialized = self._deserialize('SsisObjectMetadataListResponse', response) + + if raw: + client_raw_response = ClientRawResponse(deserialized, response) + return client_raw_response + + return deserialized + get.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/integrationRuntimes/{integrationRuntimeName}/getObjectMetadata'} From 991661944f57bb98117df5a7045b5ff6e6b591a5 Mon Sep 17 00:00:00 2001 From: Azure SDK for Python bot Date: Mon, 17 Dec 2018 09:10:43 -0800 Subject: [PATCH 38/54] Generated from 948350f5d28ef4686eaa2a4ffead98655f0262cb (#4077) typo: datafactory/resource-manager/Microsoft.DataFactory - culster -> cluster --- .../azure/mgmt/datafactory/models/hd_insight_hive_activity.py | 2 +- .../mgmt/datafactory/models/hd_insight_hive_activity_py3.py | 2 +- 2 files changed, 2 insertions(+), 2 deletions(-) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_hive_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_hive_activity.py index d927f1ed0581..4110b0f8b7de 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_hive_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_hive_activity.py @@ -55,7 +55,7 @@ class HDInsightHiveActivity(ExecutionActivity): :param variables: User specified arguments under hivevar namespace. :type variables: list[object] :param query_timeout: Query timeout value (in minutes). Effective when - the HDInsight culster is with ESP (Enterprise Security Package) + the HDInsight cluster is with ESP (Enterprise Security Package) :type query_timeout: int """ diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_hive_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_hive_activity_py3.py index f62211526203..f8a5441fe767 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_hive_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hd_insight_hive_activity_py3.py @@ -55,7 +55,7 @@ class HDInsightHiveActivity(ExecutionActivity): :param variables: User specified arguments under hivevar namespace. :type variables: list[object] :param query_timeout: Query timeout value (in minutes). Effective when - the HDInsight culster is with ESP (Enterprise Security Package) + the HDInsight cluster is with ESP (Enterprise Security Package) :type query_timeout: int """ From b33f97b288f6bb8ee8f2c2e7ca3d95636705854f Mon Sep 17 00:00:00 2001 From: Azure SDK for Python bot Date: Thu, 3 Jan 2019 15:39:00 -0800 Subject: [PATCH 39/54] Generated from f9be2a6b0020602b14ff1e64cfb47c38f096fffa (#4090) Change type of Ssis actitvity's packagePath, runtime, loggingLevel and environmentPath from string to object, so that we can use expression for these properties --- .../azure/mgmt/datafactory/models/__init__.py | 2 -- .../data_factory_management_client_enums.py | 6 ----- .../models/execute_ssis_package_activity.py | 21 +++++++++-------- .../execute_ssis_package_activity_py3.py | 23 +++++++++++-------- .../models/ssis_package_location.py | 7 +++--- .../models/ssis_package_location_py3.py | 9 ++++---- 6 files changed, 34 insertions(+), 34 deletions(-) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py index 72e1c1a551b6..c9728054c80a 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py @@ -781,7 +781,6 @@ CassandraSourceReadConsistencyLevels, StoredProcedureParameterType, SalesforceSourceReadBehavior, - SSISExecutionRuntime, HDInsightActivityDebugInfoOption, SalesforceSinkWriteBehavior, AzureSearchIndexWriteBehaviorType, @@ -1211,7 +1210,6 @@ 'CassandraSourceReadConsistencyLevels', 'StoredProcedureParameterType', 'SalesforceSourceReadBehavior', - 'SSISExecutionRuntime', 'HDInsightActivityDebugInfoOption', 'SalesforceSinkWriteBehavior', 'AzureSearchIndexWriteBehaviorType', diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_factory_management_client_enums.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_factory_management_client_enums.py index 1e4c04780d32..23334492e4d2 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_factory_management_client_enums.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_factory_management_client_enums.py @@ -346,12 +346,6 @@ class SalesforceSourceReadBehavior(str, Enum): query_all = "QueryAll" -class SSISExecutionRuntime(str, Enum): - - x64 = "x64" - x86 = "x86" - - class HDInsightActivityDebugInfoOption(str, Enum): none = "None" diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_ssis_package_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_ssis_package_activity.py index 3491cbe93982..3e5dc62320cb 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_ssis_package_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_ssis_package_activity.py @@ -37,13 +37,16 @@ class ExecuteSSISPackageActivity(ExecutionActivity): :type policy: ~azure.mgmt.datafactory.models.ActivityPolicy :param package_location: Required. SSIS package location. :type package_location: ~azure.mgmt.datafactory.models.SSISPackageLocation - :param runtime: Specifies the runtime to execute SSIS package. Possible - values include: 'x64', 'x86' - :type runtime: str or ~azure.mgmt.datafactory.models.SSISExecutionRuntime - :param logging_level: The logging level of SSIS package execution. - :type logging_level: str + :param runtime: Specifies the runtime to execute SSIS package. The value + should be "x86" or "x64". Type: string (or Expression with resultType + string). + :type runtime: object + :param logging_level: The logging level of SSIS package execution. Type: + string (or Expression with resultType string). + :type logging_level: object :param environment_path: The environment path to execute the SSIS package. - :type environment_path: str + Type: string (or Expression with resultType string). + :type environment_path: object :param connect_via: Required. The integration runtime reference. :type connect_via: ~azure.mgmt.datafactory.models.IntegrationRuntimeReference @@ -86,9 +89,9 @@ class ExecuteSSISPackageActivity(ExecutionActivity): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, 'package_location': {'key': 'typeProperties.packageLocation', 'type': 'SSISPackageLocation'}, - 'runtime': {'key': 'typeProperties.runtime', 'type': 'str'}, - 'logging_level': {'key': 'typeProperties.loggingLevel', 'type': 'str'}, - 'environment_path': {'key': 'typeProperties.environmentPath', 'type': 'str'}, + 'runtime': {'key': 'typeProperties.runtime', 'type': 'object'}, + 'logging_level': {'key': 'typeProperties.loggingLevel', 'type': 'object'}, + 'environment_path': {'key': 'typeProperties.environmentPath', 'type': 'object'}, 'connect_via': {'key': 'typeProperties.connectVia', 'type': 'IntegrationRuntimeReference'}, 'project_parameters': {'key': 'typeProperties.projectParameters', 'type': '{SSISExecutionParameter}'}, 'package_parameters': {'key': 'typeProperties.packageParameters', 'type': '{SSISExecutionParameter}'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_ssis_package_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_ssis_package_activity_py3.py index b00419f52c55..0a716e8c9482 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_ssis_package_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_ssis_package_activity_py3.py @@ -37,13 +37,16 @@ class ExecuteSSISPackageActivity(ExecutionActivity): :type policy: ~azure.mgmt.datafactory.models.ActivityPolicy :param package_location: Required. SSIS package location. :type package_location: ~azure.mgmt.datafactory.models.SSISPackageLocation - :param runtime: Specifies the runtime to execute SSIS package. Possible - values include: 'x64', 'x86' - :type runtime: str or ~azure.mgmt.datafactory.models.SSISExecutionRuntime - :param logging_level: The logging level of SSIS package execution. - :type logging_level: str + :param runtime: Specifies the runtime to execute SSIS package. The value + should be "x86" or "x64". Type: string (or Expression with resultType + string). + :type runtime: object + :param logging_level: The logging level of SSIS package execution. Type: + string (or Expression with resultType string). + :type logging_level: object :param environment_path: The environment path to execute the SSIS package. - :type environment_path: str + Type: string (or Expression with resultType string). + :type environment_path: object :param connect_via: Required. The integration runtime reference. :type connect_via: ~azure.mgmt.datafactory.models.IntegrationRuntimeReference @@ -86,9 +89,9 @@ class ExecuteSSISPackageActivity(ExecutionActivity): 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, 'package_location': {'key': 'typeProperties.packageLocation', 'type': 'SSISPackageLocation'}, - 'runtime': {'key': 'typeProperties.runtime', 'type': 'str'}, - 'logging_level': {'key': 'typeProperties.loggingLevel', 'type': 'str'}, - 'environment_path': {'key': 'typeProperties.environmentPath', 'type': 'str'}, + 'runtime': {'key': 'typeProperties.runtime', 'type': 'object'}, + 'logging_level': {'key': 'typeProperties.loggingLevel', 'type': 'object'}, + 'environment_path': {'key': 'typeProperties.environmentPath', 'type': 'object'}, 'connect_via': {'key': 'typeProperties.connectVia', 'type': 'IntegrationRuntimeReference'}, 'project_parameters': {'key': 'typeProperties.projectParameters', 'type': '{SSISExecutionParameter}'}, 'package_parameters': {'key': 'typeProperties.packageParameters', 'type': '{SSISExecutionParameter}'}, @@ -97,7 +100,7 @@ class ExecuteSSISPackageActivity(ExecutionActivity): 'property_overrides': {'key': 'typeProperties.propertyOverrides', 'type': '{SSISPropertyOverride}'}, } - def __init__(self, *, name: str, package_location, connect_via, additional_properties=None, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, runtime=None, logging_level: str=None, environment_path: str=None, project_parameters=None, package_parameters=None, project_connection_managers=None, package_connection_managers=None, property_overrides=None, **kwargs) -> None: + def __init__(self, *, name: str, package_location, connect_via, additional_properties=None, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, runtime=None, logging_level=None, environment_path=None, project_parameters=None, package_parameters=None, project_connection_managers=None, package_connection_managers=None, property_overrides=None, **kwargs) -> None: super(ExecuteSSISPackageActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) self.package_location = package_location self.runtime = runtime diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/ssis_package_location.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/ssis_package_location.py index f81a9fafec37..81a17eb8fe53 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/ssis_package_location.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/ssis_package_location.py @@ -17,8 +17,9 @@ class SSISPackageLocation(Model): All required parameters must be populated in order to send to Azure. - :param package_path: Required. The SSIS package path. - :type package_path: str + :param package_path: Required. The SSIS package path. Type: string (or + Expression with resultType string). + :type package_path: object """ _validation = { @@ -26,7 +27,7 @@ class SSISPackageLocation(Model): } _attribute_map = { - 'package_path': {'key': 'packagePath', 'type': 'str'}, + 'package_path': {'key': 'packagePath', 'type': 'object'}, } def __init__(self, **kwargs): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/ssis_package_location_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/ssis_package_location_py3.py index 0449d4ff2943..af139da47d88 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/ssis_package_location_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/ssis_package_location_py3.py @@ -17,8 +17,9 @@ class SSISPackageLocation(Model): All required parameters must be populated in order to send to Azure. - :param package_path: Required. The SSIS package path. - :type package_path: str + :param package_path: Required. The SSIS package path. Type: string (or + Expression with resultType string). + :type package_path: object """ _validation = { @@ -26,9 +27,9 @@ class SSISPackageLocation(Model): } _attribute_map = { - 'package_path': {'key': 'packagePath', 'type': 'str'}, + 'package_path': {'key': 'packagePath', 'type': 'object'}, } - def __init__(self, *, package_path: str, **kwargs) -> None: + def __init__(self, *, package_path, **kwargs) -> None: super(SSISPackageLocation, self).__init__(**kwargs) self.package_path = package_path From 3af5560985326b70b875ba2cfa144711ff3d71fa Mon Sep 17 00:00:00 2001 From: Azure SDK for Python bot Date: Thu, 10 Jan 2019 12:50:15 -0800 Subject: [PATCH 40/54] [AutoPR datafactory/resource-manager] [Datafactory] Add Webhook Activity to Azure Data Factory (#4140) * Generated from 30ce8a5d068ff6b375d77746ce44810802593ade remove timeout as required * Generated from 0d3bebe04c8ee34399f725a3581ba55cebff214f combine and update web auth definitions * Generated from dafb0014942f618644438753c10df85ea504ccfa revert auth type enum change --- .../azure/mgmt/datafactory/models/__init__.py | 5 + .../datafactory/models/control_activity.py | 4 +- .../models/control_activity_py3.py | 4 +- .../data_factory_management_client_enums.py | 5 + .../datafactory/models/web_hook_activity.py | 94 +++++++++++++++++++ .../models/web_hook_activity_py3.py | 94 +++++++++++++++++++ 6 files changed, 202 insertions(+), 4 deletions(-) create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_hook_activity.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_hook_activity_py3.py diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py index c9728054c80a..96700c048a45 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py @@ -342,6 +342,7 @@ from .wait_activity_py3 import WaitActivity from .for_each_activity_py3 import ForEachActivity from .if_condition_activity_py3 import IfConditionActivity + from .web_hook_activity_py3 import WebHookActivity from .execute_pipeline_activity_py3 import ExecutePipelineActivity from .control_activity_py3 import ControlActivity from .linked_integration_runtime_py3 import LinkedIntegrationRuntime @@ -702,6 +703,7 @@ from .wait_activity import WaitActivity from .for_each_activity import ForEachActivity from .if_condition_activity import IfConditionActivity + from .web_hook_activity import WebHookActivity from .execute_pipeline_activity import ExecutePipelineActivity from .control_activity import ControlActivity from .linked_integration_runtime import LinkedIntegrationRuntime @@ -787,6 +789,7 @@ CopyBehaviorType, PolybaseSettingsRejectType, SapCloudForCustomerSinkWriteBehavior, + WebHookActivityMethod, IntegrationRuntimeType, SelfHostedIntegrationRuntimeNodeStatus, IntegrationRuntimeUpdateResult, @@ -1132,6 +1135,7 @@ 'WaitActivity', 'ForEachActivity', 'IfConditionActivity', + 'WebHookActivity', 'ExecutePipelineActivity', 'ControlActivity', 'LinkedIntegrationRuntime', @@ -1216,6 +1220,7 @@ 'CopyBehaviorType', 'PolybaseSettingsRejectType', 'SapCloudForCustomerSinkWriteBehavior', + 'WebHookActivityMethod', 'IntegrationRuntimeType', 'SelfHostedIntegrationRuntimeNodeStatus', 'IntegrationRuntimeUpdateResult', diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/control_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/control_activity.py index 16581581786b..e0f9592f9759 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/control_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/control_activity.py @@ -18,7 +18,7 @@ class ControlActivity(Activity): You probably want to use the sub-classes and not this class directly. Known sub-classes are: AppendVariableActivity, SetVariableActivity, FilterActivity, UntilActivity, WaitActivity, ForEachActivity, - IfConditionActivity, ExecutePipelineActivity + IfConditionActivity, WebHookActivity, ExecutePipelineActivity All required parameters must be populated in order to send to Azure. @@ -52,7 +52,7 @@ class ControlActivity(Activity): } _subtype_map = { - 'type': {'AppendVariable': 'AppendVariableActivity', 'SetVariable': 'SetVariableActivity', 'Filter': 'FilterActivity', 'Until': 'UntilActivity', 'Wait': 'WaitActivity', 'ForEach': 'ForEachActivity', 'IfCondition': 'IfConditionActivity', 'ExecutePipeline': 'ExecutePipelineActivity'} + 'type': {'AppendVariable': 'AppendVariableActivity', 'SetVariable': 'SetVariableActivity', 'Filter': 'FilterActivity', 'Until': 'UntilActivity', 'Wait': 'WaitActivity', 'ForEach': 'ForEachActivity', 'IfCondition': 'IfConditionActivity', 'WebHook': 'WebHookActivity', 'ExecutePipeline': 'ExecutePipelineActivity'} } def __init__(self, **kwargs): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/control_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/control_activity_py3.py index 739d8b9c311b..8021c7a34c14 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/control_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/control_activity_py3.py @@ -18,7 +18,7 @@ class ControlActivity(Activity): You probably want to use the sub-classes and not this class directly. Known sub-classes are: AppendVariableActivity, SetVariableActivity, FilterActivity, UntilActivity, WaitActivity, ForEachActivity, - IfConditionActivity, ExecutePipelineActivity + IfConditionActivity, WebHookActivity, ExecutePipelineActivity All required parameters must be populated in order to send to Azure. @@ -52,7 +52,7 @@ class ControlActivity(Activity): } _subtype_map = { - 'type': {'AppendVariable': 'AppendVariableActivity', 'SetVariable': 'SetVariableActivity', 'Filter': 'FilterActivity', 'Until': 'UntilActivity', 'Wait': 'WaitActivity', 'ForEach': 'ForEachActivity', 'IfCondition': 'IfConditionActivity', 'ExecutePipeline': 'ExecutePipelineActivity'} + 'type': {'AppendVariable': 'AppendVariableActivity', 'SetVariable': 'SetVariableActivity', 'Filter': 'FilterActivity', 'Until': 'UntilActivity', 'Wait': 'WaitActivity', 'ForEach': 'ForEachActivity', 'IfCondition': 'IfConditionActivity', 'WebHook': 'WebHookActivity', 'ExecutePipeline': 'ExecutePipelineActivity'} } def __init__(self, *, name: str, additional_properties=None, description: str=None, depends_on=None, user_properties=None, **kwargs) -> None: diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_factory_management_client_enums.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_factory_management_client_enums.py index 23334492e4d2..a135c4b2e101 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_factory_management_client_enums.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_factory_management_client_enums.py @@ -384,6 +384,11 @@ class SapCloudForCustomerSinkWriteBehavior(str, Enum): update = "Update" +class WebHookActivityMethod(str, Enum): + + post = "POST" + + class IntegrationRuntimeType(str, Enum): managed = "Managed" diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_hook_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_hook_activity.py new file mode 100644 index 000000000000..ca8945a9d697 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_hook_activity.py @@ -0,0 +1,94 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .control_activity import ControlActivity + + +class WebHookActivity(ControlActivity): + """WebHook activity. + + Variables are only populated by the server, and will be ignored when + sending a request. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param name: Required. Activity name. + :type name: str + :param description: Activity description. + :type description: str + :param depends_on: Activity depends on condition. + :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param user_properties: Activity user properties. + :type user_properties: list[~azure.mgmt.datafactory.models.UserProperty] + :param type: Required. Constant filled by server. + :type type: str + :ivar method: Required. Rest API method for target endpoint. Default + value: "POST" . + :vartype method: str + :param url: Required. WebHook activity target endpoint and path. Type: + string (or Expression with resultType string). + :type url: object + :param timeout: Specifies the timeout within which the webhook should be + called back. If there is no value specified, it takes the value of + TimeSpan.FromMinutes(10) which is 10 minutes as default. Type: string (or + Expression with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type timeout: object + :param headers: Represents the headers that will be sent to the request. + For example, to set the language and type on a request: "headers" : { + "Accept-Language": "en-us", "Content-Type": "application/json" }. Type: + string (or Expression with resultType string). + :type headers: object + :param body: Represents the payload that will be sent to the endpoint. + Required for POST/PUT method, not allowed for GET method Type: string (or + Expression with resultType string). + :type body: object + :param authentication: Authentication method used for calling the + endpoint. + :type authentication: + ~azure.mgmt.datafactory.models.WebActivityAuthentication + """ + + _validation = { + 'name': {'required': True}, + 'type': {'required': True}, + 'method': {'required': True, 'constant': True}, + 'url': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'name': {'key': 'name', 'type': 'str'}, + 'description': {'key': 'description', 'type': 'str'}, + 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'method': {'key': 'typeProperties.method', 'type': 'str'}, + 'url': {'key': 'typeProperties.url', 'type': 'object'}, + 'timeout': {'key': 'typeProperties.timeout', 'type': 'object'}, + 'headers': {'key': 'typeProperties.headers', 'type': 'object'}, + 'body': {'key': 'typeProperties.body', 'type': 'object'}, + 'authentication': {'key': 'typeProperties.authentication', 'type': 'WebActivityAuthentication'}, + } + + method = "POST" + + def __init__(self, **kwargs): + super(WebHookActivity, self).__init__(**kwargs) + self.url = kwargs.get('url', None) + self.timeout = kwargs.get('timeout', None) + self.headers = kwargs.get('headers', None) + self.body = kwargs.get('body', None) + self.authentication = kwargs.get('authentication', None) + self.type = 'WebHook' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_hook_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_hook_activity_py3.py new file mode 100644 index 000000000000..e4779bfbca07 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_hook_activity_py3.py @@ -0,0 +1,94 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .control_activity_py3 import ControlActivity + + +class WebHookActivity(ControlActivity): + """WebHook activity. + + Variables are only populated by the server, and will be ignored when + sending a request. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param name: Required. Activity name. + :type name: str + :param description: Activity description. + :type description: str + :param depends_on: Activity depends on condition. + :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param user_properties: Activity user properties. + :type user_properties: list[~azure.mgmt.datafactory.models.UserProperty] + :param type: Required. Constant filled by server. + :type type: str + :ivar method: Required. Rest API method for target endpoint. Default + value: "POST" . + :vartype method: str + :param url: Required. WebHook activity target endpoint and path. Type: + string (or Expression with resultType string). + :type url: object + :param timeout: Specifies the timeout within which the webhook should be + called back. If there is no value specified, it takes the value of + TimeSpan.FromMinutes(10) which is 10 minutes as default. Type: string (or + Expression with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type timeout: object + :param headers: Represents the headers that will be sent to the request. + For example, to set the language and type on a request: "headers" : { + "Accept-Language": "en-us", "Content-Type": "application/json" }. Type: + string (or Expression with resultType string). + :type headers: object + :param body: Represents the payload that will be sent to the endpoint. + Required for POST/PUT method, not allowed for GET method Type: string (or + Expression with resultType string). + :type body: object + :param authentication: Authentication method used for calling the + endpoint. + :type authentication: + ~azure.mgmt.datafactory.models.WebActivityAuthentication + """ + + _validation = { + 'name': {'required': True}, + 'type': {'required': True}, + 'method': {'required': True, 'constant': True}, + 'url': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'name': {'key': 'name', 'type': 'str'}, + 'description': {'key': 'description', 'type': 'str'}, + 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'method': {'key': 'typeProperties.method', 'type': 'str'}, + 'url': {'key': 'typeProperties.url', 'type': 'object'}, + 'timeout': {'key': 'typeProperties.timeout', 'type': 'object'}, + 'headers': {'key': 'typeProperties.headers', 'type': 'object'}, + 'body': {'key': 'typeProperties.body', 'type': 'object'}, + 'authentication': {'key': 'typeProperties.authentication', 'type': 'WebActivityAuthentication'}, + } + + method = "POST" + + def __init__(self, *, name: str, url, additional_properties=None, description: str=None, depends_on=None, user_properties=None, timeout=None, headers=None, body=None, authentication=None, **kwargs) -> None: + super(WebHookActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, **kwargs) + self.url = url + self.timeout = timeout + self.headers = headers + self.body = body + self.authentication = authentication + self.type = 'WebHook' From 34952e8331ea4e263187770e0a672ca831da60a8 Mon Sep 17 00:00:00 2001 From: Azure SDK for Python bot Date: Mon, 14 Jan 2019 09:59:07 -0800 Subject: [PATCH 41/54] Generated from af8bffa10fefa97dffc4d50f127481f09e74174e (#4173) Winauth Support in SSIS Activity --- .../azure/mgmt/datafactory/models/__init__.py | 3 ++ .../models/execute_ssis_package_activity.py | 5 +++ .../execute_ssis_package_activity_py3.py | 7 ++- .../models/ssis_execution_credential.py | 44 +++++++++++++++++++ .../models/ssis_execution_credential_py3.py | 44 +++++++++++++++++++ 5 files changed, 102 insertions(+), 1 deletion(-) create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/ssis_execution_credential.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/ssis_execution_credential_py3.py diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py index 96700c048a45..144d6e2998ae 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py @@ -306,6 +306,7 @@ from .custom_activity_py3 import CustomActivity from .ssis_property_override_py3 import SSISPropertyOverride from .ssis_execution_parameter_py3 import SSISExecutionParameter + from .ssis_execution_credential_py3 import SSISExecutionCredential from .ssis_package_location_py3 import SSISPackageLocation from .execute_ssis_package_activity_py3 import ExecuteSSISPackageActivity from .hd_insight_spark_activity_py3 import HDInsightSparkActivity @@ -667,6 +668,7 @@ from .custom_activity import CustomActivity from .ssis_property_override import SSISPropertyOverride from .ssis_execution_parameter import SSISExecutionParameter + from .ssis_execution_credential import SSISExecutionCredential from .ssis_package_location import SSISPackageLocation from .execute_ssis_package_activity import ExecuteSSISPackageActivity from .hd_insight_spark_activity import HDInsightSparkActivity @@ -1099,6 +1101,7 @@ 'CustomActivity', 'SSISPropertyOverride', 'SSISExecutionParameter', + 'SSISExecutionCredential', 'SSISPackageLocation', 'ExecuteSSISPackageActivity', 'HDInsightSparkActivity', diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_ssis_package_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_ssis_package_activity.py index 3e5dc62320cb..150e8481f0df 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_ssis_package_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_ssis_package_activity.py @@ -47,6 +47,9 @@ class ExecuteSSISPackageActivity(ExecutionActivity): :param environment_path: The environment path to execute the SSIS package. Type: string (or Expression with resultType string). :type environment_path: object + :param execution_credential: The package execution credential. + :type execution_credential: dict[str, + ~azure.mgmt.datafactory.models.SSISExecutionCredential] :param connect_via: Required. The integration runtime reference. :type connect_via: ~azure.mgmt.datafactory.models.IntegrationRuntimeReference @@ -92,6 +95,7 @@ class ExecuteSSISPackageActivity(ExecutionActivity): 'runtime': {'key': 'typeProperties.runtime', 'type': 'object'}, 'logging_level': {'key': 'typeProperties.loggingLevel', 'type': 'object'}, 'environment_path': {'key': 'typeProperties.environmentPath', 'type': 'object'}, + 'execution_credential': {'key': 'typeProperties.executionCredential', 'type': '{SSISExecutionCredential}'}, 'connect_via': {'key': 'typeProperties.connectVia', 'type': 'IntegrationRuntimeReference'}, 'project_parameters': {'key': 'typeProperties.projectParameters', 'type': '{SSISExecutionParameter}'}, 'package_parameters': {'key': 'typeProperties.packageParameters', 'type': '{SSISExecutionParameter}'}, @@ -106,6 +110,7 @@ def __init__(self, **kwargs): self.runtime = kwargs.get('runtime', None) self.logging_level = kwargs.get('logging_level', None) self.environment_path = kwargs.get('environment_path', None) + self.execution_credential = kwargs.get('execution_credential', None) self.connect_via = kwargs.get('connect_via', None) self.project_parameters = kwargs.get('project_parameters', None) self.package_parameters = kwargs.get('package_parameters', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_ssis_package_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_ssis_package_activity_py3.py index 0a716e8c9482..59d65046c96f 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_ssis_package_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_ssis_package_activity_py3.py @@ -47,6 +47,9 @@ class ExecuteSSISPackageActivity(ExecutionActivity): :param environment_path: The environment path to execute the SSIS package. Type: string (or Expression with resultType string). :type environment_path: object + :param execution_credential: The package execution credential. + :type execution_credential: dict[str, + ~azure.mgmt.datafactory.models.SSISExecutionCredential] :param connect_via: Required. The integration runtime reference. :type connect_via: ~azure.mgmt.datafactory.models.IntegrationRuntimeReference @@ -92,6 +95,7 @@ class ExecuteSSISPackageActivity(ExecutionActivity): 'runtime': {'key': 'typeProperties.runtime', 'type': 'object'}, 'logging_level': {'key': 'typeProperties.loggingLevel', 'type': 'object'}, 'environment_path': {'key': 'typeProperties.environmentPath', 'type': 'object'}, + 'execution_credential': {'key': 'typeProperties.executionCredential', 'type': '{SSISExecutionCredential}'}, 'connect_via': {'key': 'typeProperties.connectVia', 'type': 'IntegrationRuntimeReference'}, 'project_parameters': {'key': 'typeProperties.projectParameters', 'type': '{SSISExecutionParameter}'}, 'package_parameters': {'key': 'typeProperties.packageParameters', 'type': '{SSISExecutionParameter}'}, @@ -100,12 +104,13 @@ class ExecuteSSISPackageActivity(ExecutionActivity): 'property_overrides': {'key': 'typeProperties.propertyOverrides', 'type': '{SSISPropertyOverride}'}, } - def __init__(self, *, name: str, package_location, connect_via, additional_properties=None, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, runtime=None, logging_level=None, environment_path=None, project_parameters=None, package_parameters=None, project_connection_managers=None, package_connection_managers=None, property_overrides=None, **kwargs) -> None: + def __init__(self, *, name: str, package_location, connect_via, additional_properties=None, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, runtime=None, logging_level=None, environment_path=None, execution_credential=None, project_parameters=None, package_parameters=None, project_connection_managers=None, package_connection_managers=None, property_overrides=None, **kwargs) -> None: super(ExecuteSSISPackageActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) self.package_location = package_location self.runtime = runtime self.logging_level = logging_level self.environment_path = environment_path + self.execution_credential = execution_credential self.connect_via = connect_via self.project_parameters = project_parameters self.package_parameters = package_parameters diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/ssis_execution_credential.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/ssis_execution_credential.py new file mode 100644 index 000000000000..c090694416a9 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/ssis_execution_credential.py @@ -0,0 +1,44 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class SSISExecutionCredential(Model): + """SSIS package execution credential. + + All required parameters must be populated in order to send to Azure. + + :param domain: Required. Domain for windows authentication. + :type domain: object + :param user_name: Required. UseName for windows authentication. + :type user_name: object + :param password: Required. Password for windows authentication. + :type password: ~azure.mgmt.datafactory.models.SecureString + """ + + _validation = { + 'domain': {'required': True}, + 'user_name': {'required': True}, + 'password': {'required': True}, + } + + _attribute_map = { + 'domain': {'key': 'domain', 'type': 'object'}, + 'user_name': {'key': 'userName', 'type': 'object'}, + 'password': {'key': 'password', 'type': 'SecureString'}, + } + + def __init__(self, **kwargs): + super(SSISExecutionCredential, self).__init__(**kwargs) + self.domain = kwargs.get('domain', None) + self.user_name = kwargs.get('user_name', None) + self.password = kwargs.get('password', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/ssis_execution_credential_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/ssis_execution_credential_py3.py new file mode 100644 index 000000000000..051eaffa2bf2 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/ssis_execution_credential_py3.py @@ -0,0 +1,44 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class SSISExecutionCredential(Model): + """SSIS package execution credential. + + All required parameters must be populated in order to send to Azure. + + :param domain: Required. Domain for windows authentication. + :type domain: object + :param user_name: Required. UseName for windows authentication. + :type user_name: object + :param password: Required. Password for windows authentication. + :type password: ~azure.mgmt.datafactory.models.SecureString + """ + + _validation = { + 'domain': {'required': True}, + 'user_name': {'required': True}, + 'password': {'required': True}, + } + + _attribute_map = { + 'domain': {'key': 'domain', 'type': 'object'}, + 'user_name': {'key': 'userName', 'type': 'object'}, + 'password': {'key': 'password', 'type': 'SecureString'}, + } + + def __init__(self, *, domain, user_name, password, **kwargs) -> None: + super(SSISExecutionCredential, self).__init__(**kwargs) + self.domain = domain + self.user_name = user_name + self.password = password From e566fa9d09399e1c92db7fe751a389b4d4c0a89d Mon Sep 17 00:00:00 2001 From: Azure SDK for Python bot Date: Tue, 15 Jan 2019 11:54:48 -0800 Subject: [PATCH 42/54] [AutoPR datafactory/resource-manager] [DataFactory]Support schema property for dataset (#4156) * Generated from c36d6d87e9a4ae714416dd7ece0528c86cc97514 Support schema property for dataset * Generated from c36d6d87e9a4ae714416dd7ece0528c86cc97514 Support schema property for dataset --- .../mgmt/datafactory/models/amazon_mws_object_dataset.py | 5 +++++ .../datafactory/models/amazon_mws_object_dataset_py3.py | 9 +++++++-- .../azure/mgmt/datafactory/models/amazon_s3_dataset.py | 5 +++++ .../mgmt/datafactory/models/amazon_s3_dataset_py3.py | 9 +++++++-- .../azure/mgmt/datafactory/models/azure_blob_dataset.py | 5 +++++ .../mgmt/datafactory/models/azure_blob_dataset_py3.py | 9 +++++++-- .../datafactory/models/azure_data_lake_store_dataset.py | 5 +++++ .../models/azure_data_lake_store_dataset_py3.py | 9 +++++++-- .../datafactory/models/azure_my_sql_table_dataset.py | 5 +++++ .../datafactory/models/azure_my_sql_table_dataset_py3.py | 9 +++++++-- .../models/azure_postgre_sql_table_dataset.py | 5 +++++ .../models/azure_postgre_sql_table_dataset_py3.py | 9 +++++++-- .../datafactory/models/azure_search_index_dataset.py | 5 +++++ .../datafactory/models/azure_search_index_dataset_py3.py | 9 +++++++-- .../datafactory/models/azure_sql_dw_table_dataset.py | 5 +++++ .../datafactory/models/azure_sql_dw_table_dataset_py3.py | 9 +++++++-- .../mgmt/datafactory/models/azure_sql_table_dataset.py | 5 +++++ .../datafactory/models/azure_sql_table_dataset_py3.py | 9 +++++++-- .../azure/mgmt/datafactory/models/azure_table_dataset.py | 5 +++++ .../mgmt/datafactory/models/azure_table_dataset_py3.py | 9 +++++++-- .../mgmt/datafactory/models/cassandra_table_dataset.py | 5 +++++ .../datafactory/models/cassandra_table_dataset_py3.py | 9 +++++++-- .../mgmt/datafactory/models/concur_object_dataset.py | 5 +++++ .../mgmt/datafactory/models/concur_object_dataset_py3.py | 9 +++++++-- .../mgmt/datafactory/models/couchbase_table_dataset.py | 5 +++++ .../datafactory/models/couchbase_table_dataset_py3.py | 9 +++++++-- .../azure/mgmt/datafactory/models/custom_dataset.py | 5 +++++ .../azure/mgmt/datafactory/models/custom_dataset_py3.py | 9 +++++++-- .../azure/mgmt/datafactory/models/dataset.py | 6 ++++++ .../azure/mgmt/datafactory/models/dataset_py3.py | 8 +++++++- .../datafactory/models/document_db_collection_dataset.py | 5 +++++ .../models/document_db_collection_dataset_py3.py | 9 +++++++-- .../azure/mgmt/datafactory/models/drill_table_dataset.py | 5 +++++ .../mgmt/datafactory/models/drill_table_dataset_py3.py | 9 +++++++-- .../mgmt/datafactory/models/dynamics_entity_dataset.py | 5 +++++ .../datafactory/models/dynamics_entity_dataset_py3.py | 9 +++++++-- .../mgmt/datafactory/models/eloqua_object_dataset.py | 5 +++++ .../mgmt/datafactory/models/eloqua_object_dataset_py3.py | 9 +++++++-- .../azure/mgmt/datafactory/models/file_share_dataset.py | 5 +++++ .../mgmt/datafactory/models/file_share_dataset_py3.py | 9 +++++++-- .../models/google_big_query_object_dataset.py | 5 +++++ .../models/google_big_query_object_dataset_py3.py | 9 +++++++-- .../mgmt/datafactory/models/greenplum_table_dataset.py | 5 +++++ .../datafactory/models/greenplum_table_dataset_py3.py | 9 +++++++-- .../mgmt/datafactory/models/hbase_object_dataset.py | 5 +++++ .../mgmt/datafactory/models/hbase_object_dataset_py3.py | 9 +++++++-- .../azure/mgmt/datafactory/models/hive_object_dataset.py | 5 +++++ .../mgmt/datafactory/models/hive_object_dataset_py3.py | 9 +++++++-- .../azure/mgmt/datafactory/models/http_dataset.py | 5 +++++ .../azure/mgmt/datafactory/models/http_dataset_py3.py | 9 +++++++-- .../mgmt/datafactory/models/hubspot_object_dataset.py | 5 +++++ .../datafactory/models/hubspot_object_dataset_py3.py | 9 +++++++-- .../mgmt/datafactory/models/impala_object_dataset.py | 5 +++++ .../mgmt/datafactory/models/impala_object_dataset_py3.py | 9 +++++++-- .../azure/mgmt/datafactory/models/jira_object_dataset.py | 5 +++++ .../mgmt/datafactory/models/jira_object_dataset_py3.py | 9 +++++++-- .../mgmt/datafactory/models/magento_object_dataset.py | 5 +++++ .../datafactory/models/magento_object_dataset_py3.py | 9 +++++++-- .../mgmt/datafactory/models/maria_db_table_dataset.py | 5 +++++ .../datafactory/models/maria_db_table_dataset_py3.py | 9 +++++++-- .../mgmt/datafactory/models/marketo_object_dataset.py | 5 +++++ .../datafactory/models/marketo_object_dataset_py3.py | 9 +++++++-- .../datafactory/models/mongo_db_collection_dataset.py | 5 +++++ .../models/mongo_db_collection_dataset_py3.py | 9 +++++++-- .../mgmt/datafactory/models/netezza_table_dataset.py | 5 +++++ .../mgmt/datafactory/models/netezza_table_dataset_py3.py | 9 +++++++-- .../mgmt/datafactory/models/odata_resource_dataset.py | 5 +++++ .../datafactory/models/odata_resource_dataset_py3.py | 9 +++++++-- .../mgmt/datafactory/models/oracle_table_dataset.py | 5 +++++ .../mgmt/datafactory/models/oracle_table_dataset_py3.py | 9 +++++++-- .../mgmt/datafactory/models/paypal_object_dataset.py | 5 +++++ .../mgmt/datafactory/models/paypal_object_dataset_py3.py | 9 +++++++-- .../mgmt/datafactory/models/phoenix_object_dataset.py | 5 +++++ .../datafactory/models/phoenix_object_dataset_py3.py | 9 +++++++-- .../mgmt/datafactory/models/presto_object_dataset.py | 5 +++++ .../mgmt/datafactory/models/presto_object_dataset_py3.py | 9 +++++++-- .../datafactory/models/quick_books_object_dataset.py | 5 +++++ .../datafactory/models/quick_books_object_dataset_py3.py | 9 +++++++-- .../mgmt/datafactory/models/relational_table_dataset.py | 5 +++++ .../datafactory/models/relational_table_dataset_py3.py | 9 +++++++-- .../mgmt/datafactory/models/responsys_object_dataset.py | 5 +++++ .../datafactory/models/responsys_object_dataset_py3.py | 9 +++++++-- .../models/salesforce_marketing_cloud_object_dataset.py | 5 +++++ .../salesforce_marketing_cloud_object_dataset_py3.py | 9 +++++++-- .../mgmt/datafactory/models/salesforce_object_dataset.py | 5 +++++ .../datafactory/models/salesforce_object_dataset_py3.py | 9 +++++++-- .../models/sap_cloud_for_customer_resource_dataset.py | 5 +++++ .../sap_cloud_for_customer_resource_dataset_py3.py | 9 +++++++-- .../mgmt/datafactory/models/sap_ecc_resource_dataset.py | 5 +++++ .../datafactory/models/sap_ecc_resource_dataset_py3.py | 9 +++++++-- .../datafactory/models/service_now_object_dataset.py | 5 +++++ .../datafactory/models/service_now_object_dataset_py3.py | 9 +++++++-- .../mgmt/datafactory/models/shopify_object_dataset.py | 5 +++++ .../datafactory/models/shopify_object_dataset_py3.py | 9 +++++++-- .../mgmt/datafactory/models/spark_object_dataset.py | 5 +++++ .../mgmt/datafactory/models/spark_object_dataset_py3.py | 9 +++++++-- .../mgmt/datafactory/models/sql_server_table_dataset.py | 5 +++++ .../datafactory/models/sql_server_table_dataset_py3.py | 9 +++++++-- .../mgmt/datafactory/models/square_object_dataset.py | 5 +++++ .../mgmt/datafactory/models/square_object_dataset_py3.py | 9 +++++++-- .../mgmt/datafactory/models/vertica_table_dataset.py | 5 +++++ .../mgmt/datafactory/models/vertica_table_dataset_py3.py | 9 +++++++-- .../azure/mgmt/datafactory/models/web_table_dataset.py | 5 +++++ .../mgmt/datafactory/models/web_table_dataset_py3.py | 9 +++++++-- .../azure/mgmt/datafactory/models/xero_object_dataset.py | 5 +++++ .../mgmt/datafactory/models/xero_object_dataset_py3.py | 9 +++++++-- .../azure/mgmt/datafactory/models/zoho_object_dataset.py | 5 +++++ .../mgmt/datafactory/models/zoho_object_dataset_py3.py | 9 +++++++-- 108 files changed, 649 insertions(+), 107 deletions(-) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_object_dataset.py index 4b6782cee208..9885f5c77d8c 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_object_dataset.py @@ -25,6 +25,10 @@ class AmazonMWSObjectDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -53,6 +57,7 @@ class AmazonMWSObjectDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_object_dataset_py3.py index abe8569df119..015ed9401c15 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_object_dataset_py3.py @@ -25,6 +25,10 @@ class AmazonMWSObjectDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -53,6 +57,7 @@ class AmazonMWSObjectDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, @@ -61,7 +66,7 @@ class AmazonMWSObjectDataset(Dataset): 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: - super(AmazonMWSObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, schema=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: + super(AmazonMWSObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.table_name = table_name self.type = 'AmazonMWSObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_s3_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_s3_dataset.py index 87540eb0eb11..d6262a013b0d 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_s3_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_s3_dataset.py @@ -25,6 +25,10 @@ class AmazonS3Dataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -68,6 +72,7 @@ class AmazonS3Dataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_s3_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_s3_dataset_py3.py index c4b602da1303..3936e9646a09 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_s3_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_s3_dataset_py3.py @@ -25,6 +25,10 @@ class AmazonS3Dataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -68,6 +72,7 @@ class AmazonS3Dataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, @@ -81,8 +86,8 @@ class AmazonS3Dataset(Dataset): 'compression': {'key': 'typeProperties.compression', 'type': 'DatasetCompression'}, } - def __init__(self, *, linked_service_name, bucket_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, key=None, prefix=None, version=None, format=None, compression=None, **kwargs) -> None: - super(AmazonS3Dataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + def __init__(self, *, linked_service_name, bucket_name, additional_properties=None, description: str=None, structure=None, schema=None, parameters=None, annotations=None, folder=None, key=None, prefix=None, version=None, format=None, compression=None, **kwargs) -> None: + super(AmazonS3Dataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.bucket_name = bucket_name self.key = key self.prefix = prefix diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_dataset.py index dfa31040d658..c3f4ffc118ba 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_dataset.py @@ -25,6 +25,10 @@ class AzureBlobDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -63,6 +67,7 @@ class AzureBlobDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_dataset_py3.py index adef2dbd6def..7567e1fba9fb 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_dataset_py3.py @@ -25,6 +25,10 @@ class AzureBlobDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -63,6 +67,7 @@ class AzureBlobDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, @@ -75,8 +80,8 @@ class AzureBlobDataset(Dataset): 'compression': {'key': 'typeProperties.compression', 'type': 'DatasetCompression'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, folder_path=None, table_root_location=None, file_name=None, format=None, compression=None, **kwargs) -> None: - super(AzureBlobDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, schema=None, parameters=None, annotations=None, folder=None, folder_path=None, table_root_location=None, file_name=None, format=None, compression=None, **kwargs) -> None: + super(AzureBlobDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.folder_path = folder_path self.table_root_location = table_root_location self.file_name = file_name diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_dataset.py index 0d3d9edf45d9..e0299ba2bcad 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_dataset.py @@ -25,6 +25,10 @@ class AzureDataLakeStoreDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -62,6 +66,7 @@ class AzureDataLakeStoreDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_dataset_py3.py index 8d108c944dcb..62e761dc9695 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_dataset_py3.py @@ -25,6 +25,10 @@ class AzureDataLakeStoreDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -62,6 +66,7 @@ class AzureDataLakeStoreDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, @@ -73,8 +78,8 @@ class AzureDataLakeStoreDataset(Dataset): 'compression': {'key': 'typeProperties.compression', 'type': 'DatasetCompression'}, } - def __init__(self, *, linked_service_name, folder_path, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, file_name=None, format=None, compression=None, **kwargs) -> None: - super(AzureDataLakeStoreDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + def __init__(self, *, linked_service_name, folder_path, additional_properties=None, description: str=None, structure=None, schema=None, parameters=None, annotations=None, folder=None, file_name=None, format=None, compression=None, **kwargs) -> None: + super(AzureDataLakeStoreDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.folder_path = folder_path self.file_name = file_name self.format = format diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_table_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_table_dataset.py index bbea45836145..8f5d43478089 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_table_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_table_dataset.py @@ -25,6 +25,10 @@ class AzureMySqlTableDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -53,6 +57,7 @@ class AzureMySqlTableDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_table_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_table_dataset_py3.py index 1fc1fef3201f..7bd7eb6f17f8 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_table_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_table_dataset_py3.py @@ -25,6 +25,10 @@ class AzureMySqlTableDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -53,6 +57,7 @@ class AzureMySqlTableDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, @@ -61,7 +66,7 @@ class AzureMySqlTableDataset(Dataset): 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: - super(AzureMySqlTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, schema=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: + super(AzureMySqlTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.table_name = table_name self.type = 'AzureMySqlTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_table_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_table_dataset.py index 02ce3e1e20f9..8960acc0df75 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_table_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_table_dataset.py @@ -25,6 +25,10 @@ class AzurePostgreSqlTableDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -53,6 +57,7 @@ class AzurePostgreSqlTableDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_table_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_table_dataset_py3.py index 88b251d4cb59..fddf0720c565 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_table_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_table_dataset_py3.py @@ -25,6 +25,10 @@ class AzurePostgreSqlTableDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -53,6 +57,7 @@ class AzurePostgreSqlTableDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, @@ -61,7 +66,7 @@ class AzurePostgreSqlTableDataset(Dataset): 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: - super(AzurePostgreSqlTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, schema=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: + super(AzurePostgreSqlTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.table_name = table_name self.type = 'AzurePostgreSqlTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_search_index_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_search_index_dataset.py index de9f38cf6793..1239bbad78fc 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_search_index_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_search_index_dataset.py @@ -25,6 +25,10 @@ class AzureSearchIndexDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -54,6 +58,7 @@ class AzureSearchIndexDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_search_index_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_search_index_dataset_py3.py index 6d57a2d3faad..da5e92dd2edd 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_search_index_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_search_index_dataset_py3.py @@ -25,6 +25,10 @@ class AzureSearchIndexDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -54,6 +58,7 @@ class AzureSearchIndexDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, @@ -62,7 +67,7 @@ class AzureSearchIndexDataset(Dataset): 'index_name': {'key': 'typeProperties.indexName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, index_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: - super(AzureSearchIndexDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + def __init__(self, *, linked_service_name, index_name, additional_properties=None, description: str=None, structure=None, schema=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + super(AzureSearchIndexDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.index_name = index_name self.type = 'AzureSearchIndex' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_dw_table_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_dw_table_dataset.py index 75dcf5778ba3..e12464efdd49 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_dw_table_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_dw_table_dataset.py @@ -25,6 +25,10 @@ class AzureSqlDWTableDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -54,6 +58,7 @@ class AzureSqlDWTableDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_dw_table_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_dw_table_dataset_py3.py index 516d9fc80d69..a06073f86c5b 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_dw_table_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_dw_table_dataset_py3.py @@ -25,6 +25,10 @@ class AzureSqlDWTableDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -54,6 +58,7 @@ class AzureSqlDWTableDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, @@ -62,7 +67,7 @@ class AzureSqlDWTableDataset(Dataset): 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, table_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: - super(AzureSqlDWTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + def __init__(self, *, linked_service_name, table_name, additional_properties=None, description: str=None, structure=None, schema=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + super(AzureSqlDWTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.table_name = table_name self.type = 'AzureSqlDWTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_table_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_table_dataset.py index 6e7cd0acc419..9a078241d620 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_table_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_table_dataset.py @@ -25,6 +25,10 @@ class AzureSqlTableDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -54,6 +58,7 @@ class AzureSqlTableDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_table_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_table_dataset_py3.py index 1769551763da..b3c388202f52 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_table_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_sql_table_dataset_py3.py @@ -25,6 +25,10 @@ class AzureSqlTableDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -54,6 +58,7 @@ class AzureSqlTableDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, @@ -62,7 +67,7 @@ class AzureSqlTableDataset(Dataset): 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, table_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: - super(AzureSqlTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + def __init__(self, *, linked_service_name, table_name, additional_properties=None, description: str=None, structure=None, schema=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + super(AzureSqlTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.table_name = table_name self.type = 'AzureSqlTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_table_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_table_dataset.py index dae0a5b3bf02..eb8dacbfbb98 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_table_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_table_dataset.py @@ -25,6 +25,10 @@ class AzureTableDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -54,6 +58,7 @@ class AzureTableDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_table_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_table_dataset_py3.py index 74f0ae0b7096..d70a15fdd6f1 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_table_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_table_dataset_py3.py @@ -25,6 +25,10 @@ class AzureTableDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -54,6 +58,7 @@ class AzureTableDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, @@ -62,7 +67,7 @@ class AzureTableDataset(Dataset): 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, table_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: - super(AzureTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + def __init__(self, *, linked_service_name, table_name, additional_properties=None, description: str=None, structure=None, schema=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + super(AzureTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.table_name = table_name self.type = 'AzureTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cassandra_table_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cassandra_table_dataset.py index b356830de832..b89c324fd4d4 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cassandra_table_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cassandra_table_dataset.py @@ -25,6 +25,10 @@ class CassandraTableDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -56,6 +60,7 @@ class CassandraTableDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cassandra_table_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cassandra_table_dataset_py3.py index 914c8b38cb62..256358ce50cb 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cassandra_table_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cassandra_table_dataset_py3.py @@ -25,6 +25,10 @@ class CassandraTableDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -56,6 +60,7 @@ class CassandraTableDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, @@ -65,8 +70,8 @@ class CassandraTableDataset(Dataset): 'keyspace': {'key': 'typeProperties.keyspace', 'type': 'object'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, table_name=None, keyspace=None, **kwargs) -> None: - super(CassandraTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, schema=None, parameters=None, annotations=None, folder=None, table_name=None, keyspace=None, **kwargs) -> None: + super(CassandraTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.table_name = table_name self.keyspace = keyspace self.type = 'CassandraTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_object_dataset.py index 1aab27f4f8f3..e2595f9d8aff 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_object_dataset.py @@ -25,6 +25,10 @@ class ConcurObjectDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -53,6 +57,7 @@ class ConcurObjectDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_object_dataset_py3.py index 9ae3833082c6..9543a6395a32 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_object_dataset_py3.py @@ -25,6 +25,10 @@ class ConcurObjectDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -53,6 +57,7 @@ class ConcurObjectDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, @@ -61,7 +66,7 @@ class ConcurObjectDataset(Dataset): 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: - super(ConcurObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, schema=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: + super(ConcurObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.table_name = table_name self.type = 'ConcurObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_table_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_table_dataset.py index cb87312e1377..821274b9aae4 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_table_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_table_dataset.py @@ -25,6 +25,10 @@ class CouchbaseTableDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -53,6 +57,7 @@ class CouchbaseTableDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_table_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_table_dataset_py3.py index 1fa41220cfc1..cf5299fd55a5 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_table_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_table_dataset_py3.py @@ -25,6 +25,10 @@ class CouchbaseTableDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -53,6 +57,7 @@ class CouchbaseTableDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, @@ -61,7 +66,7 @@ class CouchbaseTableDataset(Dataset): 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: - super(CouchbaseTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, schema=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: + super(CouchbaseTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.table_name = table_name self.type = 'CouchbaseTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_dataset.py index b74f0e44cf8e..8a6a8ac662a8 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_dataset.py @@ -25,6 +25,10 @@ class CustomDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -53,6 +57,7 @@ class CustomDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_dataset_py3.py index 65927670ca6c..da681e8360b8 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/custom_dataset_py3.py @@ -25,6 +25,10 @@ class CustomDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -53,6 +57,7 @@ class CustomDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, @@ -61,7 +66,7 @@ class CustomDataset(Dataset): 'type_properties': {'key': 'typeProperties', 'type': 'object'}, } - def __init__(self, *, linked_service_name, type_properties, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: - super(CustomDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + def __init__(self, *, linked_service_name, type_properties, additional_properties=None, description: str=None, structure=None, schema=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + super(CustomDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.type_properties = type_properties self.type = 'CustomDataset' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset.py index 626015fd5d11..de812815bd26 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset.py @@ -48,6 +48,10 @@ class Dataset(Model): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -73,6 +77,7 @@ class Dataset(Model): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, @@ -89,6 +94,7 @@ def __init__(self, **kwargs): self.additional_properties = kwargs.get('additional_properties', None) self.description = kwargs.get('description', None) self.structure = kwargs.get('structure', None) + self.schema = kwargs.get('schema', None) self.linked_service_name = kwargs.get('linked_service_name', None) self.parameters = kwargs.get('parameters', None) self.annotations = kwargs.get('annotations', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_py3.py index 3d67a1328cd4..9538e6105a8f 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_py3.py @@ -48,6 +48,10 @@ class Dataset(Model): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -73,6 +77,7 @@ class Dataset(Model): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, @@ -84,11 +89,12 @@ class Dataset(Model): 'type': {'ResponsysObject': 'ResponsysObjectDataset', 'SalesforceMarketingCloudObject': 'SalesforceMarketingCloudObjectDataset', 'VerticaTable': 'VerticaTableDataset', 'NetezzaTable': 'NetezzaTableDataset', 'ZohoObject': 'ZohoObjectDataset', 'XeroObject': 'XeroObjectDataset', 'SquareObject': 'SquareObjectDataset', 'SparkObject': 'SparkObjectDataset', 'ShopifyObject': 'ShopifyObjectDataset', 'ServiceNowObject': 'ServiceNowObjectDataset', 'QuickBooksObject': 'QuickBooksObjectDataset', 'PrestoObject': 'PrestoObjectDataset', 'PhoenixObject': 'PhoenixObjectDataset', 'PaypalObject': 'PaypalObjectDataset', 'MarketoObject': 'MarketoObjectDataset', 'MariaDBTable': 'MariaDBTableDataset', 'MagentoObject': 'MagentoObjectDataset', 'JiraObject': 'JiraObjectDataset', 'ImpalaObject': 'ImpalaObjectDataset', 'HubspotObject': 'HubspotObjectDataset', 'HiveObject': 'HiveObjectDataset', 'HBaseObject': 'HBaseObjectDataset', 'GreenplumTable': 'GreenplumTableDataset', 'GoogleBigQueryObject': 'GoogleBigQueryObjectDataset', 'EloquaObject': 'EloquaObjectDataset', 'DrillTable': 'DrillTableDataset', 'CouchbaseTable': 'CouchbaseTableDataset', 'ConcurObject': 'ConcurObjectDataset', 'AzurePostgreSqlTable': 'AzurePostgreSqlTableDataset', 'AmazonMWSObject': 'AmazonMWSObjectDataset', 'HttpFile': 'HttpDataset', 'AzureSearchIndex': 'AzureSearchIndexDataset', 'WebTable': 'WebTableDataset', 'SqlServerTable': 'SqlServerTableDataset', 'SapEccResource': 'SapEccResourceDataset', 'SapCloudForCustomerResource': 'SapCloudForCustomerResourceDataset', 'SalesforceObject': 'SalesforceObjectDataset', 'RelationalTable': 'RelationalTableDataset', 'AzureMySqlTable': 'AzureMySqlTableDataset', 'OracleTable': 'OracleTableDataset', 'ODataResource': 'ODataResourceDataset', 'MongoDbCollection': 'MongoDbCollectionDataset', 'FileShare': 'FileShareDataset', 'AzureDataLakeStoreFile': 'AzureDataLakeStoreDataset', 'DynamicsEntity': 'DynamicsEntityDataset', 'DocumentDbCollection': 'DocumentDbCollectionDataset', 'CustomDataset': 'CustomDataset', 'CassandraTable': 'CassandraTableDataset', 'AzureSqlDWTable': 'AzureSqlDWTableDataset', 'AzureSqlTable': 'AzureSqlTableDataset', 'AzureTable': 'AzureTableDataset', 'AzureBlob': 'AzureBlobDataset', 'AmazonS3Object': 'AmazonS3Dataset'} } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, schema=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: super(Dataset, self).__init__(**kwargs) self.additional_properties = additional_properties self.description = description self.structure = structure + self.schema = schema self.linked_service_name = linked_service_name self.parameters = parameters self.annotations = annotations diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/document_db_collection_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/document_db_collection_dataset.py index 84a9e50bd82d..fb2b8d46fa9c 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/document_db_collection_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/document_db_collection_dataset.py @@ -25,6 +25,10 @@ class DocumentDbCollectionDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -54,6 +58,7 @@ class DocumentDbCollectionDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/document_db_collection_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/document_db_collection_dataset_py3.py index c3566bf9e0f8..5eb4dbbf0997 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/document_db_collection_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/document_db_collection_dataset_py3.py @@ -25,6 +25,10 @@ class DocumentDbCollectionDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -54,6 +58,7 @@ class DocumentDbCollectionDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, @@ -62,7 +67,7 @@ class DocumentDbCollectionDataset(Dataset): 'collection_name': {'key': 'typeProperties.collectionName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, collection_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: - super(DocumentDbCollectionDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + def __init__(self, *, linked_service_name, collection_name, additional_properties=None, description: str=None, structure=None, schema=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + super(DocumentDbCollectionDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.collection_name = collection_name self.type = 'DocumentDbCollection' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_table_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_table_dataset.py index d392331627ad..c12b086b7824 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_table_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_table_dataset.py @@ -25,6 +25,10 @@ class DrillTableDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -53,6 +57,7 @@ class DrillTableDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_table_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_table_dataset_py3.py index bec695cdeaee..f4f5712f29e3 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_table_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_table_dataset_py3.py @@ -25,6 +25,10 @@ class DrillTableDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -53,6 +57,7 @@ class DrillTableDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, @@ -61,7 +66,7 @@ class DrillTableDataset(Dataset): 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: - super(DrillTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, schema=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: + super(DrillTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.table_name = table_name self.type = 'DrillTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_entity_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_entity_dataset.py index 2048d589759c..435c6d153066 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_entity_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_entity_dataset.py @@ -25,6 +25,10 @@ class DynamicsEntityDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -53,6 +57,7 @@ class DynamicsEntityDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_entity_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_entity_dataset_py3.py index e87e511d062a..7ee671890354 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_entity_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_entity_dataset_py3.py @@ -25,6 +25,10 @@ class DynamicsEntityDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -53,6 +57,7 @@ class DynamicsEntityDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, @@ -61,7 +66,7 @@ class DynamicsEntityDataset(Dataset): 'entity_name': {'key': 'typeProperties.entityName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, entity_name=None, **kwargs) -> None: - super(DynamicsEntityDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, schema=None, parameters=None, annotations=None, folder=None, entity_name=None, **kwargs) -> None: + super(DynamicsEntityDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.entity_name = entity_name self.type = 'DynamicsEntity' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/eloqua_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/eloqua_object_dataset.py index d077cd28c5b4..56adc0ce47c4 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/eloqua_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/eloqua_object_dataset.py @@ -25,6 +25,10 @@ class EloquaObjectDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -53,6 +57,7 @@ class EloquaObjectDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/eloqua_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/eloqua_object_dataset_py3.py index 1220f816cbf4..705f43cd225c 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/eloqua_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/eloqua_object_dataset_py3.py @@ -25,6 +25,10 @@ class EloquaObjectDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -53,6 +57,7 @@ class EloquaObjectDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, @@ -61,7 +66,7 @@ class EloquaObjectDataset(Dataset): 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: - super(EloquaObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, schema=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: + super(EloquaObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.table_name = table_name self.type = 'EloquaObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_share_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_share_dataset.py index d5cbf4ac6822..a851956ea319 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_share_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_share_dataset.py @@ -25,6 +25,10 @@ class FileShareDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -64,6 +68,7 @@ class FileShareDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_share_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_share_dataset_py3.py index fe980181543e..675583ae2f2c 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_share_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_share_dataset_py3.py @@ -25,6 +25,10 @@ class FileShareDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -64,6 +68,7 @@ class FileShareDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, @@ -76,8 +81,8 @@ class FileShareDataset(Dataset): 'compression': {'key': 'typeProperties.compression', 'type': 'DatasetCompression'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, folder_path=None, file_name=None, format=None, file_filter=None, compression=None, **kwargs) -> None: - super(FileShareDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, schema=None, parameters=None, annotations=None, folder=None, folder_path=None, file_name=None, format=None, file_filter=None, compression=None, **kwargs) -> None: + super(FileShareDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.folder_path = folder_path self.file_name = file_name self.format = format diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_big_query_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_big_query_object_dataset.py index 6404a7ae5110..5750875dc3a0 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_big_query_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_big_query_object_dataset.py @@ -25,6 +25,10 @@ class GoogleBigQueryObjectDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -53,6 +57,7 @@ class GoogleBigQueryObjectDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_big_query_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_big_query_object_dataset_py3.py index 30ca4d4efe46..625cd068b731 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_big_query_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_big_query_object_dataset_py3.py @@ -25,6 +25,10 @@ class GoogleBigQueryObjectDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -53,6 +57,7 @@ class GoogleBigQueryObjectDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, @@ -61,7 +66,7 @@ class GoogleBigQueryObjectDataset(Dataset): 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: - super(GoogleBigQueryObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, schema=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: + super(GoogleBigQueryObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.table_name = table_name self.type = 'GoogleBigQueryObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_table_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_table_dataset.py index 338042ec56d2..fa4a066f11a9 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_table_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_table_dataset.py @@ -25,6 +25,10 @@ class GreenplumTableDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -53,6 +57,7 @@ class GreenplumTableDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_table_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_table_dataset_py3.py index 1b189fbfdb4b..7c698db22339 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_table_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_table_dataset_py3.py @@ -25,6 +25,10 @@ class GreenplumTableDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -53,6 +57,7 @@ class GreenplumTableDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, @@ -61,7 +66,7 @@ class GreenplumTableDataset(Dataset): 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: - super(GreenplumTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, schema=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: + super(GreenplumTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.table_name = table_name self.type = 'GreenplumTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hbase_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hbase_object_dataset.py index f0f7b5735861..5de32bcb6871 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hbase_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hbase_object_dataset.py @@ -25,6 +25,10 @@ class HBaseObjectDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -53,6 +57,7 @@ class HBaseObjectDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hbase_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hbase_object_dataset_py3.py index d16c7bfba5d1..27fc0d1514ea 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hbase_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hbase_object_dataset_py3.py @@ -25,6 +25,10 @@ class HBaseObjectDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -53,6 +57,7 @@ class HBaseObjectDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, @@ -61,7 +66,7 @@ class HBaseObjectDataset(Dataset): 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: - super(HBaseObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, schema=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: + super(HBaseObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.table_name = table_name self.type = 'HBaseObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hive_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hive_object_dataset.py index 064772f3e32f..7dc4fd367f8a 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hive_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hive_object_dataset.py @@ -25,6 +25,10 @@ class HiveObjectDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -53,6 +57,7 @@ class HiveObjectDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hive_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hive_object_dataset_py3.py index b36690c692a8..c007333721be 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hive_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hive_object_dataset_py3.py @@ -25,6 +25,10 @@ class HiveObjectDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -53,6 +57,7 @@ class HiveObjectDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, @@ -61,7 +66,7 @@ class HiveObjectDataset(Dataset): 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: - super(HiveObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, schema=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: + super(HiveObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.table_name = table_name self.type = 'HiveObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/http_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/http_dataset.py index 532cd654ccef..f2184dea151f 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/http_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/http_dataset.py @@ -25,6 +25,10 @@ class HttpDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -70,6 +74,7 @@ class HttpDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/http_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/http_dataset_py3.py index 6cacff526753..09f97a03a95d 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/http_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/http_dataset_py3.py @@ -25,6 +25,10 @@ class HttpDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -70,6 +74,7 @@ class HttpDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, @@ -83,8 +88,8 @@ class HttpDataset(Dataset): 'compression': {'key': 'typeProperties.compression', 'type': 'DatasetCompression'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, relative_url=None, request_method=None, request_body=None, additional_headers=None, format=None, compression=None, **kwargs) -> None: - super(HttpDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, schema=None, parameters=None, annotations=None, folder=None, relative_url=None, request_method=None, request_body=None, additional_headers=None, format=None, compression=None, **kwargs) -> None: + super(HttpDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.relative_url = relative_url self.request_method = request_method self.request_body = request_body diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_object_dataset.py index 7a54a96e855a..ce8994b4db4a 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_object_dataset.py @@ -25,6 +25,10 @@ class HubspotObjectDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -53,6 +57,7 @@ class HubspotObjectDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_object_dataset_py3.py index c7886d783d95..bd2309101f72 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_object_dataset_py3.py @@ -25,6 +25,10 @@ class HubspotObjectDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -53,6 +57,7 @@ class HubspotObjectDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, @@ -61,7 +66,7 @@ class HubspotObjectDataset(Dataset): 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: - super(HubspotObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, schema=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: + super(HubspotObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.table_name = table_name self.type = 'HubspotObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/impala_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/impala_object_dataset.py index 13a1a548fd59..d9bf591d8021 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/impala_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/impala_object_dataset.py @@ -25,6 +25,10 @@ class ImpalaObjectDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -53,6 +57,7 @@ class ImpalaObjectDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/impala_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/impala_object_dataset_py3.py index c0dc4d490e78..d103603b2586 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/impala_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/impala_object_dataset_py3.py @@ -25,6 +25,10 @@ class ImpalaObjectDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -53,6 +57,7 @@ class ImpalaObjectDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, @@ -61,7 +66,7 @@ class ImpalaObjectDataset(Dataset): 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: - super(ImpalaObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, schema=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: + super(ImpalaObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.table_name = table_name self.type = 'ImpalaObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_object_dataset.py index e197d0d9eb97..1c2b12c18e15 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_object_dataset.py @@ -25,6 +25,10 @@ class JiraObjectDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -53,6 +57,7 @@ class JiraObjectDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_object_dataset_py3.py index 022e88aa0575..3c061b238cde 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_object_dataset_py3.py @@ -25,6 +25,10 @@ class JiraObjectDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -53,6 +57,7 @@ class JiraObjectDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, @@ -61,7 +66,7 @@ class JiraObjectDataset(Dataset): 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: - super(JiraObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, schema=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: + super(JiraObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.table_name = table_name self.type = 'JiraObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/magento_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/magento_object_dataset.py index d0c83370bd47..ad540093ca55 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/magento_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/magento_object_dataset.py @@ -25,6 +25,10 @@ class MagentoObjectDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -53,6 +57,7 @@ class MagentoObjectDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/magento_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/magento_object_dataset_py3.py index 8d95ca580679..481732bb688a 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/magento_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/magento_object_dataset_py3.py @@ -25,6 +25,10 @@ class MagentoObjectDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -53,6 +57,7 @@ class MagentoObjectDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, @@ -61,7 +66,7 @@ class MagentoObjectDataset(Dataset): 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: - super(MagentoObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, schema=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: + super(MagentoObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.table_name = table_name self.type = 'MagentoObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_table_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_table_dataset.py index 58a2b3f4da0e..66dc9c8ea9b7 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_table_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_table_dataset.py @@ -25,6 +25,10 @@ class MariaDBTableDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -53,6 +57,7 @@ class MariaDBTableDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_table_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_table_dataset_py3.py index af0a17cac20f..ac3c8cf2ea72 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_table_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_table_dataset_py3.py @@ -25,6 +25,10 @@ class MariaDBTableDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -53,6 +57,7 @@ class MariaDBTableDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, @@ -61,7 +66,7 @@ class MariaDBTableDataset(Dataset): 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: - super(MariaDBTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, schema=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: + super(MariaDBTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.table_name = table_name self.type = 'MariaDBTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/marketo_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/marketo_object_dataset.py index dba47a291f96..63daa10047b9 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/marketo_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/marketo_object_dataset.py @@ -25,6 +25,10 @@ class MarketoObjectDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -53,6 +57,7 @@ class MarketoObjectDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/marketo_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/marketo_object_dataset_py3.py index 1322fee85893..7179d5af53dd 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/marketo_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/marketo_object_dataset_py3.py @@ -25,6 +25,10 @@ class MarketoObjectDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -53,6 +57,7 @@ class MarketoObjectDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, @@ -61,7 +66,7 @@ class MarketoObjectDataset(Dataset): 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: - super(MarketoObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, schema=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: + super(MarketoObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.table_name = table_name self.type = 'MarketoObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/mongo_db_collection_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/mongo_db_collection_dataset.py index 733d6454492c..796c5e14eaca 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/mongo_db_collection_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/mongo_db_collection_dataset.py @@ -25,6 +25,10 @@ class MongoDbCollectionDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -54,6 +58,7 @@ class MongoDbCollectionDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/mongo_db_collection_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/mongo_db_collection_dataset_py3.py index 7b9de8741d64..68fe2affb0e4 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/mongo_db_collection_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/mongo_db_collection_dataset_py3.py @@ -25,6 +25,10 @@ class MongoDbCollectionDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -54,6 +58,7 @@ class MongoDbCollectionDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, @@ -62,7 +67,7 @@ class MongoDbCollectionDataset(Dataset): 'collection_name': {'key': 'typeProperties.collectionName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, collection_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: - super(MongoDbCollectionDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + def __init__(self, *, linked_service_name, collection_name, additional_properties=None, description: str=None, structure=None, schema=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + super(MongoDbCollectionDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.collection_name = collection_name self.type = 'MongoDbCollection' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/netezza_table_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/netezza_table_dataset.py index 8780abac2f6c..cf3b9205846c 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/netezza_table_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/netezza_table_dataset.py @@ -25,6 +25,10 @@ class NetezzaTableDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -53,6 +57,7 @@ class NetezzaTableDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/netezza_table_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/netezza_table_dataset_py3.py index 1bc693017215..39de0032e8c9 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/netezza_table_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/netezza_table_dataset_py3.py @@ -25,6 +25,10 @@ class NetezzaTableDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -53,6 +57,7 @@ class NetezzaTableDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, @@ -61,7 +66,7 @@ class NetezzaTableDataset(Dataset): 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: - super(NetezzaTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, schema=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: + super(NetezzaTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.table_name = table_name self.type = 'NetezzaTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odata_resource_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odata_resource_dataset.py index e1c429b6ad62..658cf40c8d2b 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odata_resource_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odata_resource_dataset.py @@ -25,6 +25,10 @@ class ODataResourceDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -53,6 +57,7 @@ class ODataResourceDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odata_resource_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odata_resource_dataset_py3.py index ea103df22f30..5951a2cf6d80 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odata_resource_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/odata_resource_dataset_py3.py @@ -25,6 +25,10 @@ class ODataResourceDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -53,6 +57,7 @@ class ODataResourceDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, @@ -61,7 +66,7 @@ class ODataResourceDataset(Dataset): 'path': {'key': 'typeProperties.path', 'type': 'object'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, path=None, **kwargs) -> None: - super(ODataResourceDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, schema=None, parameters=None, annotations=None, folder=None, path=None, **kwargs) -> None: + super(ODataResourceDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.path = path self.type = 'ODataResource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_table_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_table_dataset.py index 7362e0bd42cf..4af8faaca8db 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_table_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_table_dataset.py @@ -25,6 +25,10 @@ class OracleTableDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -54,6 +58,7 @@ class OracleTableDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_table_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_table_dataset_py3.py index af1e9443e1cb..aaa1291c8f76 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_table_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_table_dataset_py3.py @@ -25,6 +25,10 @@ class OracleTableDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -54,6 +58,7 @@ class OracleTableDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, @@ -62,7 +67,7 @@ class OracleTableDataset(Dataset): 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, table_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: - super(OracleTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + def __init__(self, *, linked_service_name, table_name, additional_properties=None, description: str=None, structure=None, schema=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + super(OracleTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.table_name = table_name self.type = 'OracleTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_object_dataset.py index 69ac88775d42..d0fdc678841b 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_object_dataset.py @@ -25,6 +25,10 @@ class PaypalObjectDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -53,6 +57,7 @@ class PaypalObjectDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_object_dataset_py3.py index 080e482cf978..55df7c97166d 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_object_dataset_py3.py @@ -25,6 +25,10 @@ class PaypalObjectDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -53,6 +57,7 @@ class PaypalObjectDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, @@ -61,7 +66,7 @@ class PaypalObjectDataset(Dataset): 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: - super(PaypalObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, schema=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: + super(PaypalObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.table_name = table_name self.type = 'PaypalObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/phoenix_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/phoenix_object_dataset.py index 44b271aeb1dd..2d9cd5dcd581 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/phoenix_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/phoenix_object_dataset.py @@ -25,6 +25,10 @@ class PhoenixObjectDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -53,6 +57,7 @@ class PhoenixObjectDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/phoenix_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/phoenix_object_dataset_py3.py index 1f9f6e66dec2..32c6e5f9836f 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/phoenix_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/phoenix_object_dataset_py3.py @@ -25,6 +25,10 @@ class PhoenixObjectDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -53,6 +57,7 @@ class PhoenixObjectDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, @@ -61,7 +66,7 @@ class PhoenixObjectDataset(Dataset): 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: - super(PhoenixObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, schema=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: + super(PhoenixObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.table_name = table_name self.type = 'PhoenixObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/presto_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/presto_object_dataset.py index f94090dbd359..35ceaa1389a7 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/presto_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/presto_object_dataset.py @@ -25,6 +25,10 @@ class PrestoObjectDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -53,6 +57,7 @@ class PrestoObjectDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/presto_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/presto_object_dataset_py3.py index c65b6d6d6520..193004e2c381 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/presto_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/presto_object_dataset_py3.py @@ -25,6 +25,10 @@ class PrestoObjectDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -53,6 +57,7 @@ class PrestoObjectDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, @@ -61,7 +66,7 @@ class PrestoObjectDataset(Dataset): 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: - super(PrestoObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, schema=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: + super(PrestoObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.table_name = table_name self.type = 'PrestoObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_object_dataset.py index 1653ee62d028..73446d0ed938 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_object_dataset.py @@ -25,6 +25,10 @@ class QuickBooksObjectDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -53,6 +57,7 @@ class QuickBooksObjectDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_object_dataset_py3.py index 9273418585b8..65f67d2b20af 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_object_dataset_py3.py @@ -25,6 +25,10 @@ class QuickBooksObjectDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -53,6 +57,7 @@ class QuickBooksObjectDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, @@ -61,7 +66,7 @@ class QuickBooksObjectDataset(Dataset): 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: - super(QuickBooksObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, schema=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: + super(QuickBooksObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.table_name = table_name self.type = 'QuickBooksObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/relational_table_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/relational_table_dataset.py index b0c91815e554..e5dd2e0786c8 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/relational_table_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/relational_table_dataset.py @@ -25,6 +25,10 @@ class RelationalTableDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -53,6 +57,7 @@ class RelationalTableDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/relational_table_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/relational_table_dataset_py3.py index 638b0b8a3eef..3c85d95f8033 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/relational_table_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/relational_table_dataset_py3.py @@ -25,6 +25,10 @@ class RelationalTableDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -53,6 +57,7 @@ class RelationalTableDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, @@ -61,7 +66,7 @@ class RelationalTableDataset(Dataset): 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: - super(RelationalTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, schema=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: + super(RelationalTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.table_name = table_name self.type = 'RelationalTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_object_dataset.py index 78b2e6859f7e..f459e69113a1 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_object_dataset.py @@ -25,6 +25,10 @@ class ResponsysObjectDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -53,6 +57,7 @@ class ResponsysObjectDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_object_dataset_py3.py index 59d3e2d001d4..c5f375910aaf 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/responsys_object_dataset_py3.py @@ -25,6 +25,10 @@ class ResponsysObjectDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -53,6 +57,7 @@ class ResponsysObjectDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, @@ -61,7 +66,7 @@ class ResponsysObjectDataset(Dataset): 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: - super(ResponsysObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, schema=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: + super(ResponsysObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.table_name = table_name self.type = 'ResponsysObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_marketing_cloud_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_marketing_cloud_object_dataset.py index 3dfba59c2dab..20f581ce1c50 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_marketing_cloud_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_marketing_cloud_object_dataset.py @@ -25,6 +25,10 @@ class SalesforceMarketingCloudObjectDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -53,6 +57,7 @@ class SalesforceMarketingCloudObjectDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_marketing_cloud_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_marketing_cloud_object_dataset_py3.py index 31490518a869..526ac806649f 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_marketing_cloud_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_marketing_cloud_object_dataset_py3.py @@ -25,6 +25,10 @@ class SalesforceMarketingCloudObjectDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -53,6 +57,7 @@ class SalesforceMarketingCloudObjectDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, @@ -61,7 +66,7 @@ class SalesforceMarketingCloudObjectDataset(Dataset): 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: - super(SalesforceMarketingCloudObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, schema=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: + super(SalesforceMarketingCloudObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.table_name = table_name self.type = 'SalesforceMarketingCloudObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_object_dataset.py index a55974c6428c..10cfce97fe0f 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_object_dataset.py @@ -25,6 +25,10 @@ class SalesforceObjectDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -53,6 +57,7 @@ class SalesforceObjectDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_object_dataset_py3.py index fa55f29420dd..3c3f75d6059e 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_object_dataset_py3.py @@ -25,6 +25,10 @@ class SalesforceObjectDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -53,6 +57,7 @@ class SalesforceObjectDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, @@ -61,7 +66,7 @@ class SalesforceObjectDataset(Dataset): 'object_api_name': {'key': 'typeProperties.objectApiName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, object_api_name=None, **kwargs) -> None: - super(SalesforceObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, schema=None, parameters=None, annotations=None, folder=None, object_api_name=None, **kwargs) -> None: + super(SalesforceObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.object_api_name = object_api_name self.type = 'SalesforceObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_resource_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_resource_dataset.py index b975a3c99cc1..436b251207a4 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_resource_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_resource_dataset.py @@ -25,6 +25,10 @@ class SapCloudForCustomerResourceDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -54,6 +58,7 @@ class SapCloudForCustomerResourceDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_resource_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_resource_dataset_py3.py index 09ab4693c1fb..455bad7c9095 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_resource_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_resource_dataset_py3.py @@ -25,6 +25,10 @@ class SapCloudForCustomerResourceDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -54,6 +58,7 @@ class SapCloudForCustomerResourceDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, @@ -62,7 +67,7 @@ class SapCloudForCustomerResourceDataset(Dataset): 'path': {'key': 'typeProperties.path', 'type': 'object'}, } - def __init__(self, *, linked_service_name, path, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: - super(SapCloudForCustomerResourceDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + def __init__(self, *, linked_service_name, path, additional_properties=None, description: str=None, structure=None, schema=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + super(SapCloudForCustomerResourceDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.path = path self.type = 'SapCloudForCustomerResource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_ecc_resource_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_ecc_resource_dataset.py index f26232ea34f6..e4f10113aecd 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_ecc_resource_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_ecc_resource_dataset.py @@ -25,6 +25,10 @@ class SapEccResourceDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -54,6 +58,7 @@ class SapEccResourceDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_ecc_resource_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_ecc_resource_dataset_py3.py index aced6ccc0f9e..08bf742dc415 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_ecc_resource_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_ecc_resource_dataset_py3.py @@ -25,6 +25,10 @@ class SapEccResourceDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -54,6 +58,7 @@ class SapEccResourceDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, @@ -62,7 +67,7 @@ class SapEccResourceDataset(Dataset): 'path': {'key': 'typeProperties.path', 'type': 'str'}, } - def __init__(self, *, linked_service_name, path: str, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: - super(SapEccResourceDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + def __init__(self, *, linked_service_name, path: str, additional_properties=None, description: str=None, structure=None, schema=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + super(SapEccResourceDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.path = path self.type = 'SapEccResource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_object_dataset.py index 96b86b1f8f9c..a9821ba0fd10 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_object_dataset.py @@ -25,6 +25,10 @@ class ServiceNowObjectDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -53,6 +57,7 @@ class ServiceNowObjectDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_object_dataset_py3.py index 54c7f0225a0f..fcd2fd537a31 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_object_dataset_py3.py @@ -25,6 +25,10 @@ class ServiceNowObjectDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -53,6 +57,7 @@ class ServiceNowObjectDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, @@ -61,7 +66,7 @@ class ServiceNowObjectDataset(Dataset): 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: - super(ServiceNowObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, schema=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: + super(ServiceNowObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.table_name = table_name self.type = 'ServiceNowObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_object_dataset.py index 73569a397edc..ab3e475b9c97 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_object_dataset.py @@ -25,6 +25,10 @@ class ShopifyObjectDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -53,6 +57,7 @@ class ShopifyObjectDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_object_dataset_py3.py index ed18a9f34769..98b9c43c21e8 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_object_dataset_py3.py @@ -25,6 +25,10 @@ class ShopifyObjectDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -53,6 +57,7 @@ class ShopifyObjectDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, @@ -61,7 +66,7 @@ class ShopifyObjectDataset(Dataset): 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: - super(ShopifyObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, schema=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: + super(ShopifyObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.table_name = table_name self.type = 'ShopifyObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/spark_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/spark_object_dataset.py index a33390e391fe..8d1493ea9c7f 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/spark_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/spark_object_dataset.py @@ -25,6 +25,10 @@ class SparkObjectDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -53,6 +57,7 @@ class SparkObjectDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/spark_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/spark_object_dataset_py3.py index de4736d96ead..3ab167dd3540 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/spark_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/spark_object_dataset_py3.py @@ -25,6 +25,10 @@ class SparkObjectDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -53,6 +57,7 @@ class SparkObjectDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, @@ -61,7 +66,7 @@ class SparkObjectDataset(Dataset): 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: - super(SparkObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, schema=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: + super(SparkObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.table_name = table_name self.type = 'SparkObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_table_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_table_dataset.py index 99afdbf99ce0..c0bc0b66a5e2 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_table_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_table_dataset.py @@ -25,6 +25,10 @@ class SqlServerTableDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -54,6 +58,7 @@ class SqlServerTableDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_table_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_table_dataset_py3.py index 9567fc1e6bb8..0fb8d10ea111 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_table_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sql_server_table_dataset_py3.py @@ -25,6 +25,10 @@ class SqlServerTableDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -54,6 +58,7 @@ class SqlServerTableDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, @@ -62,7 +67,7 @@ class SqlServerTableDataset(Dataset): 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, table_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: - super(SqlServerTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + def __init__(self, *, linked_service_name, table_name, additional_properties=None, description: str=None, structure=None, schema=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: + super(SqlServerTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.table_name = table_name self.type = 'SqlServerTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_object_dataset.py index 515bab8bbff0..3903382d2e3a 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_object_dataset.py @@ -25,6 +25,10 @@ class SquareObjectDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -53,6 +57,7 @@ class SquareObjectDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_object_dataset_py3.py index 730f11108723..6d624dc6feef 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_object_dataset_py3.py @@ -25,6 +25,10 @@ class SquareObjectDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -53,6 +57,7 @@ class SquareObjectDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, @@ -61,7 +66,7 @@ class SquareObjectDataset(Dataset): 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: - super(SquareObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, schema=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: + super(SquareObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.table_name = table_name self.type = 'SquareObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/vertica_table_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/vertica_table_dataset.py index 8ef0d1a07d99..e84465f8ba07 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/vertica_table_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/vertica_table_dataset.py @@ -25,6 +25,10 @@ class VerticaTableDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -53,6 +57,7 @@ class VerticaTableDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/vertica_table_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/vertica_table_dataset_py3.py index 82be33c77831..87d69bb9a443 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/vertica_table_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/vertica_table_dataset_py3.py @@ -25,6 +25,10 @@ class VerticaTableDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -53,6 +57,7 @@ class VerticaTableDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, @@ -61,7 +66,7 @@ class VerticaTableDataset(Dataset): 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: - super(VerticaTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, schema=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: + super(VerticaTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.table_name = table_name self.type = 'VerticaTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_table_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_table_dataset.py index a9b30245d2da..3980fe3d885a 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_table_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_table_dataset.py @@ -25,6 +25,10 @@ class WebTableDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -57,6 +61,7 @@ class WebTableDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_table_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_table_dataset_py3.py index d5ea9d5f66ba..edb2344c35d2 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_table_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_table_dataset_py3.py @@ -25,6 +25,10 @@ class WebTableDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -57,6 +61,7 @@ class WebTableDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, @@ -66,8 +71,8 @@ class WebTableDataset(Dataset): 'path': {'key': 'typeProperties.path', 'type': 'object'}, } - def __init__(self, *, linked_service_name, index, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, path=None, **kwargs) -> None: - super(WebTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + def __init__(self, *, linked_service_name, index, additional_properties=None, description: str=None, structure=None, schema=None, parameters=None, annotations=None, folder=None, path=None, **kwargs) -> None: + super(WebTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.index = index self.path = path self.type = 'WebTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_object_dataset.py index 8623618a99d8..53c5edd44cec 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_object_dataset.py @@ -25,6 +25,10 @@ class XeroObjectDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -53,6 +57,7 @@ class XeroObjectDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_object_dataset_py3.py index b23685dda4ca..673d41e1771e 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_object_dataset_py3.py @@ -25,6 +25,10 @@ class XeroObjectDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -53,6 +57,7 @@ class XeroObjectDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, @@ -61,7 +66,7 @@ class XeroObjectDataset(Dataset): 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: - super(XeroObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, schema=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: + super(XeroObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.table_name = table_name self.type = 'XeroObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/zoho_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/zoho_object_dataset.py index e2df390d74ea..062d508860a6 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/zoho_object_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/zoho_object_dataset.py @@ -25,6 +25,10 @@ class ZohoObjectDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -53,6 +57,7 @@ class ZohoObjectDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/zoho_object_dataset_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/zoho_object_dataset_py3.py index 465e922ecd34..ef5a67d4fe35 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/zoho_object_dataset_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/zoho_object_dataset_py3.py @@ -25,6 +25,10 @@ class ZohoObjectDataset(Dataset): :param structure: Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. :type structure: object + :param schema: Columns that define the physical type schema of the + dataset. Type: array (or Expression with resultType array), itemType: + DatasetSchemaDataElement. + :type schema: object :param linked_service_name: Required. Linked service reference. :type linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference @@ -53,6 +57,7 @@ class ZohoObjectDataset(Dataset): 'additional_properties': {'key': '', 'type': '{object}'}, 'description': {'key': 'description', 'type': 'str'}, 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, @@ -61,7 +66,7 @@ class ZohoObjectDataset(Dataset): 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, } - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: - super(ZohoObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, schema=None, parameters=None, annotations=None, folder=None, table_name=None, **kwargs) -> None: + super(ZohoObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) self.table_name = table_name self.type = 'ZohoObject' From ddd2d4b85d69c9ec088ac7af0ab76b722a3066f9 Mon Sep 17 00:00:00 2001 From: Azure SDK for Python bot Date: Tue, 15 Jan 2019 13:06:30 -0800 Subject: [PATCH 43/54] Generated from 795e21a260c01ba1f6be0fd1699dc0cbe5150346 (#4185) For preview: Change type of Ssis actitvity's packagePath, runtime, loggingLevel and environmentPath from string to object, so that we can use expression for these properties# Please enter the commit message for your changes. Lines starting --- .../azure/mgmt/datafactory/models/__init__.py | 3 + .../datafactory/models/control_activity.py | 7 +- .../models/control_activity_py3.py | 7 +- .../datafactory/models/validation_activity.py | 77 +++++++++++++++++++ .../models/validation_activity_py3.py | 77 +++++++++++++++++++ 5 files changed, 165 insertions(+), 6 deletions(-) create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/validation_activity.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/validation_activity_py3.py diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py index 144d6e2998ae..058405862120 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py @@ -339,6 +339,7 @@ from .append_variable_activity_py3 import AppendVariableActivity from .set_variable_activity_py3 import SetVariableActivity from .filter_activity_py3 import FilterActivity + from .validation_activity_py3 import ValidationActivity from .until_activity_py3 import UntilActivity from .wait_activity_py3 import WaitActivity from .for_each_activity_py3 import ForEachActivity @@ -701,6 +702,7 @@ from .append_variable_activity import AppendVariableActivity from .set_variable_activity import SetVariableActivity from .filter_activity import FilterActivity + from .validation_activity import ValidationActivity from .until_activity import UntilActivity from .wait_activity import WaitActivity from .for_each_activity import ForEachActivity @@ -1134,6 +1136,7 @@ 'AppendVariableActivity', 'SetVariableActivity', 'FilterActivity', + 'ValidationActivity', 'UntilActivity', 'WaitActivity', 'ForEachActivity', diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/control_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/control_activity.py index e0f9592f9759..5d64648774da 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/control_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/control_activity.py @@ -17,8 +17,9 @@ class ControlActivity(Activity): You probably want to use the sub-classes and not this class directly. Known sub-classes are: AppendVariableActivity, SetVariableActivity, - FilterActivity, UntilActivity, WaitActivity, ForEachActivity, - IfConditionActivity, WebHookActivity, ExecutePipelineActivity + FilterActivity, ValidationActivity, UntilActivity, WaitActivity, + ForEachActivity, IfConditionActivity, WebHookActivity, + ExecutePipelineActivity All required parameters must be populated in order to send to Azure. @@ -52,7 +53,7 @@ class ControlActivity(Activity): } _subtype_map = { - 'type': {'AppendVariable': 'AppendVariableActivity', 'SetVariable': 'SetVariableActivity', 'Filter': 'FilterActivity', 'Until': 'UntilActivity', 'Wait': 'WaitActivity', 'ForEach': 'ForEachActivity', 'IfCondition': 'IfConditionActivity', 'WebHook': 'WebHookActivity', 'ExecutePipeline': 'ExecutePipelineActivity'} + 'type': {'AppendVariable': 'AppendVariableActivity', 'SetVariable': 'SetVariableActivity', 'Filter': 'FilterActivity', 'Validation': 'ValidationActivity', 'Until': 'UntilActivity', 'Wait': 'WaitActivity', 'ForEach': 'ForEachActivity', 'IfCondition': 'IfConditionActivity', 'WebHook': 'WebHookActivity', 'ExecutePipeline': 'ExecutePipelineActivity'} } def __init__(self, **kwargs): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/control_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/control_activity_py3.py index 8021c7a34c14..be9282168ac1 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/control_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/control_activity_py3.py @@ -17,8 +17,9 @@ class ControlActivity(Activity): You probably want to use the sub-classes and not this class directly. Known sub-classes are: AppendVariableActivity, SetVariableActivity, - FilterActivity, UntilActivity, WaitActivity, ForEachActivity, - IfConditionActivity, WebHookActivity, ExecutePipelineActivity + FilterActivity, ValidationActivity, UntilActivity, WaitActivity, + ForEachActivity, IfConditionActivity, WebHookActivity, + ExecutePipelineActivity All required parameters must be populated in order to send to Azure. @@ -52,7 +53,7 @@ class ControlActivity(Activity): } _subtype_map = { - 'type': {'AppendVariable': 'AppendVariableActivity', 'SetVariable': 'SetVariableActivity', 'Filter': 'FilterActivity', 'Until': 'UntilActivity', 'Wait': 'WaitActivity', 'ForEach': 'ForEachActivity', 'IfCondition': 'IfConditionActivity', 'WebHook': 'WebHookActivity', 'ExecutePipeline': 'ExecutePipelineActivity'} + 'type': {'AppendVariable': 'AppendVariableActivity', 'SetVariable': 'SetVariableActivity', 'Filter': 'FilterActivity', 'Validation': 'ValidationActivity', 'Until': 'UntilActivity', 'Wait': 'WaitActivity', 'ForEach': 'ForEachActivity', 'IfCondition': 'IfConditionActivity', 'WebHook': 'WebHookActivity', 'ExecutePipeline': 'ExecutePipelineActivity'} } def __init__(self, *, name: str, additional_properties=None, description: str=None, depends_on=None, user_properties=None, **kwargs) -> None: diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/validation_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/validation_activity.py new file mode 100644 index 000000000000..6e20563ffc97 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/validation_activity.py @@ -0,0 +1,77 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .control_activity import ControlActivity + + +class ValidationActivity(ControlActivity): + """This activity blocks execution until a file has been validated to exist, + with an optional minimum size, or the timeout is reached, whichever is + earlier. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param name: Required. Activity name. + :type name: str + :param description: Activity description. + :type description: str + :param depends_on: Activity depends on condition. + :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param user_properties: Activity user properties. + :type user_properties: list[~azure.mgmt.datafactory.models.UserProperty] + :param type: Required. Constant filled by server. + :type type: str + :param timeout: Specifies the timeout for the activity to run. If there is + no value specified, it takes the value of TimeSpan.FromDays(7) which is 1 + week as default. Type: string (or Expression with resultType string), + pattern: ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). Type: + string (or Expression with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type timeout: object + :param sleep: A delay in seconds between validation attempts. If no value + is specified, 10 seconds will be used as the default. + :type sleep: int + :param minimum_size: Minimum size of a file in byte. If no value is + specified, 0 byte will be used as the default. + :type minimum_size: int + :param dataset: Required. Validation activity dataset reference. + :type dataset: ~azure.mgmt.datafactory.models.DatasetReference + """ + + _validation = { + 'name': {'required': True}, + 'type': {'required': True}, + 'dataset': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'name': {'key': 'name', 'type': 'str'}, + 'description': {'key': 'description', 'type': 'str'}, + 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'timeout': {'key': 'typeProperties.timeout', 'type': 'object'}, + 'sleep': {'key': 'typeProperties.sleep', 'type': 'int'}, + 'minimum_size': {'key': 'typeProperties.minimumSize', 'type': 'int'}, + 'dataset': {'key': 'typeProperties.dataset', 'type': 'DatasetReference'}, + } + + def __init__(self, **kwargs): + super(ValidationActivity, self).__init__(**kwargs) + self.timeout = kwargs.get('timeout', None) + self.sleep = kwargs.get('sleep', None) + self.minimum_size = kwargs.get('minimum_size', None) + self.dataset = kwargs.get('dataset', None) + self.type = 'Validation' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/validation_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/validation_activity_py3.py new file mode 100644 index 000000000000..997b5fb4c6e7 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/validation_activity_py3.py @@ -0,0 +1,77 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .control_activity_py3 import ControlActivity + + +class ValidationActivity(ControlActivity): + """This activity blocks execution until a file has been validated to exist, + with an optional minimum size, or the timeout is reached, whichever is + earlier. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param name: Required. Activity name. + :type name: str + :param description: Activity description. + :type description: str + :param depends_on: Activity depends on condition. + :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param user_properties: Activity user properties. + :type user_properties: list[~azure.mgmt.datafactory.models.UserProperty] + :param type: Required. Constant filled by server. + :type type: str + :param timeout: Specifies the timeout for the activity to run. If there is + no value specified, it takes the value of TimeSpan.FromDays(7) which is 1 + week as default. Type: string (or Expression with resultType string), + pattern: ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). Type: + string (or Expression with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type timeout: object + :param sleep: A delay in seconds between validation attempts. If no value + is specified, 10 seconds will be used as the default. + :type sleep: int + :param minimum_size: Minimum size of a file in byte. If no value is + specified, 0 byte will be used as the default. + :type minimum_size: int + :param dataset: Required. Validation activity dataset reference. + :type dataset: ~azure.mgmt.datafactory.models.DatasetReference + """ + + _validation = { + 'name': {'required': True}, + 'type': {'required': True}, + 'dataset': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'name': {'key': 'name', 'type': 'str'}, + 'description': {'key': 'description', 'type': 'str'}, + 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'timeout': {'key': 'typeProperties.timeout', 'type': 'object'}, + 'sleep': {'key': 'typeProperties.sleep', 'type': 'int'}, + 'minimum_size': {'key': 'typeProperties.minimumSize', 'type': 'int'}, + 'dataset': {'key': 'typeProperties.dataset', 'type': 'DatasetReference'}, + } + + def __init__(self, *, name: str, dataset, additional_properties=None, description: str=None, depends_on=None, user_properties=None, timeout=None, sleep: int=None, minimum_size: int=None, **kwargs) -> None: + super(ValidationActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, **kwargs) + self.timeout = timeout + self.sleep = sleep + self.minimum_size = minimum_size + self.dataset = dataset + self.type = 'Validation' From c90bc9063f0115e8c98853bc21499745028176ac Mon Sep 17 00:00:00 2001 From: Azure SDK for Python bot Date: Wed, 16 Jan 2019 07:36:12 -0800 Subject: [PATCH 44/54] [AutoPR datafactory/resource-manager] Add exposure control APIs according to the compliance request. (#4105) * Generated from f81ef7f40910a3ecb87c8b6cce9aa7bb919344fd Add exposure control APIs. * Generated from 741fceb5f54f36ac55da1e93567549588c5cdb69 Resolve reivew comments * Generated from 0c64157708f71a189f8e9726dff0959e63440ed0 Delete API ExposureControl_GetFeatureBySubscription * Generated from 32f235ca299351d02b27f0546cb1f06d02125b63 Resolve review comments --- .../data_factory_management_client.py | 5 + .../azure/mgmt/datafactory/models/__init__.py | 6 + .../models/exposure_control_request.py | 32 ++++++ .../models/exposure_control_request_py3.py | 32 ++++++ .../models/exposure_control_response.py | 40 +++++++ .../models/exposure_control_response_py3.py | 40 +++++++ .../mgmt/datafactory/operations/__init__.py | 2 + .../operations/exposure_control_operations.py | 107 ++++++++++++++++++ 8 files changed, 264 insertions(+) create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/exposure_control_request.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/exposure_control_request_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/exposure_control_response.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/exposure_control_response_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/operations/exposure_control_operations.py diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/data_factory_management_client.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/data_factory_management_client.py index bfb43b111214..e49abccce72a 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/data_factory_management_client.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/data_factory_management_client.py @@ -15,6 +15,7 @@ from .version import VERSION from .operations.operations import Operations from .operations.factories_operations import FactoriesOperations +from .operations.exposure_control_operations import ExposureControlOperations from .operations.integration_runtimes_operations import IntegrationRuntimesOperations from .operations.integration_runtime_object_metadata_operations import IntegrationRuntimeObjectMetadataOperations from .operations.integration_runtime_nodes_operations import IntegrationRuntimeNodesOperations @@ -71,6 +72,8 @@ class DataFactoryManagementClient(SDKClient): :vartype operations: azure.mgmt.datafactory.operations.Operations :ivar factories: Factories operations :vartype factories: azure.mgmt.datafactory.operations.FactoriesOperations + :ivar exposure_control: ExposureControl operations + :vartype exposure_control: azure.mgmt.datafactory.operations.ExposureControlOperations :ivar integration_runtimes: IntegrationRuntimes operations :vartype integration_runtimes: azure.mgmt.datafactory.operations.IntegrationRuntimesOperations :ivar integration_runtime_object_metadata: IntegrationRuntimeObjectMetadata operations @@ -117,6 +120,8 @@ def __init__( self._client, self.config, self._serialize, self._deserialize) self.factories = FactoriesOperations( self._client, self.config, self._serialize, self._deserialize) + self.exposure_control = ExposureControlOperations( + self._client, self.config, self._serialize, self._deserialize) self.integration_runtimes = IntegrationRuntimesOperations( self._client, self.config, self._serialize, self._deserialize) self.integration_runtime_object_metadata = IntegrationRuntimeObjectMetadataOperations( diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py index 058405862120..9c3893952ab6 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py @@ -78,6 +78,8 @@ from .operation_py3 import Operation from .get_ssis_object_metadata_request_py3 import GetSsisObjectMetadataRequest from .ssis_object_metadata_status_response_py3 import SsisObjectMetadataStatusResponse + from .exposure_control_request_py3 import ExposureControlRequest + from .exposure_control_response_py3 import ExposureControlResponse from .self_dependency_tumbling_window_trigger_reference_py3 import SelfDependencyTumblingWindowTriggerReference from .trigger_reference_py3 import TriggerReference from .tumbling_window_trigger_dependency_reference_py3 import TumblingWindowTriggerDependencyReference @@ -441,6 +443,8 @@ from .operation import Operation from .get_ssis_object_metadata_request import GetSsisObjectMetadataRequest from .ssis_object_metadata_status_response import SsisObjectMetadataStatusResponse + from .exposure_control_request import ExposureControlRequest + from .exposure_control_response import ExposureControlResponse from .self_dependency_tumbling_window_trigger_reference import SelfDependencyTumblingWindowTriggerReference from .trigger_reference import TriggerReference from .tumbling_window_trigger_dependency_reference import TumblingWindowTriggerDependencyReference @@ -875,6 +879,8 @@ 'Operation', 'GetSsisObjectMetadataRequest', 'SsisObjectMetadataStatusResponse', + 'ExposureControlRequest', + 'ExposureControlResponse', 'SelfDependencyTumblingWindowTriggerReference', 'TriggerReference', 'TumblingWindowTriggerDependencyReference', diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/exposure_control_request.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/exposure_control_request.py new file mode 100644 index 000000000000..a6a2cc280b4d --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/exposure_control_request.py @@ -0,0 +1,32 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class ExposureControlRequest(Model): + """The exposure control request. + + :param feature_name: The feature name. + :type feature_name: str + :param feature_type: The feature type. + :type feature_type: str + """ + + _attribute_map = { + 'feature_name': {'key': 'featureName', 'type': 'str'}, + 'feature_type': {'key': 'featureType', 'type': 'str'}, + } + + def __init__(self, **kwargs): + super(ExposureControlRequest, self).__init__(**kwargs) + self.feature_name = kwargs.get('feature_name', None) + self.feature_type = kwargs.get('feature_type', None) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/exposure_control_request_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/exposure_control_request_py3.py new file mode 100644 index 000000000000..b3f4099fb972 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/exposure_control_request_py3.py @@ -0,0 +1,32 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class ExposureControlRequest(Model): + """The exposure control request. + + :param feature_name: The feature name. + :type feature_name: str + :param feature_type: The feature type. + :type feature_type: str + """ + + _attribute_map = { + 'feature_name': {'key': 'featureName', 'type': 'str'}, + 'feature_type': {'key': 'featureType', 'type': 'str'}, + } + + def __init__(self, *, feature_name: str=None, feature_type: str=None, **kwargs) -> None: + super(ExposureControlRequest, self).__init__(**kwargs) + self.feature_name = feature_name + self.feature_type = feature_type diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/exposure_control_response.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/exposure_control_response.py new file mode 100644 index 000000000000..868647e3c5b3 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/exposure_control_response.py @@ -0,0 +1,40 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class ExposureControlResponse(Model): + """The exposure control response. + + Variables are only populated by the server, and will be ignored when + sending a request. + + :ivar feature_name: The feature name. + :vartype feature_name: str + :ivar value: The feature value. + :vartype value: str + """ + + _validation = { + 'feature_name': {'readonly': True}, + 'value': {'readonly': True}, + } + + _attribute_map = { + 'feature_name': {'key': 'featureName', 'type': 'str'}, + 'value': {'key': 'value', 'type': 'str'}, + } + + def __init__(self, **kwargs): + super(ExposureControlResponse, self).__init__(**kwargs) + self.feature_name = None + self.value = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/exposure_control_response_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/exposure_control_response_py3.py new file mode 100644 index 000000000000..1ac7138e7984 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/exposure_control_response_py3.py @@ -0,0 +1,40 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class ExposureControlResponse(Model): + """The exposure control response. + + Variables are only populated by the server, and will be ignored when + sending a request. + + :ivar feature_name: The feature name. + :vartype feature_name: str + :ivar value: The feature value. + :vartype value: str + """ + + _validation = { + 'feature_name': {'readonly': True}, + 'value': {'readonly': True}, + } + + _attribute_map = { + 'feature_name': {'key': 'featureName', 'type': 'str'}, + 'value': {'key': 'value', 'type': 'str'}, + } + + def __init__(self, **kwargs) -> None: + super(ExposureControlResponse, self).__init__(**kwargs) + self.feature_name = None + self.value = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/__init__.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/__init__.py index c0dccd86b56a..ffc98f67bed2 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/__init__.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/__init__.py @@ -11,6 +11,7 @@ from .operations import Operations from .factories_operations import FactoriesOperations +from .exposure_control_operations import ExposureControlOperations from .integration_runtimes_operations import IntegrationRuntimesOperations from .integration_runtime_object_metadata_operations import IntegrationRuntimeObjectMetadataOperations from .integration_runtime_nodes_operations import IntegrationRuntimeNodesOperations @@ -26,6 +27,7 @@ __all__ = [ 'Operations', 'FactoriesOperations', + 'ExposureControlOperations', 'IntegrationRuntimesOperations', 'IntegrationRuntimeObjectMetadataOperations', 'IntegrationRuntimeNodesOperations', diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/exposure_control_operations.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/exposure_control_operations.py new file mode 100644 index 000000000000..080e8c87ba18 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/exposure_control_operations.py @@ -0,0 +1,107 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +import uuid +from msrest.pipeline import ClientRawResponse +from msrestazure.azure_exceptions import CloudError + +from .. import models + + +class ExposureControlOperations(object): + """ExposureControlOperations operations. + + :param client: Client for service requests. + :param config: Configuration of service client. + :param serializer: An object model serializer. + :param deserializer: An object model deserializer. + :ivar api_version: The API version. Constant value: "2018-06-01". + """ + + models = models + + def __init__(self, client, config, serializer, deserializer): + + self._client = client + self._serialize = serializer + self._deserialize = deserializer + self.api_version = "2018-06-01" + + self.config = config + + def get_feature_value( + self, location_id, feature_name=None, feature_type=None, custom_headers=None, raw=False, **operation_config): + """Get exposure control feature for specific location. + + :param location_id: The location identifier. + :type location_id: str + :param feature_name: The feature name. + :type feature_name: str + :param feature_type: The feature type. + :type feature_type: str + :param dict custom_headers: headers that will be added to the request + :param bool raw: returns the direct response alongside the + deserialized response + :param operation_config: :ref:`Operation configuration + overrides`. + :return: ExposureControlResponse or ClientRawResponse if raw=true + :rtype: ~azure.mgmt.datafactory.models.ExposureControlResponse or + ~msrest.pipeline.ClientRawResponse + :raises: :class:`CloudError` + """ + exposure_control_request = models.ExposureControlRequest(feature_name=feature_name, feature_type=feature_type) + + # Construct URL + url = self.get_feature_value.metadata['url'] + path_format_arguments = { + 'subscriptionId': self._serialize.url("self.config.subscription_id", self.config.subscription_id, 'str'), + 'locationId': self._serialize.url("location_id", location_id, 'str') + } + url = self._client.format_url(url, **path_format_arguments) + + # Construct parameters + query_parameters = {} + query_parameters['api-version'] = self._serialize.query("self.api_version", self.api_version, 'str') + + # Construct headers + header_parameters = {} + header_parameters['Accept'] = 'application/json' + header_parameters['Content-Type'] = 'application/json; charset=utf-8' + if self.config.generate_client_request_id: + header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) + if custom_headers: + header_parameters.update(custom_headers) + if self.config.accept_language is not None: + header_parameters['accept-language'] = self._serialize.header("self.config.accept_language", self.config.accept_language, 'str') + + # Construct body + body_content = self._serialize.body(exposure_control_request, 'ExposureControlRequest') + + # Construct and send request + request = self._client.post(url, query_parameters, header_parameters, body_content) + response = self._client.send(request, stream=False, **operation_config) + + if response.status_code not in [200]: + exp = CloudError(response) + exp.request_id = response.headers.get('x-ms-request-id') + raise exp + + deserialized = None + + if response.status_code == 200: + deserialized = self._deserialize('ExposureControlResponse', response) + + if raw: + client_raw_response = ClientRawResponse(deserialized, response) + return client_raw_response + + return deserialized + get_feature_value.metadata = {'url': '/subscriptions/{subscriptionId}/providers/Microsoft.DataFactory/locations/{locationId}/getFeatureValue'} From e9e2954862fafc2231ec1886ecec2f21ac10592b Mon Sep 17 00:00:00 2001 From: Azure SDK for Python bot Date: Wed, 16 Jan 2019 11:36:50 -0800 Subject: [PATCH 45/54] [AutoPR datafactory/resource-manager] Fix ssis execution credential (#4178) * Generated from 8b1f253665d913fa9220085a4de852f5f3dcd66b Fix ssis execute credential issue * Generated from 8b1f253665d913fa9220085a4de852f5f3dcd66b Fix ssis execute credential issue --- .../datafactory/models/execute_ssis_package_activity.py | 6 +++--- .../datafactory/models/execute_ssis_package_activity_py3.py | 6 +++--- 2 files changed, 6 insertions(+), 6 deletions(-) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_ssis_package_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_ssis_package_activity.py index 150e8481f0df..3ea2abd2e734 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_ssis_package_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_ssis_package_activity.py @@ -48,8 +48,8 @@ class ExecuteSSISPackageActivity(ExecutionActivity): Type: string (or Expression with resultType string). :type environment_path: object :param execution_credential: The package execution credential. - :type execution_credential: dict[str, - ~azure.mgmt.datafactory.models.SSISExecutionCredential] + :type execution_credential: + ~azure.mgmt.datafactory.models.SSISExecutionCredential :param connect_via: Required. The integration runtime reference. :type connect_via: ~azure.mgmt.datafactory.models.IntegrationRuntimeReference @@ -95,7 +95,7 @@ class ExecuteSSISPackageActivity(ExecutionActivity): 'runtime': {'key': 'typeProperties.runtime', 'type': 'object'}, 'logging_level': {'key': 'typeProperties.loggingLevel', 'type': 'object'}, 'environment_path': {'key': 'typeProperties.environmentPath', 'type': 'object'}, - 'execution_credential': {'key': 'typeProperties.executionCredential', 'type': '{SSISExecutionCredential}'}, + 'execution_credential': {'key': 'typeProperties.executionCredential', 'type': 'SSISExecutionCredential'}, 'connect_via': {'key': 'typeProperties.connectVia', 'type': 'IntegrationRuntimeReference'}, 'project_parameters': {'key': 'typeProperties.projectParameters', 'type': '{SSISExecutionParameter}'}, 'package_parameters': {'key': 'typeProperties.packageParameters', 'type': '{SSISExecutionParameter}'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_ssis_package_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_ssis_package_activity_py3.py index 59d65046c96f..fb72bacf03d9 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_ssis_package_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execute_ssis_package_activity_py3.py @@ -48,8 +48,8 @@ class ExecuteSSISPackageActivity(ExecutionActivity): Type: string (or Expression with resultType string). :type environment_path: object :param execution_credential: The package execution credential. - :type execution_credential: dict[str, - ~azure.mgmt.datafactory.models.SSISExecutionCredential] + :type execution_credential: + ~azure.mgmt.datafactory.models.SSISExecutionCredential :param connect_via: Required. The integration runtime reference. :type connect_via: ~azure.mgmt.datafactory.models.IntegrationRuntimeReference @@ -95,7 +95,7 @@ class ExecuteSSISPackageActivity(ExecutionActivity): 'runtime': {'key': 'typeProperties.runtime', 'type': 'object'}, 'logging_level': {'key': 'typeProperties.loggingLevel', 'type': 'object'}, 'environment_path': {'key': 'typeProperties.environmentPath', 'type': 'object'}, - 'execution_credential': {'key': 'typeProperties.executionCredential', 'type': '{SSISExecutionCredential}'}, + 'execution_credential': {'key': 'typeProperties.executionCredential', 'type': 'SSISExecutionCredential'}, 'connect_via': {'key': 'typeProperties.connectVia', 'type': 'IntegrationRuntimeReference'}, 'project_parameters': {'key': 'typeProperties.projectParameters', 'type': '{SSISExecutionParameter}'}, 'package_parameters': {'key': 'typeProperties.packageParameters', 'type': '{SSISExecutionParameter}'}, From a220bc0e8ad35c63781c6596a480f925224ef44b Mon Sep 17 00:00:00 2001 From: Azure SDK for Python bot Date: Thu, 17 Jan 2019 16:44:07 -0800 Subject: [PATCH 46/54] Generated from f33ee24e7884fc7cdd79dfa4276779dfffcb660e (#4204) removing webhook --- .../azure/mgmt/datafactory/models/__init__.py | 5 - .../datafactory/models/control_activity.py | 5 +- .../models/control_activity_py3.py | 5 +- .../data_factory_management_client_enums.py | 5 - .../datafactory/models/web_hook_activity.py | 94 ------------------- .../models/web_hook_activity_py3.py | 94 ------------------- 6 files changed, 4 insertions(+), 204 deletions(-) delete mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_hook_activity.py delete mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_hook_activity_py3.py diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py index 9c3893952ab6..8b56abee27cd 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py @@ -346,7 +346,6 @@ from .wait_activity_py3 import WaitActivity from .for_each_activity_py3 import ForEachActivity from .if_condition_activity_py3 import IfConditionActivity - from .web_hook_activity_py3 import WebHookActivity from .execute_pipeline_activity_py3 import ExecutePipelineActivity from .control_activity_py3 import ControlActivity from .linked_integration_runtime_py3 import LinkedIntegrationRuntime @@ -711,7 +710,6 @@ from .wait_activity import WaitActivity from .for_each_activity import ForEachActivity from .if_condition_activity import IfConditionActivity - from .web_hook_activity import WebHookActivity from .execute_pipeline_activity import ExecutePipelineActivity from .control_activity import ControlActivity from .linked_integration_runtime import LinkedIntegrationRuntime @@ -797,7 +795,6 @@ CopyBehaviorType, PolybaseSettingsRejectType, SapCloudForCustomerSinkWriteBehavior, - WebHookActivityMethod, IntegrationRuntimeType, SelfHostedIntegrationRuntimeNodeStatus, IntegrationRuntimeUpdateResult, @@ -1147,7 +1144,6 @@ 'WaitActivity', 'ForEachActivity', 'IfConditionActivity', - 'WebHookActivity', 'ExecutePipelineActivity', 'ControlActivity', 'LinkedIntegrationRuntime', @@ -1232,7 +1228,6 @@ 'CopyBehaviorType', 'PolybaseSettingsRejectType', 'SapCloudForCustomerSinkWriteBehavior', - 'WebHookActivityMethod', 'IntegrationRuntimeType', 'SelfHostedIntegrationRuntimeNodeStatus', 'IntegrationRuntimeUpdateResult', diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/control_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/control_activity.py index 5d64648774da..3c7023c73911 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/control_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/control_activity.py @@ -18,8 +18,7 @@ class ControlActivity(Activity): You probably want to use the sub-classes and not this class directly. Known sub-classes are: AppendVariableActivity, SetVariableActivity, FilterActivity, ValidationActivity, UntilActivity, WaitActivity, - ForEachActivity, IfConditionActivity, WebHookActivity, - ExecutePipelineActivity + ForEachActivity, IfConditionActivity, ExecutePipelineActivity All required parameters must be populated in order to send to Azure. @@ -53,7 +52,7 @@ class ControlActivity(Activity): } _subtype_map = { - 'type': {'AppendVariable': 'AppendVariableActivity', 'SetVariable': 'SetVariableActivity', 'Filter': 'FilterActivity', 'Validation': 'ValidationActivity', 'Until': 'UntilActivity', 'Wait': 'WaitActivity', 'ForEach': 'ForEachActivity', 'IfCondition': 'IfConditionActivity', 'WebHook': 'WebHookActivity', 'ExecutePipeline': 'ExecutePipelineActivity'} + 'type': {'AppendVariable': 'AppendVariableActivity', 'SetVariable': 'SetVariableActivity', 'Filter': 'FilterActivity', 'Validation': 'ValidationActivity', 'Until': 'UntilActivity', 'Wait': 'WaitActivity', 'ForEach': 'ForEachActivity', 'IfCondition': 'IfConditionActivity', 'ExecutePipeline': 'ExecutePipelineActivity'} } def __init__(self, **kwargs): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/control_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/control_activity_py3.py index be9282168ac1..75798a1b05e1 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/control_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/control_activity_py3.py @@ -18,8 +18,7 @@ class ControlActivity(Activity): You probably want to use the sub-classes and not this class directly. Known sub-classes are: AppendVariableActivity, SetVariableActivity, FilterActivity, ValidationActivity, UntilActivity, WaitActivity, - ForEachActivity, IfConditionActivity, WebHookActivity, - ExecutePipelineActivity + ForEachActivity, IfConditionActivity, ExecutePipelineActivity All required parameters must be populated in order to send to Azure. @@ -53,7 +52,7 @@ class ControlActivity(Activity): } _subtype_map = { - 'type': {'AppendVariable': 'AppendVariableActivity', 'SetVariable': 'SetVariableActivity', 'Filter': 'FilterActivity', 'Validation': 'ValidationActivity', 'Until': 'UntilActivity', 'Wait': 'WaitActivity', 'ForEach': 'ForEachActivity', 'IfCondition': 'IfConditionActivity', 'WebHook': 'WebHookActivity', 'ExecutePipeline': 'ExecutePipelineActivity'} + 'type': {'AppendVariable': 'AppendVariableActivity', 'SetVariable': 'SetVariableActivity', 'Filter': 'FilterActivity', 'Validation': 'ValidationActivity', 'Until': 'UntilActivity', 'Wait': 'WaitActivity', 'ForEach': 'ForEachActivity', 'IfCondition': 'IfConditionActivity', 'ExecutePipeline': 'ExecutePipelineActivity'} } def __init__(self, *, name: str, additional_properties=None, description: str=None, depends_on=None, user_properties=None, **kwargs) -> None: diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_factory_management_client_enums.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_factory_management_client_enums.py index a135c4b2e101..23334492e4d2 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_factory_management_client_enums.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_factory_management_client_enums.py @@ -384,11 +384,6 @@ class SapCloudForCustomerSinkWriteBehavior(str, Enum): update = "Update" -class WebHookActivityMethod(str, Enum): - - post = "POST" - - class IntegrationRuntimeType(str, Enum): managed = "Managed" diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_hook_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_hook_activity.py deleted file mode 100644 index ca8945a9d697..000000000000 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_hook_activity.py +++ /dev/null @@ -1,94 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for -# license information. -# -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is -# regenerated. -# -------------------------------------------------------------------------- - -from .control_activity import ControlActivity - - -class WebHookActivity(ControlActivity): - """WebHook activity. - - Variables are only populated by the server, and will be ignored when - sending a request. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] - :param name: Required. Activity name. - :type name: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.mgmt.datafactory.models.UserProperty] - :param type: Required. Constant filled by server. - :type type: str - :ivar method: Required. Rest API method for target endpoint. Default - value: "POST" . - :vartype method: str - :param url: Required. WebHook activity target endpoint and path. Type: - string (or Expression with resultType string). - :type url: object - :param timeout: Specifies the timeout within which the webhook should be - called back. If there is no value specified, it takes the value of - TimeSpan.FromMinutes(10) which is 10 minutes as default. Type: string (or - Expression with resultType string), pattern: - ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type timeout: object - :param headers: Represents the headers that will be sent to the request. - For example, to set the language and type on a request: "headers" : { - "Accept-Language": "en-us", "Content-Type": "application/json" }. Type: - string (or Expression with resultType string). - :type headers: object - :param body: Represents the payload that will be sent to the endpoint. - Required for POST/PUT method, not allowed for GET method Type: string (or - Expression with resultType string). - :type body: object - :param authentication: Authentication method used for calling the - endpoint. - :type authentication: - ~azure.mgmt.datafactory.models.WebActivityAuthentication - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - 'method': {'required': True, 'constant': True}, - 'url': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'type': {'key': 'type', 'type': 'str'}, - 'method': {'key': 'typeProperties.method', 'type': 'str'}, - 'url': {'key': 'typeProperties.url', 'type': 'object'}, - 'timeout': {'key': 'typeProperties.timeout', 'type': 'object'}, - 'headers': {'key': 'typeProperties.headers', 'type': 'object'}, - 'body': {'key': 'typeProperties.body', 'type': 'object'}, - 'authentication': {'key': 'typeProperties.authentication', 'type': 'WebActivityAuthentication'}, - } - - method = "POST" - - def __init__(self, **kwargs): - super(WebHookActivity, self).__init__(**kwargs) - self.url = kwargs.get('url', None) - self.timeout = kwargs.get('timeout', None) - self.headers = kwargs.get('headers', None) - self.body = kwargs.get('body', None) - self.authentication = kwargs.get('authentication', None) - self.type = 'WebHook' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_hook_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_hook_activity_py3.py deleted file mode 100644 index e4779bfbca07..000000000000 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/web_hook_activity_py3.py +++ /dev/null @@ -1,94 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for -# license information. -# -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is -# regenerated. -# -------------------------------------------------------------------------- - -from .control_activity_py3 import ControlActivity - - -class WebHookActivity(ControlActivity): - """WebHook activity. - - Variables are only populated by the server, and will be ignored when - sending a request. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] - :param name: Required. Activity name. - :type name: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.mgmt.datafactory.models.UserProperty] - :param type: Required. Constant filled by server. - :type type: str - :ivar method: Required. Rest API method for target endpoint. Default - value: "POST" . - :vartype method: str - :param url: Required. WebHook activity target endpoint and path. Type: - string (or Expression with resultType string). - :type url: object - :param timeout: Specifies the timeout within which the webhook should be - called back. If there is no value specified, it takes the value of - TimeSpan.FromMinutes(10) which is 10 minutes as default. Type: string (or - Expression with resultType string), pattern: - ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type timeout: object - :param headers: Represents the headers that will be sent to the request. - For example, to set the language and type on a request: "headers" : { - "Accept-Language": "en-us", "Content-Type": "application/json" }. Type: - string (or Expression with resultType string). - :type headers: object - :param body: Represents the payload that will be sent to the endpoint. - Required for POST/PUT method, not allowed for GET method Type: string (or - Expression with resultType string). - :type body: object - :param authentication: Authentication method used for calling the - endpoint. - :type authentication: - ~azure.mgmt.datafactory.models.WebActivityAuthentication - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - 'method': {'required': True, 'constant': True}, - 'url': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'type': {'key': 'type', 'type': 'str'}, - 'method': {'key': 'typeProperties.method', 'type': 'str'}, - 'url': {'key': 'typeProperties.url', 'type': 'object'}, - 'timeout': {'key': 'typeProperties.timeout', 'type': 'object'}, - 'headers': {'key': 'typeProperties.headers', 'type': 'object'}, - 'body': {'key': 'typeProperties.body', 'type': 'object'}, - 'authentication': {'key': 'typeProperties.authentication', 'type': 'WebActivityAuthentication'}, - } - - method = "POST" - - def __init__(self, *, name: str, url, additional_properties=None, description: str=None, depends_on=None, user_properties=None, timeout=None, headers=None, body=None, authentication=None, **kwargs) -> None: - super(WebHookActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, **kwargs) - self.url = url - self.timeout = timeout - self.headers = headers - self.body = body - self.authentication = authentication - self.type = 'WebHook' From 97918c80906c9b76937ca5ca6aeb78bba81a53e6 Mon Sep 17 00:00:00 2001 From: Azure SDK for Python bot Date: Fri, 18 Jan 2019 06:42:46 -0800 Subject: [PATCH 47/54] [AutoPR datafactory/resource-manager] [Datafactory] Add Azure Function to Swagger (#4146) * Generated from fe85d670b4a1a49325414614087c0c4558353da6 Azure Function changes * Generated from 19263c53213686a6807712e57fb85b1ab1840266 update defn * Generated from 41fb59ae051421c105bddd407e98a3605d17e297 update object to secret base --- .../azure/mgmt/datafactory/models/__init__.py | 8 ++ .../models/azure_function_activity.py | 85 +++++++++++++++++++ .../models/azure_function_activity_py3.py | 85 +++++++++++++++++++ .../models/azure_function_linked_service.py | 69 +++++++++++++++ .../azure_function_linked_service_py3.py | 69 +++++++++++++++ .../data_factory_management_client_enums.py | 11 +++ .../datafactory/models/execution_activity.py | 14 +-- .../models/execution_activity_py3.py | 14 +-- .../mgmt/datafactory/models/linked_service.py | 34 ++++---- .../datafactory/models/linked_service_py3.py | 34 ++++---- 10 files changed, 375 insertions(+), 48 deletions(-) create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_function_activity.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_function_activity_py3.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_function_linked_service.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_function_linked_service_py3.py diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py index 8b56abee27cd..5b345b0ff707 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py @@ -94,6 +94,7 @@ from .schedule_trigger_recurrence_py3 import ScheduleTriggerRecurrence from .schedule_trigger_py3 import ScheduleTrigger from .multiple_pipeline_trigger_py3 import MultiplePipelineTrigger + from .azure_function_linked_service_py3 import AzureFunctionLinkedService from .responsys_linked_service_py3 import ResponsysLinkedService from .azure_databricks_linked_service_py3 import AzureDatabricksLinkedService from .azure_data_lake_analytics_linked_service_py3 import AzureDataLakeAnalyticsLinkedService @@ -236,6 +237,7 @@ from .azure_blob_dataset_py3 import AzureBlobDataset from .amazon_s3_dataset_py3 import AmazonS3Dataset from .activity_policy_py3 import ActivityPolicy + from .azure_function_activity_py3 import AzureFunctionActivity from .databricks_spark_python_activity_py3 import DatabricksSparkPythonActivity from .databricks_spark_jar_activity_py3 import DatabricksSparkJarActivity from .databricks_notebook_activity_py3 import DatabricksNotebookActivity @@ -458,6 +460,7 @@ from .schedule_trigger_recurrence import ScheduleTriggerRecurrence from .schedule_trigger import ScheduleTrigger from .multiple_pipeline_trigger import MultiplePipelineTrigger + from .azure_function_linked_service import AzureFunctionLinkedService from .responsys_linked_service import ResponsysLinkedService from .azure_databricks_linked_service import AzureDatabricksLinkedService from .azure_data_lake_analytics_linked_service import AzureDataLakeAnalyticsLinkedService @@ -600,6 +603,7 @@ from .azure_blob_dataset import AzureBlobDataset from .amazon_s3_dataset import AmazonS3Dataset from .activity_policy import ActivityPolicy + from .azure_function_activity import AzureFunctionActivity from .databricks_spark_python_activity import DatabricksSparkPythonActivity from .databricks_spark_jar_activity import DatabricksSparkJarActivity from .databricks_notebook_activity import DatabricksNotebookActivity @@ -785,6 +789,7 @@ SybaseAuthenticationType, DatasetCompressionLevel, JsonFormatFilePattern, + AzureFunctionActivityMethod, WebActivityMethod, CassandraSourceReadConsistencyLevels, StoredProcedureParameterType, @@ -892,6 +897,7 @@ 'ScheduleTriggerRecurrence', 'ScheduleTrigger', 'MultiplePipelineTrigger', + 'AzureFunctionLinkedService', 'ResponsysLinkedService', 'AzureDatabricksLinkedService', 'AzureDataLakeAnalyticsLinkedService', @@ -1034,6 +1040,7 @@ 'AzureBlobDataset', 'AmazonS3Dataset', 'ActivityPolicy', + 'AzureFunctionActivity', 'DatabricksSparkPythonActivity', 'DatabricksSparkJarActivity', 'DatabricksNotebookActivity', @@ -1218,6 +1225,7 @@ 'SybaseAuthenticationType', 'DatasetCompressionLevel', 'JsonFormatFilePattern', + 'AzureFunctionActivityMethod', 'WebActivityMethod', 'CassandraSourceReadConsistencyLevels', 'StoredProcedureParameterType', diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_function_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_function_activity.py new file mode 100644 index 000000000000..68b02e5f771f --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_function_activity.py @@ -0,0 +1,85 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .execution_activity import ExecutionActivity + + +class AzureFunctionActivity(ExecutionActivity): + """Azure Function activity. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param name: Required. Activity name. + :type name: str + :param description: Activity description. + :type description: str + :param depends_on: Activity depends on condition. + :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param user_properties: Activity user properties. + :type user_properties: list[~azure.mgmt.datafactory.models.UserProperty] + :param type: Required. Constant filled by server. + :type type: str + :param linked_service_name: Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param policy: Activity policy. + :type policy: ~azure.mgmt.datafactory.models.ActivityPolicy + :param method: Required. Rest API method for target endpoint. Possible + values include: 'GET', 'POST', 'PUT', 'DELETE', 'OPTIONS', 'HEAD', 'TRACE' + :type method: str or + ~azure.mgmt.datafactory.models.AzureFunctionActivityMethod + :param function_name: Required. Name of the Function that the Azure + Function Activity will call. Type: string (or Expression with resultType + string) + :type function_name: object + :param headers: Represents the headers that will be sent to the request. + For example, to set the language and type on a request: "headers" : { + "Accept-Language": "en-us", "Content-Type": "application/json" }. Type: + string (or Expression with resultType string). + :type headers: object + :param body: Represents the payload that will be sent to the endpoint. + Required for POST/PUT method, not allowed for GET method Type: string (or + Expression with resultType string). + :type body: object + """ + + _validation = { + 'name': {'required': True}, + 'type': {'required': True}, + 'method': {'required': True}, + 'function_name': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'name': {'key': 'name', 'type': 'str'}, + 'description': {'key': 'description', 'type': 'str'}, + 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, + 'method': {'key': 'typeProperties.method', 'type': 'str'}, + 'function_name': {'key': 'typeProperties.functionName', 'type': 'object'}, + 'headers': {'key': 'typeProperties.headers', 'type': 'object'}, + 'body': {'key': 'typeProperties.body', 'type': 'object'}, + } + + def __init__(self, **kwargs): + super(AzureFunctionActivity, self).__init__(**kwargs) + self.method = kwargs.get('method', None) + self.function_name = kwargs.get('function_name', None) + self.headers = kwargs.get('headers', None) + self.body = kwargs.get('body', None) + self.type = 'AzureFunctionActivity' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_function_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_function_activity_py3.py new file mode 100644 index 000000000000..95bb1ca260e7 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_function_activity_py3.py @@ -0,0 +1,85 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .execution_activity_py3 import ExecutionActivity + + +class AzureFunctionActivity(ExecutionActivity): + """Azure Function activity. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param name: Required. Activity name. + :type name: str + :param description: Activity description. + :type description: str + :param depends_on: Activity depends on condition. + :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param user_properties: Activity user properties. + :type user_properties: list[~azure.mgmt.datafactory.models.UserProperty] + :param type: Required. Constant filled by server. + :type type: str + :param linked_service_name: Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param policy: Activity policy. + :type policy: ~azure.mgmt.datafactory.models.ActivityPolicy + :param method: Required. Rest API method for target endpoint. Possible + values include: 'GET', 'POST', 'PUT', 'DELETE', 'OPTIONS', 'HEAD', 'TRACE' + :type method: str or + ~azure.mgmt.datafactory.models.AzureFunctionActivityMethod + :param function_name: Required. Name of the Function that the Azure + Function Activity will call. Type: string (or Expression with resultType + string) + :type function_name: object + :param headers: Represents the headers that will be sent to the request. + For example, to set the language and type on a request: "headers" : { + "Accept-Language": "en-us", "Content-Type": "application/json" }. Type: + string (or Expression with resultType string). + :type headers: object + :param body: Represents the payload that will be sent to the endpoint. + Required for POST/PUT method, not allowed for GET method Type: string (or + Expression with resultType string). + :type body: object + """ + + _validation = { + 'name': {'required': True}, + 'type': {'required': True}, + 'method': {'required': True}, + 'function_name': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'name': {'key': 'name', 'type': 'str'}, + 'description': {'key': 'description', 'type': 'str'}, + 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, + 'method': {'key': 'typeProperties.method', 'type': 'str'}, + 'function_name': {'key': 'typeProperties.functionName', 'type': 'object'}, + 'headers': {'key': 'typeProperties.headers', 'type': 'object'}, + 'body': {'key': 'typeProperties.body', 'type': 'object'}, + } + + def __init__(self, *, name: str, method, function_name, additional_properties=None, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, headers=None, body=None, **kwargs) -> None: + super(AzureFunctionActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) + self.method = method + self.function_name = function_name + self.headers = headers + self.body = body + self.type = 'AzureFunctionActivity' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_function_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_function_linked_service.py new file mode 100644 index 000000000000..44917d8d23b9 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_function_linked_service.py @@ -0,0 +1,69 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class AzureFunctionLinkedService(LinkedService): + """Azure Function linked service. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param function_app_url: Required. The endpoint of the Azure Function App. + URL will be in the format https://.azurewebsites.net. + :type function_app_url: object + :param function_key: Function or Host key for Azure Function App. + :type function_key: ~azure.mgmt.datafactory.models.SecretBase + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'function_app_url': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'function_app_url': {'key': 'typeProperties.functionAppUrl', 'type': 'object'}, + 'function_key': {'key': 'typeProperties.functionKey', 'type': 'SecretBase'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, **kwargs): + super(AzureFunctionLinkedService, self).__init__(**kwargs) + self.function_app_url = kwargs.get('function_app_url', None) + self.function_key = kwargs.get('function_key', None) + self.encrypted_credential = kwargs.get('encrypted_credential', None) + self.type = 'AzureFunction' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_function_linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_function_linked_service_py3.py new file mode 100644 index 000000000000..b6b0f9600da1 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_function_linked_service_py3.py @@ -0,0 +1,69 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service_py3 import LinkedService + + +class AzureFunctionLinkedService(LinkedService): + """Azure Function linked service. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are + deserialized this collection + :type additional_properties: dict[str, object] + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the + Dataset. + :type annotations: list[object] + :param type: Required. Constant filled by server. + :type type: str + :param function_app_url: Required. The endpoint of the Azure Function App. + URL will be in the format https://.azurewebsites.net. + :type function_app_url: object + :param function_key: Function or Host key for Azure Function App. + :type function_key: ~azure.mgmt.datafactory.models.SecretBase + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'function_app_url': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'type': {'key': 'type', 'type': 'str'}, + 'function_app_url': {'key': 'typeProperties.functionAppUrl', 'type': 'object'}, + 'function_key': {'key': 'typeProperties.functionKey', 'type': 'SecretBase'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, *, function_app_url, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, function_key=None, encrypted_credential=None, **kwargs) -> None: + super(AzureFunctionLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.function_app_url = function_app_url + self.function_key = function_key + self.encrypted_credential = encrypted_credential + self.type = 'AzureFunction' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_factory_management_client_enums.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_factory_management_client_enums.py index 23334492e4d2..2992964b1799 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_factory_management_client_enums.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_factory_management_client_enums.py @@ -308,6 +308,17 @@ class JsonFormatFilePattern(str, Enum): array_of_objects = "arrayOfObjects" +class AzureFunctionActivityMethod(str, Enum): + + get = "GET" + post = "POST" + put = "PUT" + delete = "DELETE" + options = "OPTIONS" + head = "HEAD" + trace = "TRACE" + + class WebActivityMethod(str, Enum): get = "GET" diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execution_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execution_activity.py index e061d595a3ba..aca89a009b8e 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execution_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execution_activity.py @@ -16,12 +16,12 @@ class ExecutionActivity(Activity): """Base class for all execution activities. You probably want to use the sub-classes and not this class directly. Known - sub-classes are: DatabricksSparkPythonActivity, DatabricksSparkJarActivity, - DatabricksNotebookActivity, DataLakeAnalyticsUSQLActivity, - AzureMLUpdateResourceActivity, AzureMLBatchExecutionActivity, - GetMetadataActivity, WebActivity, LookupActivity, DeleteActivity, - SqlServerStoredProcedureActivity, CustomActivity, - ExecuteSSISPackageActivity, HDInsightSparkActivity, + sub-classes are: AzureFunctionActivity, DatabricksSparkPythonActivity, + DatabricksSparkJarActivity, DatabricksNotebookActivity, + DataLakeAnalyticsUSQLActivity, AzureMLUpdateResourceActivity, + AzureMLBatchExecutionActivity, GetMetadataActivity, WebActivity, + LookupActivity, DeleteActivity, SqlServerStoredProcedureActivity, + CustomActivity, ExecuteSSISPackageActivity, HDInsightSparkActivity, HDInsightStreamingActivity, HDInsightMapReduceActivity, HDInsightPigActivity, HDInsightHiveActivity, CopyActivity @@ -64,7 +64,7 @@ class ExecutionActivity(Activity): } _subtype_map = { - 'type': {'DatabricksSparkPython': 'DatabricksSparkPythonActivity', 'DatabricksSparkJar': 'DatabricksSparkJarActivity', 'DatabricksNotebook': 'DatabricksNotebookActivity', 'DataLakeAnalyticsU-SQL': 'DataLakeAnalyticsUSQLActivity', 'AzureMLUpdateResource': 'AzureMLUpdateResourceActivity', 'AzureMLBatchExecution': 'AzureMLBatchExecutionActivity', 'GetMetadata': 'GetMetadataActivity', 'WebActivity': 'WebActivity', 'Lookup': 'LookupActivity', 'Delete': 'DeleteActivity', 'SqlServerStoredProcedure': 'SqlServerStoredProcedureActivity', 'Custom': 'CustomActivity', 'ExecuteSSISPackage': 'ExecuteSSISPackageActivity', 'HDInsightSpark': 'HDInsightSparkActivity', 'HDInsightStreaming': 'HDInsightStreamingActivity', 'HDInsightMapReduce': 'HDInsightMapReduceActivity', 'HDInsightPig': 'HDInsightPigActivity', 'HDInsightHive': 'HDInsightHiveActivity', 'Copy': 'CopyActivity'} + 'type': {'AzureFunctionActivity': 'AzureFunctionActivity', 'DatabricksSparkPython': 'DatabricksSparkPythonActivity', 'DatabricksSparkJar': 'DatabricksSparkJarActivity', 'DatabricksNotebook': 'DatabricksNotebookActivity', 'DataLakeAnalyticsU-SQL': 'DataLakeAnalyticsUSQLActivity', 'AzureMLUpdateResource': 'AzureMLUpdateResourceActivity', 'AzureMLBatchExecution': 'AzureMLBatchExecutionActivity', 'GetMetadata': 'GetMetadataActivity', 'WebActivity': 'WebActivity', 'Lookup': 'LookupActivity', 'Delete': 'DeleteActivity', 'SqlServerStoredProcedure': 'SqlServerStoredProcedureActivity', 'Custom': 'CustomActivity', 'ExecuteSSISPackage': 'ExecuteSSISPackageActivity', 'HDInsightSpark': 'HDInsightSparkActivity', 'HDInsightStreaming': 'HDInsightStreamingActivity', 'HDInsightMapReduce': 'HDInsightMapReduceActivity', 'HDInsightPig': 'HDInsightPigActivity', 'HDInsightHive': 'HDInsightHiveActivity', 'Copy': 'CopyActivity'} } def __init__(self, **kwargs): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execution_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execution_activity_py3.py index e3fcc573bc48..7f3b452fc3f9 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execution_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/execution_activity_py3.py @@ -16,12 +16,12 @@ class ExecutionActivity(Activity): """Base class for all execution activities. You probably want to use the sub-classes and not this class directly. Known - sub-classes are: DatabricksSparkPythonActivity, DatabricksSparkJarActivity, - DatabricksNotebookActivity, DataLakeAnalyticsUSQLActivity, - AzureMLUpdateResourceActivity, AzureMLBatchExecutionActivity, - GetMetadataActivity, WebActivity, LookupActivity, DeleteActivity, - SqlServerStoredProcedureActivity, CustomActivity, - ExecuteSSISPackageActivity, HDInsightSparkActivity, + sub-classes are: AzureFunctionActivity, DatabricksSparkPythonActivity, + DatabricksSparkJarActivity, DatabricksNotebookActivity, + DataLakeAnalyticsUSQLActivity, AzureMLUpdateResourceActivity, + AzureMLBatchExecutionActivity, GetMetadataActivity, WebActivity, + LookupActivity, DeleteActivity, SqlServerStoredProcedureActivity, + CustomActivity, ExecuteSSISPackageActivity, HDInsightSparkActivity, HDInsightStreamingActivity, HDInsightMapReduceActivity, HDInsightPigActivity, HDInsightHiveActivity, CopyActivity @@ -64,7 +64,7 @@ class ExecutionActivity(Activity): } _subtype_map = { - 'type': {'DatabricksSparkPython': 'DatabricksSparkPythonActivity', 'DatabricksSparkJar': 'DatabricksSparkJarActivity', 'DatabricksNotebook': 'DatabricksNotebookActivity', 'DataLakeAnalyticsU-SQL': 'DataLakeAnalyticsUSQLActivity', 'AzureMLUpdateResource': 'AzureMLUpdateResourceActivity', 'AzureMLBatchExecution': 'AzureMLBatchExecutionActivity', 'GetMetadata': 'GetMetadataActivity', 'WebActivity': 'WebActivity', 'Lookup': 'LookupActivity', 'Delete': 'DeleteActivity', 'SqlServerStoredProcedure': 'SqlServerStoredProcedureActivity', 'Custom': 'CustomActivity', 'ExecuteSSISPackage': 'ExecuteSSISPackageActivity', 'HDInsightSpark': 'HDInsightSparkActivity', 'HDInsightStreaming': 'HDInsightStreamingActivity', 'HDInsightMapReduce': 'HDInsightMapReduceActivity', 'HDInsightPig': 'HDInsightPigActivity', 'HDInsightHive': 'HDInsightHiveActivity', 'Copy': 'CopyActivity'} + 'type': {'AzureFunctionActivity': 'AzureFunctionActivity', 'DatabricksSparkPython': 'DatabricksSparkPythonActivity', 'DatabricksSparkJar': 'DatabricksSparkJarActivity', 'DatabricksNotebook': 'DatabricksNotebookActivity', 'DataLakeAnalyticsU-SQL': 'DataLakeAnalyticsUSQLActivity', 'AzureMLUpdateResource': 'AzureMLUpdateResourceActivity', 'AzureMLBatchExecution': 'AzureMLBatchExecutionActivity', 'GetMetadata': 'GetMetadataActivity', 'WebActivity': 'WebActivity', 'Lookup': 'LookupActivity', 'Delete': 'DeleteActivity', 'SqlServerStoredProcedure': 'SqlServerStoredProcedureActivity', 'Custom': 'CustomActivity', 'ExecuteSSISPackage': 'ExecuteSSISPackageActivity', 'HDInsightSpark': 'HDInsightSparkActivity', 'HDInsightStreaming': 'HDInsightStreamingActivity', 'HDInsightMapReduce': 'HDInsightMapReduceActivity', 'HDInsightPig': 'HDInsightPigActivity', 'HDInsightHive': 'HDInsightHiveActivity', 'Copy': 'CopyActivity'} } def __init__(self, *, name: str, additional_properties=None, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, **kwargs) -> None: diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service.py index 3dec3c0dcc24..62f172fded76 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service.py @@ -18,22 +18,22 @@ class LinkedService(Model): resource. You probably want to use the sub-classes and not this class directly. Known - sub-classes are: ResponsysLinkedService, AzureDatabricksLinkedService, - AzureDataLakeAnalyticsLinkedService, HDInsightOnDemandLinkedService, - SalesforceMarketingCloudLinkedService, NetezzaLinkedService, - VerticaLinkedService, ZohoLinkedService, XeroLinkedService, - SquareLinkedService, SparkLinkedService, ShopifyLinkedService, - ServiceNowLinkedService, QuickBooksLinkedService, PrestoLinkedService, - PhoenixLinkedService, PaypalLinkedService, MarketoLinkedService, - MariaDBLinkedService, MagentoLinkedService, JiraLinkedService, - ImpalaLinkedService, HubspotLinkedService, HiveLinkedService, - HBaseLinkedService, GreenplumLinkedService, GoogleBigQueryLinkedService, - EloquaLinkedService, DrillLinkedService, CouchbaseLinkedService, - ConcurLinkedService, AzurePostgreSqlLinkedService, AmazonMWSLinkedService, - SapHanaLinkedService, SapBWLinkedService, SftpServerLinkedService, - FtpServerLinkedService, HttpLinkedService, AzureSearchLinkedService, - CustomDataSourceLinkedService, AmazonRedshiftLinkedService, - AmazonS3LinkedService, SapEccLinkedService, + sub-classes are: AzureFunctionLinkedService, ResponsysLinkedService, + AzureDatabricksLinkedService, AzureDataLakeAnalyticsLinkedService, + HDInsightOnDemandLinkedService, SalesforceMarketingCloudLinkedService, + NetezzaLinkedService, VerticaLinkedService, ZohoLinkedService, + XeroLinkedService, SquareLinkedService, SparkLinkedService, + ShopifyLinkedService, ServiceNowLinkedService, QuickBooksLinkedService, + PrestoLinkedService, PhoenixLinkedService, PaypalLinkedService, + MarketoLinkedService, MariaDBLinkedService, MagentoLinkedService, + JiraLinkedService, ImpalaLinkedService, HubspotLinkedService, + HiveLinkedService, HBaseLinkedService, GreenplumLinkedService, + GoogleBigQueryLinkedService, EloquaLinkedService, DrillLinkedService, + CouchbaseLinkedService, ConcurLinkedService, AzurePostgreSqlLinkedService, + AmazonMWSLinkedService, SapHanaLinkedService, SapBWLinkedService, + SftpServerLinkedService, FtpServerLinkedService, HttpLinkedService, + AzureSearchLinkedService, CustomDataSourceLinkedService, + AmazonRedshiftLinkedService, AmazonS3LinkedService, SapEccLinkedService, SapCloudForCustomerLinkedService, SalesforceLinkedService, AzureDataLakeStoreLinkedService, MongoDbLinkedService, CassandraLinkedService, WebLinkedService, ODataLinkedService, @@ -81,7 +81,7 @@ class LinkedService(Model): } _subtype_map = { - 'type': {'Responsys': 'ResponsysLinkedService', 'AzureDatabricks': 'AzureDatabricksLinkedService', 'AzureDataLakeAnalytics': 'AzureDataLakeAnalyticsLinkedService', 'HDInsightOnDemand': 'HDInsightOnDemandLinkedService', 'SalesforceMarketingCloud': 'SalesforceMarketingCloudLinkedService', 'Netezza': 'NetezzaLinkedService', 'Vertica': 'VerticaLinkedService', 'Zoho': 'ZohoLinkedService', 'Xero': 'XeroLinkedService', 'Square': 'SquareLinkedService', 'Spark': 'SparkLinkedService', 'Shopify': 'ShopifyLinkedService', 'ServiceNow': 'ServiceNowLinkedService', 'QuickBooks': 'QuickBooksLinkedService', 'Presto': 'PrestoLinkedService', 'Phoenix': 'PhoenixLinkedService', 'Paypal': 'PaypalLinkedService', 'Marketo': 'MarketoLinkedService', 'MariaDB': 'MariaDBLinkedService', 'Magento': 'MagentoLinkedService', 'Jira': 'JiraLinkedService', 'Impala': 'ImpalaLinkedService', 'Hubspot': 'HubspotLinkedService', 'Hive': 'HiveLinkedService', 'HBase': 'HBaseLinkedService', 'Greenplum': 'GreenplumLinkedService', 'GoogleBigQuery': 'GoogleBigQueryLinkedService', 'Eloqua': 'EloquaLinkedService', 'Drill': 'DrillLinkedService', 'Couchbase': 'CouchbaseLinkedService', 'Concur': 'ConcurLinkedService', 'AzurePostgreSql': 'AzurePostgreSqlLinkedService', 'AmazonMWS': 'AmazonMWSLinkedService', 'SapHana': 'SapHanaLinkedService', 'SapBW': 'SapBWLinkedService', 'Sftp': 'SftpServerLinkedService', 'FtpServer': 'FtpServerLinkedService', 'HttpServer': 'HttpLinkedService', 'AzureSearch': 'AzureSearchLinkedService', 'CustomDataSource': 'CustomDataSourceLinkedService', 'AmazonRedshift': 'AmazonRedshiftLinkedService', 'AmazonS3': 'AmazonS3LinkedService', 'SapEcc': 'SapEccLinkedService', 'SapCloudForCustomer': 'SapCloudForCustomerLinkedService', 'Salesforce': 'SalesforceLinkedService', 'AzureDataLakeStore': 'AzureDataLakeStoreLinkedService', 'MongoDb': 'MongoDbLinkedService', 'Cassandra': 'CassandraLinkedService', 'Web': 'WebLinkedService', 'OData': 'ODataLinkedService', 'Hdfs': 'HdfsLinkedService', 'Odbc': 'OdbcLinkedService', 'AzureML': 'AzureMLLinkedService', 'Teradata': 'TeradataLinkedService', 'Db2': 'Db2LinkedService', 'Sybase': 'SybaseLinkedService', 'PostgreSql': 'PostgreSqlLinkedService', 'MySql': 'MySqlLinkedService', 'AzureMySql': 'AzureMySqlLinkedService', 'Oracle': 'OracleLinkedService', 'FileServer': 'FileServerLinkedService', 'HDInsight': 'HDInsightLinkedService', 'Dynamics': 'DynamicsLinkedService', 'CosmosDb': 'CosmosDbLinkedService', 'AzureKeyVault': 'AzureKeyVaultLinkedService', 'AzureBatch': 'AzureBatchLinkedService', 'AzureSqlDatabase': 'AzureSqlDatabaseLinkedService', 'SqlServer': 'SqlServerLinkedService', 'AzureSqlDW': 'AzureSqlDWLinkedService', 'AzureTableStorage': 'AzureTableStorageLinkedService', 'AzureBlobStorage': 'AzureBlobStorageLinkedService', 'AzureStorage': 'AzureStorageLinkedService'} + 'type': {'AzureFunction': 'AzureFunctionLinkedService', 'Responsys': 'ResponsysLinkedService', 'AzureDatabricks': 'AzureDatabricksLinkedService', 'AzureDataLakeAnalytics': 'AzureDataLakeAnalyticsLinkedService', 'HDInsightOnDemand': 'HDInsightOnDemandLinkedService', 'SalesforceMarketingCloud': 'SalesforceMarketingCloudLinkedService', 'Netezza': 'NetezzaLinkedService', 'Vertica': 'VerticaLinkedService', 'Zoho': 'ZohoLinkedService', 'Xero': 'XeroLinkedService', 'Square': 'SquareLinkedService', 'Spark': 'SparkLinkedService', 'Shopify': 'ShopifyLinkedService', 'ServiceNow': 'ServiceNowLinkedService', 'QuickBooks': 'QuickBooksLinkedService', 'Presto': 'PrestoLinkedService', 'Phoenix': 'PhoenixLinkedService', 'Paypal': 'PaypalLinkedService', 'Marketo': 'MarketoLinkedService', 'MariaDB': 'MariaDBLinkedService', 'Magento': 'MagentoLinkedService', 'Jira': 'JiraLinkedService', 'Impala': 'ImpalaLinkedService', 'Hubspot': 'HubspotLinkedService', 'Hive': 'HiveLinkedService', 'HBase': 'HBaseLinkedService', 'Greenplum': 'GreenplumLinkedService', 'GoogleBigQuery': 'GoogleBigQueryLinkedService', 'Eloqua': 'EloquaLinkedService', 'Drill': 'DrillLinkedService', 'Couchbase': 'CouchbaseLinkedService', 'Concur': 'ConcurLinkedService', 'AzurePostgreSql': 'AzurePostgreSqlLinkedService', 'AmazonMWS': 'AmazonMWSLinkedService', 'SapHana': 'SapHanaLinkedService', 'SapBW': 'SapBWLinkedService', 'Sftp': 'SftpServerLinkedService', 'FtpServer': 'FtpServerLinkedService', 'HttpServer': 'HttpLinkedService', 'AzureSearch': 'AzureSearchLinkedService', 'CustomDataSource': 'CustomDataSourceLinkedService', 'AmazonRedshift': 'AmazonRedshiftLinkedService', 'AmazonS3': 'AmazonS3LinkedService', 'SapEcc': 'SapEccLinkedService', 'SapCloudForCustomer': 'SapCloudForCustomerLinkedService', 'Salesforce': 'SalesforceLinkedService', 'AzureDataLakeStore': 'AzureDataLakeStoreLinkedService', 'MongoDb': 'MongoDbLinkedService', 'Cassandra': 'CassandraLinkedService', 'Web': 'WebLinkedService', 'OData': 'ODataLinkedService', 'Hdfs': 'HdfsLinkedService', 'Odbc': 'OdbcLinkedService', 'AzureML': 'AzureMLLinkedService', 'Teradata': 'TeradataLinkedService', 'Db2': 'Db2LinkedService', 'Sybase': 'SybaseLinkedService', 'PostgreSql': 'PostgreSqlLinkedService', 'MySql': 'MySqlLinkedService', 'AzureMySql': 'AzureMySqlLinkedService', 'Oracle': 'OracleLinkedService', 'FileServer': 'FileServerLinkedService', 'HDInsight': 'HDInsightLinkedService', 'Dynamics': 'DynamicsLinkedService', 'CosmosDb': 'CosmosDbLinkedService', 'AzureKeyVault': 'AzureKeyVaultLinkedService', 'AzureBatch': 'AzureBatchLinkedService', 'AzureSqlDatabase': 'AzureSqlDatabaseLinkedService', 'SqlServer': 'SqlServerLinkedService', 'AzureSqlDW': 'AzureSqlDWLinkedService', 'AzureTableStorage': 'AzureTableStorageLinkedService', 'AzureBlobStorage': 'AzureBlobStorageLinkedService', 'AzureStorage': 'AzureStorageLinkedService'} } def __init__(self, **kwargs): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service_py3.py index 885af520bcda..ff4bb8c7605d 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service_py3.py @@ -18,22 +18,22 @@ class LinkedService(Model): resource. You probably want to use the sub-classes and not this class directly. Known - sub-classes are: ResponsysLinkedService, AzureDatabricksLinkedService, - AzureDataLakeAnalyticsLinkedService, HDInsightOnDemandLinkedService, - SalesforceMarketingCloudLinkedService, NetezzaLinkedService, - VerticaLinkedService, ZohoLinkedService, XeroLinkedService, - SquareLinkedService, SparkLinkedService, ShopifyLinkedService, - ServiceNowLinkedService, QuickBooksLinkedService, PrestoLinkedService, - PhoenixLinkedService, PaypalLinkedService, MarketoLinkedService, - MariaDBLinkedService, MagentoLinkedService, JiraLinkedService, - ImpalaLinkedService, HubspotLinkedService, HiveLinkedService, - HBaseLinkedService, GreenplumLinkedService, GoogleBigQueryLinkedService, - EloquaLinkedService, DrillLinkedService, CouchbaseLinkedService, - ConcurLinkedService, AzurePostgreSqlLinkedService, AmazonMWSLinkedService, - SapHanaLinkedService, SapBWLinkedService, SftpServerLinkedService, - FtpServerLinkedService, HttpLinkedService, AzureSearchLinkedService, - CustomDataSourceLinkedService, AmazonRedshiftLinkedService, - AmazonS3LinkedService, SapEccLinkedService, + sub-classes are: AzureFunctionLinkedService, ResponsysLinkedService, + AzureDatabricksLinkedService, AzureDataLakeAnalyticsLinkedService, + HDInsightOnDemandLinkedService, SalesforceMarketingCloudLinkedService, + NetezzaLinkedService, VerticaLinkedService, ZohoLinkedService, + XeroLinkedService, SquareLinkedService, SparkLinkedService, + ShopifyLinkedService, ServiceNowLinkedService, QuickBooksLinkedService, + PrestoLinkedService, PhoenixLinkedService, PaypalLinkedService, + MarketoLinkedService, MariaDBLinkedService, MagentoLinkedService, + JiraLinkedService, ImpalaLinkedService, HubspotLinkedService, + HiveLinkedService, HBaseLinkedService, GreenplumLinkedService, + GoogleBigQueryLinkedService, EloquaLinkedService, DrillLinkedService, + CouchbaseLinkedService, ConcurLinkedService, AzurePostgreSqlLinkedService, + AmazonMWSLinkedService, SapHanaLinkedService, SapBWLinkedService, + SftpServerLinkedService, FtpServerLinkedService, HttpLinkedService, + AzureSearchLinkedService, CustomDataSourceLinkedService, + AmazonRedshiftLinkedService, AmazonS3LinkedService, SapEccLinkedService, SapCloudForCustomerLinkedService, SalesforceLinkedService, AzureDataLakeStoreLinkedService, MongoDbLinkedService, CassandraLinkedService, WebLinkedService, ODataLinkedService, @@ -81,7 +81,7 @@ class LinkedService(Model): } _subtype_map = { - 'type': {'Responsys': 'ResponsysLinkedService', 'AzureDatabricks': 'AzureDatabricksLinkedService', 'AzureDataLakeAnalytics': 'AzureDataLakeAnalyticsLinkedService', 'HDInsightOnDemand': 'HDInsightOnDemandLinkedService', 'SalesforceMarketingCloud': 'SalesforceMarketingCloudLinkedService', 'Netezza': 'NetezzaLinkedService', 'Vertica': 'VerticaLinkedService', 'Zoho': 'ZohoLinkedService', 'Xero': 'XeroLinkedService', 'Square': 'SquareLinkedService', 'Spark': 'SparkLinkedService', 'Shopify': 'ShopifyLinkedService', 'ServiceNow': 'ServiceNowLinkedService', 'QuickBooks': 'QuickBooksLinkedService', 'Presto': 'PrestoLinkedService', 'Phoenix': 'PhoenixLinkedService', 'Paypal': 'PaypalLinkedService', 'Marketo': 'MarketoLinkedService', 'MariaDB': 'MariaDBLinkedService', 'Magento': 'MagentoLinkedService', 'Jira': 'JiraLinkedService', 'Impala': 'ImpalaLinkedService', 'Hubspot': 'HubspotLinkedService', 'Hive': 'HiveLinkedService', 'HBase': 'HBaseLinkedService', 'Greenplum': 'GreenplumLinkedService', 'GoogleBigQuery': 'GoogleBigQueryLinkedService', 'Eloqua': 'EloquaLinkedService', 'Drill': 'DrillLinkedService', 'Couchbase': 'CouchbaseLinkedService', 'Concur': 'ConcurLinkedService', 'AzurePostgreSql': 'AzurePostgreSqlLinkedService', 'AmazonMWS': 'AmazonMWSLinkedService', 'SapHana': 'SapHanaLinkedService', 'SapBW': 'SapBWLinkedService', 'Sftp': 'SftpServerLinkedService', 'FtpServer': 'FtpServerLinkedService', 'HttpServer': 'HttpLinkedService', 'AzureSearch': 'AzureSearchLinkedService', 'CustomDataSource': 'CustomDataSourceLinkedService', 'AmazonRedshift': 'AmazonRedshiftLinkedService', 'AmazonS3': 'AmazonS3LinkedService', 'SapEcc': 'SapEccLinkedService', 'SapCloudForCustomer': 'SapCloudForCustomerLinkedService', 'Salesforce': 'SalesforceLinkedService', 'AzureDataLakeStore': 'AzureDataLakeStoreLinkedService', 'MongoDb': 'MongoDbLinkedService', 'Cassandra': 'CassandraLinkedService', 'Web': 'WebLinkedService', 'OData': 'ODataLinkedService', 'Hdfs': 'HdfsLinkedService', 'Odbc': 'OdbcLinkedService', 'AzureML': 'AzureMLLinkedService', 'Teradata': 'TeradataLinkedService', 'Db2': 'Db2LinkedService', 'Sybase': 'SybaseLinkedService', 'PostgreSql': 'PostgreSqlLinkedService', 'MySql': 'MySqlLinkedService', 'AzureMySql': 'AzureMySqlLinkedService', 'Oracle': 'OracleLinkedService', 'FileServer': 'FileServerLinkedService', 'HDInsight': 'HDInsightLinkedService', 'Dynamics': 'DynamicsLinkedService', 'CosmosDb': 'CosmosDbLinkedService', 'AzureKeyVault': 'AzureKeyVaultLinkedService', 'AzureBatch': 'AzureBatchLinkedService', 'AzureSqlDatabase': 'AzureSqlDatabaseLinkedService', 'SqlServer': 'SqlServerLinkedService', 'AzureSqlDW': 'AzureSqlDWLinkedService', 'AzureTableStorage': 'AzureTableStorageLinkedService', 'AzureBlobStorage': 'AzureBlobStorageLinkedService', 'AzureStorage': 'AzureStorageLinkedService'} + 'type': {'AzureFunction': 'AzureFunctionLinkedService', 'Responsys': 'ResponsysLinkedService', 'AzureDatabricks': 'AzureDatabricksLinkedService', 'AzureDataLakeAnalytics': 'AzureDataLakeAnalyticsLinkedService', 'HDInsightOnDemand': 'HDInsightOnDemandLinkedService', 'SalesforceMarketingCloud': 'SalesforceMarketingCloudLinkedService', 'Netezza': 'NetezzaLinkedService', 'Vertica': 'VerticaLinkedService', 'Zoho': 'ZohoLinkedService', 'Xero': 'XeroLinkedService', 'Square': 'SquareLinkedService', 'Spark': 'SparkLinkedService', 'Shopify': 'ShopifyLinkedService', 'ServiceNow': 'ServiceNowLinkedService', 'QuickBooks': 'QuickBooksLinkedService', 'Presto': 'PrestoLinkedService', 'Phoenix': 'PhoenixLinkedService', 'Paypal': 'PaypalLinkedService', 'Marketo': 'MarketoLinkedService', 'MariaDB': 'MariaDBLinkedService', 'Magento': 'MagentoLinkedService', 'Jira': 'JiraLinkedService', 'Impala': 'ImpalaLinkedService', 'Hubspot': 'HubspotLinkedService', 'Hive': 'HiveLinkedService', 'HBase': 'HBaseLinkedService', 'Greenplum': 'GreenplumLinkedService', 'GoogleBigQuery': 'GoogleBigQueryLinkedService', 'Eloqua': 'EloquaLinkedService', 'Drill': 'DrillLinkedService', 'Couchbase': 'CouchbaseLinkedService', 'Concur': 'ConcurLinkedService', 'AzurePostgreSql': 'AzurePostgreSqlLinkedService', 'AmazonMWS': 'AmazonMWSLinkedService', 'SapHana': 'SapHanaLinkedService', 'SapBW': 'SapBWLinkedService', 'Sftp': 'SftpServerLinkedService', 'FtpServer': 'FtpServerLinkedService', 'HttpServer': 'HttpLinkedService', 'AzureSearch': 'AzureSearchLinkedService', 'CustomDataSource': 'CustomDataSourceLinkedService', 'AmazonRedshift': 'AmazonRedshiftLinkedService', 'AmazonS3': 'AmazonS3LinkedService', 'SapEcc': 'SapEccLinkedService', 'SapCloudForCustomer': 'SapCloudForCustomerLinkedService', 'Salesforce': 'SalesforceLinkedService', 'AzureDataLakeStore': 'AzureDataLakeStoreLinkedService', 'MongoDb': 'MongoDbLinkedService', 'Cassandra': 'CassandraLinkedService', 'Web': 'WebLinkedService', 'OData': 'ODataLinkedService', 'Hdfs': 'HdfsLinkedService', 'Odbc': 'OdbcLinkedService', 'AzureML': 'AzureMLLinkedService', 'Teradata': 'TeradataLinkedService', 'Db2': 'Db2LinkedService', 'Sybase': 'SybaseLinkedService', 'PostgreSql': 'PostgreSqlLinkedService', 'MySql': 'MySqlLinkedService', 'AzureMySql': 'AzureMySqlLinkedService', 'Oracle': 'OracleLinkedService', 'FileServer': 'FileServerLinkedService', 'HDInsight': 'HDInsightLinkedService', 'Dynamics': 'DynamicsLinkedService', 'CosmosDb': 'CosmosDbLinkedService', 'AzureKeyVault': 'AzureKeyVaultLinkedService', 'AzureBatch': 'AzureBatchLinkedService', 'AzureSqlDatabase': 'AzureSqlDatabaseLinkedService', 'SqlServer': 'SqlServerLinkedService', 'AzureSqlDW': 'AzureSqlDWLinkedService', 'AzureTableStorage': 'AzureTableStorageLinkedService', 'AzureBlobStorage': 'AzureBlobStorageLinkedService', 'AzureStorage': 'AzureStorageLinkedService'} } def __init__(self, *, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, **kwargs) -> None: From 9326e9a2f9a7937588ccb88d82e835ed7be13d4e Mon Sep 17 00:00:00 2001 From: azuresdkci Date: Fri, 25 Jan 2019 01:06:39 +0000 Subject: [PATCH 48/54] Packaging update of azure-mgmt-datafactory --- azure-mgmt-datafactory/MANIFEST.in | 1 + azure-mgmt-datafactory/README.rst | 2 +- azure-mgmt-datafactory/setup.py | 2 +- 3 files changed, 3 insertions(+), 2 deletions(-) diff --git a/azure-mgmt-datafactory/MANIFEST.in b/azure-mgmt-datafactory/MANIFEST.in index ac87972df424..6ceb27f7a96e 100644 --- a/azure-mgmt-datafactory/MANIFEST.in +++ b/azure-mgmt-datafactory/MANIFEST.in @@ -1,3 +1,4 @@ include *.rst include azure/__init__.py include azure/mgmt/__init__.py + diff --git a/azure-mgmt-datafactory/README.rst b/azure-mgmt-datafactory/README.rst index c0d74ecb16d4..04ec9cafb387 100644 --- a/azure-mgmt-datafactory/README.rst +++ b/azure-mgmt-datafactory/README.rst @@ -1,7 +1,7 @@ Microsoft Azure SDK for Python ============================== -This is the Microsoft Azure MyService Management Client Library. +This is the Microsoft Azure Data Factory Management Client Library. Azure Resource Manager (ARM) is the next generation of management APIs that replace the old Azure Service Management (ASM). diff --git a/azure-mgmt-datafactory/setup.py b/azure-mgmt-datafactory/setup.py index 8fb4b233c1b8..4b3ca4777aca 100644 --- a/azure-mgmt-datafactory/setup.py +++ b/azure-mgmt-datafactory/setup.py @@ -13,7 +13,7 @@ # Change the PACKAGE_NAME only to change folder and different name PACKAGE_NAME = "azure-mgmt-datafactory" -PACKAGE_PPRINT_NAME = "MyService Management" +PACKAGE_PPRINT_NAME = "Data Factory Management" # a-b-c => a/b/c package_folder_path = PACKAGE_NAME.replace('-', '/') From 3702c80cdf7c3b5efd2892f8fb52be439516aeb5 Mon Sep 17 00:00:00 2001 From: Azure SDK for Python bot Date: Fri, 25 Jan 2019 17:27:27 +0000 Subject: [PATCH 49/54] Rebuild by https://github.com/Azure/azure-sdk-for-python/pull/2293 --- .../azure/mgmt/datafactory/models/__init__.py | 3 - .../datafactory/models/control_activity.py | 6 +- .../models/control_activity_py3.py | 6 +- .../datafactory/models/validation_activity.py | 77 ------------------- .../models/validation_activity_py3.py | 77 ------------------- 5 files changed, 6 insertions(+), 163 deletions(-) delete mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/validation_activity.py delete mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/validation_activity_py3.py diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py index 5b345b0ff707..46e9bf12bf1a 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py @@ -343,7 +343,6 @@ from .append_variable_activity_py3 import AppendVariableActivity from .set_variable_activity_py3 import SetVariableActivity from .filter_activity_py3 import FilterActivity - from .validation_activity_py3 import ValidationActivity from .until_activity_py3 import UntilActivity from .wait_activity_py3 import WaitActivity from .for_each_activity_py3 import ForEachActivity @@ -709,7 +708,6 @@ from .append_variable_activity import AppendVariableActivity from .set_variable_activity import SetVariableActivity from .filter_activity import FilterActivity - from .validation_activity import ValidationActivity from .until_activity import UntilActivity from .wait_activity import WaitActivity from .for_each_activity import ForEachActivity @@ -1146,7 +1144,6 @@ 'AppendVariableActivity', 'SetVariableActivity', 'FilterActivity', - 'ValidationActivity', 'UntilActivity', 'WaitActivity', 'ForEachActivity', diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/control_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/control_activity.py index 3c7023c73911..16581581786b 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/control_activity.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/control_activity.py @@ -17,8 +17,8 @@ class ControlActivity(Activity): You probably want to use the sub-classes and not this class directly. Known sub-classes are: AppendVariableActivity, SetVariableActivity, - FilterActivity, ValidationActivity, UntilActivity, WaitActivity, - ForEachActivity, IfConditionActivity, ExecutePipelineActivity + FilterActivity, UntilActivity, WaitActivity, ForEachActivity, + IfConditionActivity, ExecutePipelineActivity All required parameters must be populated in order to send to Azure. @@ -52,7 +52,7 @@ class ControlActivity(Activity): } _subtype_map = { - 'type': {'AppendVariable': 'AppendVariableActivity', 'SetVariable': 'SetVariableActivity', 'Filter': 'FilterActivity', 'Validation': 'ValidationActivity', 'Until': 'UntilActivity', 'Wait': 'WaitActivity', 'ForEach': 'ForEachActivity', 'IfCondition': 'IfConditionActivity', 'ExecutePipeline': 'ExecutePipelineActivity'} + 'type': {'AppendVariable': 'AppendVariableActivity', 'SetVariable': 'SetVariableActivity', 'Filter': 'FilterActivity', 'Until': 'UntilActivity', 'Wait': 'WaitActivity', 'ForEach': 'ForEachActivity', 'IfCondition': 'IfConditionActivity', 'ExecutePipeline': 'ExecutePipelineActivity'} } def __init__(self, **kwargs): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/control_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/control_activity_py3.py index 75798a1b05e1..739d8b9c311b 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/control_activity_py3.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/control_activity_py3.py @@ -17,8 +17,8 @@ class ControlActivity(Activity): You probably want to use the sub-classes and not this class directly. Known sub-classes are: AppendVariableActivity, SetVariableActivity, - FilterActivity, ValidationActivity, UntilActivity, WaitActivity, - ForEachActivity, IfConditionActivity, ExecutePipelineActivity + FilterActivity, UntilActivity, WaitActivity, ForEachActivity, + IfConditionActivity, ExecutePipelineActivity All required parameters must be populated in order to send to Azure. @@ -52,7 +52,7 @@ class ControlActivity(Activity): } _subtype_map = { - 'type': {'AppendVariable': 'AppendVariableActivity', 'SetVariable': 'SetVariableActivity', 'Filter': 'FilterActivity', 'Validation': 'ValidationActivity', 'Until': 'UntilActivity', 'Wait': 'WaitActivity', 'ForEach': 'ForEachActivity', 'IfCondition': 'IfConditionActivity', 'ExecutePipeline': 'ExecutePipelineActivity'} + 'type': {'AppendVariable': 'AppendVariableActivity', 'SetVariable': 'SetVariableActivity', 'Filter': 'FilterActivity', 'Until': 'UntilActivity', 'Wait': 'WaitActivity', 'ForEach': 'ForEachActivity', 'IfCondition': 'IfConditionActivity', 'ExecutePipeline': 'ExecutePipelineActivity'} } def __init__(self, *, name: str, additional_properties=None, description: str=None, depends_on=None, user_properties=None, **kwargs) -> None: diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/validation_activity.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/validation_activity.py deleted file mode 100644 index 6e20563ffc97..000000000000 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/validation_activity.py +++ /dev/null @@ -1,77 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for -# license information. -# -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is -# regenerated. -# -------------------------------------------------------------------------- - -from .control_activity import ControlActivity - - -class ValidationActivity(ControlActivity): - """This activity blocks execution until a file has been validated to exist, - with an optional minimum size, or the timeout is reached, whichever is - earlier. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] - :param name: Required. Activity name. - :type name: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.mgmt.datafactory.models.UserProperty] - :param type: Required. Constant filled by server. - :type type: str - :param timeout: Specifies the timeout for the activity to run. If there is - no value specified, it takes the value of TimeSpan.FromDays(7) which is 1 - week as default. Type: string (or Expression with resultType string), - pattern: ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). Type: - string (or Expression with resultType string), pattern: - ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type timeout: object - :param sleep: A delay in seconds between validation attempts. If no value - is specified, 10 seconds will be used as the default. - :type sleep: int - :param minimum_size: Minimum size of a file in byte. If no value is - specified, 0 byte will be used as the default. - :type minimum_size: int - :param dataset: Required. Validation activity dataset reference. - :type dataset: ~azure.mgmt.datafactory.models.DatasetReference - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - 'dataset': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'type': {'key': 'type', 'type': 'str'}, - 'timeout': {'key': 'typeProperties.timeout', 'type': 'object'}, - 'sleep': {'key': 'typeProperties.sleep', 'type': 'int'}, - 'minimum_size': {'key': 'typeProperties.minimumSize', 'type': 'int'}, - 'dataset': {'key': 'typeProperties.dataset', 'type': 'DatasetReference'}, - } - - def __init__(self, **kwargs): - super(ValidationActivity, self).__init__(**kwargs) - self.timeout = kwargs.get('timeout', None) - self.sleep = kwargs.get('sleep', None) - self.minimum_size = kwargs.get('minimum_size', None) - self.dataset = kwargs.get('dataset', None) - self.type = 'Validation' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/validation_activity_py3.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/validation_activity_py3.py deleted file mode 100644 index 997b5fb4c6e7..000000000000 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/validation_activity_py3.py +++ /dev/null @@ -1,77 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for -# license information. -# -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is -# regenerated. -# -------------------------------------------------------------------------- - -from .control_activity_py3 import ControlActivity - - -class ValidationActivity(ControlActivity): - """This activity blocks execution until a file has been validated to exist, - with an optional minimum size, or the timeout is reached, whichever is - earlier. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] - :param name: Required. Activity name. - :type name: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.mgmt.datafactory.models.UserProperty] - :param type: Required. Constant filled by server. - :type type: str - :param timeout: Specifies the timeout for the activity to run. If there is - no value specified, it takes the value of TimeSpan.FromDays(7) which is 1 - week as default. Type: string (or Expression with resultType string), - pattern: ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). Type: - string (or Expression with resultType string), pattern: - ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type timeout: object - :param sleep: A delay in seconds between validation attempts. If no value - is specified, 10 seconds will be used as the default. - :type sleep: int - :param minimum_size: Minimum size of a file in byte. If no value is - specified, 0 byte will be used as the default. - :type minimum_size: int - :param dataset: Required. Validation activity dataset reference. - :type dataset: ~azure.mgmt.datafactory.models.DatasetReference - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - 'dataset': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'type': {'key': 'type', 'type': 'str'}, - 'timeout': {'key': 'typeProperties.timeout', 'type': 'object'}, - 'sleep': {'key': 'typeProperties.sleep', 'type': 'int'}, - 'minimum_size': {'key': 'typeProperties.minimumSize', 'type': 'int'}, - 'dataset': {'key': 'typeProperties.dataset', 'type': 'DatasetReference'}, - } - - def __init__(self, *, name: str, dataset, additional_properties=None, description: str=None, depends_on=None, user_properties=None, timeout=None, sleep: int=None, minimum_size: int=None, **kwargs) -> None: - super(ValidationActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, **kwargs) - self.timeout = timeout - self.sleep = sleep - self.minimum_size = minimum_size - self.dataset = dataset - self.type = 'Validation' From 53064b7cd3d87e63abe8021ea39b7feca2a4d6bb Mon Sep 17 00:00:00 2001 From: Laurent Mazuel Date: Tue, 29 Jan 2019 17:11:10 -0800 Subject: [PATCH 50/54] Update version.py --- azure-mgmt-datafactory/azure/mgmt/datafactory/version.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/version.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/version.py index a39916c162ce..981739e4ff95 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/version.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/version.py @@ -9,5 +9,5 @@ # regenerated. # -------------------------------------------------------------------------- -VERSION = "1.0.0" +VERSION = "0.7.0" From 95d1561e49348fba5f430cc5ad55ff87daed1087 Mon Sep 17 00:00:00 2001 From: azuresdkci Date: Wed, 30 Jan 2019 02:35:17 +0000 Subject: [PATCH 51/54] Packaging update of azure-mgmt-datafactory --- azure-mgmt-datafactory/README.rst | 19 ------------------- 1 file changed, 19 deletions(-) diff --git a/azure-mgmt-datafactory/README.rst b/azure-mgmt-datafactory/README.rst index 04ec9cafb387..3614d6b6b809 100644 --- a/azure-mgmt-datafactory/README.rst +++ b/azure-mgmt-datafactory/README.rst @@ -14,25 +14,6 @@ For the older Azure Service Management (ASM) libraries, see For a more complete set of Azure libraries, see the `azure `__ bundle package. -Compatibility -============= - -**IMPORTANT**: If you have an earlier version of the azure package -(version < 1.0), you should uninstall it before installing this package. - -You can check the version using pip: - -.. code:: shell - - pip freeze - -If you see azure==0.11.0 (or any version below 1.0), uninstall it first: - -.. code:: shell - - pip uninstall azure - - Usage ===== From 63a044d119044c506a42b93c7e87891f08968d72 Mon Sep 17 00:00:00 2001 From: Laurent Mazuel Date: Wed, 30 Jan 2019 14:33:23 -0800 Subject: [PATCH 52/54] ChangeLog --- azure-mgmt-datafactory/HISTORY.rst | 253 ++++++++++++++++++++++++++++- 1 file changed, 252 insertions(+), 1 deletion(-) diff --git a/azure-mgmt-datafactory/HISTORY.rst b/azure-mgmt-datafactory/HISTORY.rst index de3d528919b0..c2beec74971f 100644 --- a/azure-mgmt-datafactory/HISTORY.rst +++ b/azure-mgmt-datafactory/HISTORY.rst @@ -3,6 +3,257 @@ Release History =============== +0.7.0 (2019-01-31) +++++++++++++++++++ + +**Features** + +- Model MarketoObjectDataset has a new parameter folder +- Model MarketoObjectDataset has a new parameter schema +- Model MarketoObjectDataset has a new parameter table_name +- Model AzureTableDataset has a new parameter folder +- Model AzureTableDataset has a new parameter schema +- Model VerticaTableDataset has a new parameter folder +- Model VerticaTableDataset has a new parameter schema +- Model VerticaTableDataset has a new parameter table_name +- Model VerticaLinkedService has a new parameter pwd +- Model DocumentDbCollectionDataset has a new parameter folder +- Model DocumentDbCollectionDataset has a new parameter schema +- Model HubspotObjectDataset has a new parameter folder +- Model HubspotObjectDataset has a new parameter schema +- Model HubspotObjectDataset has a new parameter table_name +- Model GetMetadataActivity has a new parameter user_properties +- Model SalesforceObjectDataset has a new parameter folder +- Model SalesforceObjectDataset has a new parameter schema +- Model AzureStorageLinkedService has a new parameter account_key +- Model AzureStorageLinkedService has a new parameter sas_token +- Model OracleLinkedService has a new parameter password +- Model ZohoObjectDataset has a new parameter folder +- Model ZohoObjectDataset has a new parameter schema +- Model ZohoObjectDataset has a new parameter table_name +- Model HDInsightHiveActivity has a new parameter variables +- Model HDInsightHiveActivity has a new parameter query_timeout +- Model HDInsightHiveActivity has a new parameter user_properties +- Model AmazonS3Dataset has a new parameter folder +- Model AmazonS3Dataset has a new parameter schema +- Model AzureSqlTableDataset has a new parameter folder +- Model AzureSqlTableDataset has a new parameter schema +- Model Activity has a new parameter user_properties +- Model AzurePostgreSqlLinkedService has a new parameter password +- Model HDInsightMapReduceActivity has a new parameter user_properties +- Model HttpDataset has a new parameter folder +- Model HttpDataset has a new parameter schema +- Model MagentoObjectDataset has a new parameter folder +- Model MagentoObjectDataset has a new parameter schema +- Model MagentoObjectDataset has a new parameter table_name +- Model NetezzaLinkedService has a new parameter pwd +- Model ImpalaObjectDataset has a new parameter folder +- Model ImpalaObjectDataset has a new parameter schema +- Model ImpalaObjectDataset has a new parameter table_name +- Model DrillLinkedService has a new parameter pwd +- Model XeroObjectDataset has a new parameter folder +- Model XeroObjectDataset has a new parameter schema +- Model XeroObjectDataset has a new parameter table_name +- Model ODataResourceDataset has a new parameter folder +- Model ODataResourceDataset has a new parameter schema +- Model MariaDBTableDataset has a new parameter folder +- Model MariaDBTableDataset has a new parameter schema +- Model MariaDBTableDataset has a new parameter table_name +- Model PhoenixObjectDataset has a new parameter folder +- Model PhoenixObjectDataset has a new parameter schema +- Model PhoenixObjectDataset has a new parameter table_name +- Model ShopifyObjectDataset has a new parameter folder +- Model ShopifyObjectDataset has a new parameter schema +- Model ShopifyObjectDataset has a new parameter table_name +- Model DatabricksNotebookActivity has a new parameter libraries +- Model DatabricksNotebookActivity has a new parameter user_properties +- Model HDInsightStreamingActivity has a new parameter user_properties +- Model MariaDBLinkedService has a new parameter pwd +- Model OracleTableDataset has a new parameter folder +- Model OracleTableDataset has a new parameter schema +- Model AzureDatabricksLinkedService has a new parameter new_cluster_spark_env_vars +- Model AzureDatabricksLinkedService has a new parameter new_cluster_custom_tags +- Model ControlActivity has a new parameter user_properties +- Model AzurePostgreSqlTableDataset has a new parameter folder +- Model AzurePostgreSqlTableDataset has a new parameter schema +- Model AzurePostgreSqlTableDataset has a new parameter table_name +- Model EloquaObjectDataset has a new parameter folder +- Model EloquaObjectDataset has a new parameter schema +- Model EloquaObjectDataset has a new parameter table_name +- Model ForEachActivity has a new parameter user_properties +- Model HDInsightPigActivity has a new parameter user_properties +- Model WaitActivity has a new parameter user_properties +- Model DrillTableDataset has a new parameter folder +- Model DrillTableDataset has a new parameter schema +- Model DrillTableDataset has a new parameter table_name +- Model ExecutePipelineActivity has a new parameter user_properties +- Model UntilActivity has a new parameter user_properties +- Model AzureDataLakeStoreDataset has a new parameter folder +- Model AzureDataLakeStoreDataset has a new parameter schema +- Model HDInsightLinkedService has a new parameter is_esp_enabled +- Model SelfHostedIntegrationRuntimeStatus has a new parameter auto_update_eta +- Model SelfHostedIntegrationRuntimeStatus has a new parameter pushed_version +- Model SelfHostedIntegrationRuntimeStatus has a new parameter latest_version +- Model ServiceNowObjectDataset has a new parameter folder +- Model ServiceNowObjectDataset has a new parameter schema +- Model ServiceNowObjectDataset has a new parameter table_name +- Model WebActivity has a new parameter user_properties +- Model QuickBooksObjectDataset has a new parameter folder +- Model QuickBooksObjectDataset has a new parameter schema +- Model QuickBooksObjectDataset has a new parameter table_name +- Model CustomDataset has a new parameter folder +- Model CustomDataset has a new parameter schema +- Model GreenplumTableDataset has a new parameter folder +- Model GreenplumTableDataset has a new parameter schema +- Model GreenplumTableDataset has a new parameter table_name +- Model JiraObjectDataset has a new parameter folder +- Model JiraObjectDataset has a new parameter schema +- Model JiraObjectDataset has a new parameter table_name +- Model CouchbaseLinkedService has a new parameter cred_string +- Model PrestoObjectDataset has a new parameter folder +- Model PrestoObjectDataset has a new parameter schema +- Model PrestoObjectDataset has a new parameter table_name +- Model TabularTranslator has a new parameter schema_mapping +- Model Factory has a new parameter e_tag +- Model Factory has a new parameter repo_configuration +- Model AzureSearchIndexDataset has a new parameter folder +- Model AzureSearchIndexDataset has a new parameter schema +- Model WebTableDataset has a new parameter folder +- Model WebTableDataset has a new parameter schema +- Model FilterActivity has a new parameter user_properties +- Model PipelineRunInvokedBy has a new parameter invoked_by_type +- Model Resource has a new parameter e_tag +- Model RelationalTableDataset has a new parameter folder +- Model RelationalTableDataset has a new parameter schema +- Model AzureSqlDWTableDataset has a new parameter folder +- Model AzureSqlDWTableDataset has a new parameter schema +- Model Dataset has a new parameter folder +- Model Dataset has a new parameter schema +- Model AzureMLBatchExecutionActivity has a new parameter user_properties +- Model CouchbaseTableDataset has a new parameter folder +- Model CouchbaseTableDataset has a new parameter schema +- Model CouchbaseTableDataset has a new parameter table_name +- Model HDInsightSparkActivity has a new parameter user_properties +- Model AzureSqlDWLinkedService has a new parameter password +- Model AzureMLUpdateResourceActivity has a new parameter user_properties +- Model SapEccResourceDataset has a new parameter folder +- Model SapEccResourceDataset has a new parameter schema +- Model LookupActivity has a new parameter user_properties +- Model AzureMySqlLinkedService has a new parameter password +- Model DataLakeAnalyticsUSQLActivity has a new parameter user_properties +- Model CassandraTableDataset has a new parameter folder +- Model CassandraTableDataset has a new parameter schema +- Model SquareObjectDataset has a new parameter folder +- Model SquareObjectDataset has a new parameter schema +- Model SquareObjectDataset has a new parameter table_name +- Model HDInsightOnDemandLinkedService has a new parameter script_actions +- Model PaypalObjectDataset has a new parameter folder +- Model PaypalObjectDataset has a new parameter schema +- Model PaypalObjectDataset has a new parameter table_name +- Model PipelineResource has a new parameter variables +- Model PipelineResource has a new parameter folder +- Model DynamicsEntityDataset has a new parameter folder +- Model DynamicsEntityDataset has a new parameter schema +- Model ActivityPolicy has a new parameter secure_input +- Model FileShareDataset has a new parameter folder +- Model FileShareDataset has a new parameter schema +- Model AzureMySqlTableDataset has a new parameter folder +- Model AzureMySqlTableDataset has a new parameter schema +- Model ExecuteSSISPackageActivity has a new parameter project_connection_managers +- Model ExecuteSSISPackageActivity has a new parameter user_properties +- Model ExecuteSSISPackageActivity has a new parameter package_connection_managers +- Model ExecuteSSISPackageActivity has a new parameter package_parameters +- Model ExecuteSSISPackageActivity has a new parameter property_overrides +- Model ExecuteSSISPackageActivity has a new parameter project_parameters +- Model ExecuteSSISPackageActivity has a new parameter execution_credential +- Model HiveObjectDataset has a new parameter folder +- Model HiveObjectDataset has a new parameter schema +- Model HiveObjectDataset has a new parameter table_name +- Model IfConditionActivity has a new parameter user_properties +- Model CosmosDbLinkedService has a new parameter account_key +- Model GoogleBigQueryObjectDataset has a new parameter folder +- Model GoogleBigQueryObjectDataset has a new parameter schema +- Model GoogleBigQueryObjectDataset has a new parameter table_name +- Model SqlServerTableDataset has a new parameter folder +- Model SqlServerTableDataset has a new parameter schema +- Model SparkObjectDataset has a new parameter folder +- Model SparkObjectDataset has a new parameter schema +- Model SparkObjectDataset has a new parameter table_name +- Model CustomActivity has a new parameter user_properties +- Model SapCloudForCustomerResourceDataset has a new parameter folder +- Model SapCloudForCustomerResourceDataset has a new parameter schema +- Model TumblingWindowTrigger has a new parameter depends_on +- Model SqlServerStoredProcedureActivity has a new parameter user_properties +- Model ConcurObjectDataset has a new parameter folder +- Model ConcurObjectDataset has a new parameter schema +- Model ConcurObjectDataset has a new parameter table_name +- Model OperationMetricSpecification has a new parameter dimensions +- Model HBaseObjectDataset has a new parameter folder +- Model HBaseObjectDataset has a new parameter schema +- Model HBaseObjectDataset has a new parameter table_name +- Model AmazonMWSObjectDataset has a new parameter folder +- Model AmazonMWSObjectDataset has a new parameter schema +- Model AmazonMWSObjectDataset has a new parameter table_name +- Model ExecutionActivity has a new parameter user_properties +- Model AzureBlobDataset has a new parameter folder +- Model AzureBlobDataset has a new parameter schema +- Model AzureSqlDatabaseLinkedService has a new parameter password +- Model MongoDbCollectionDataset has a new parameter folder +- Model MongoDbCollectionDataset has a new parameter schema +- Model CopyActivity has a new parameter data_integration_units +- Model CopyActivity has a new parameter user_properties +- Model SalesforceMarketingCloudObjectDataset has a new parameter folder +- Model SalesforceMarketingCloudObjectDataset has a new parameter schema +- Model SalesforceMarketingCloudObjectDataset has a new parameter table_name +- Model GreenplumLinkedService has a new parameter pwd +- Model NetezzaTableDataset has a new parameter folder +- Model NetezzaTableDataset has a new parameter schema +- Model NetezzaTableDataset has a new parameter table_name +- Added operation PipelineRunsOperations.cancel +- Added operation FactoriesOperations.configure_factory_repo +- Added operation FactoriesOperations.get_data_plane_access +- Added operation FactoriesOperations.get_git_hub_access_token +- Added operation IntegrationRuntimeNodesOperations.get +- Added operation IntegrationRuntimesOperations.create_linked_integration_runtime +- Added operation IntegrationRuntimesOperations.remove_links +- Added operation ActivityRunsOperations.query_by_pipeline_run +- Added operation group RerunTriggersOperations +- Added operation group TriggerRunsOperations +- Added operation group IntegrationRuntimeObjectMetadataOperations +- Added operation group ExposureControlOperations + +**Breaking changes** + +- Parameter access_token_secret of model QuickBooksLinkedService is now required +- Parameter access_token of model QuickBooksLinkedService is now required +- Operation DatasetsOperations.get has a new signature +- Operation FactoriesOperations.create_or_update has a new signature +- Operation FactoriesOperations.get has a new signature +- Operation IntegrationRuntimesOperations.get has a new signature +- Operation LinkedServicesOperations.get has a new signature +- Operation PipelinesOperations.get has a new signature +- Operation TriggersOperations.get has a new signature +- Operation PipelinesOperations.create_run has a new signature +- Model Db2LinkedService no longer has parameter schema +- Model QuickBooksLinkedService has a new required parameter consumer_key +- Model QuickBooksLinkedService has a new required parameter consumer_secret +- Model PostgreSqlLinkedService no longer has parameter database +- Model PostgreSqlLinkedService no longer has parameter username +- Model PostgreSqlLinkedService no longer has parameter schema +- Model PostgreSqlLinkedService no longer has parameter server +- Model PostgreSqlLinkedService has a new required parameter connection_string +- Model TeradataLinkedService no longer has parameter schema +- Model CopyActivity no longer has parameter cloud_data_movement_units +- Model MySqlLinkedService no longer has parameter database +- Model MySqlLinkedService no longer has parameter username +- Model MySqlLinkedService no longer has parameter schema +- Model MySqlLinkedService no longer has parameter server +- Model MySqlLinkedService has a new required parameter connection_string +- Removed operation FactoriesOperations.cancel_pipeline_run +- Removed operation IntegrationRuntimesOperations.remove_node +- Removed operation TriggersOperations.list_runs +- Removed operation ActivityRunsOperations.list_by_pipeline_run + 0.6.0 (2018-03-22) ++++++++++++++++++ @@ -47,7 +298,7 @@ Release History - Add LinkedService, Dataset, Source for 26 RFI connectors, including: PostgreSQL,Google BigQuery,Impala,ServiceNow,Greenplum/Hawq,HBase,Hive ODBC,Spark ODBC,HBase Phoenix,MariaDB,Presto,Couchbase,Concur,Zoho CRM,Amazon Marketplace Services,PayPal,Square,Shopify,QuickBooks Online,Hubspot,Atlassian Jira,Magento,Xero,Drill,Marketo,Eloqua.  - Support round tripping of new properties using additionalProperties for some types  - Add new integration runtime API's: patch integration runtime; patch integration runtime node; upgrade integration runtime, get node IP address  -- Add integration runtime naming validation +- Add integration runtime naming validation 0.2.2 (2017-11-13) ++++++++++++++++++ From 41a4c652a4183333968b81adcd129794aa76d8e3 Mon Sep 17 00:00:00 2001 From: Laurent Mazuel Date: Wed, 30 Jan 2019 14:33:51 -0800 Subject: [PATCH 53/54] Deprecated file --- azure-mgmt-datafactory/build.json | 339 ------------------------------ 1 file changed, 339 deletions(-) delete mode 100644 azure-mgmt-datafactory/build.json diff --git a/azure-mgmt-datafactory/build.json b/azure-mgmt-datafactory/build.json deleted file mode 100644 index dc296c016c97..000000000000 --- a/azure-mgmt-datafactory/build.json +++ /dev/null @@ -1,339 +0,0 @@ -{ - "autorest": [ - { - "resolvedInfo": null, - "packageMetadata": { - "name": "@microsoft.azure/autorest-core", - "version": "2.0.4210", - "engines": { - "node": ">=7.10.0" - }, - "dependencies": {}, - "optionalDependencies": {}, - "devDependencies": { - "@types/commonmark": "^0.27.0", - "@types/js-yaml": "^3.10.0", - "@types/jsonpath": "^0.1.29", - "@types/node": "^8.0.28", - "@types/source-map": "^0.5.0", - "@types/yargs": "^8.0.2", - "dts-generator": "^2.1.0", - "mocha": "^4.0.1", - "mocha-typescript": "^1.1.7", - "shx": "0.2.2", - "static-link": "^0.2.3", - "vscode-jsonrpc": "^3.3.1" - }, - "bundleDependencies": false, - "peerDependencies": {}, - "deprecated": false, - "_resolved": "/root/.autorest/@microsoft.azure_autorest-core@2.0.4210/node_modules/@microsoft.azure/autorest-core", - "_shasum": "bcf1afb414f3a31bc3c8be1753da1e704cd1242a", - "_shrinkwrap": null, - "bin": { - "autorest-core": "./dist/app.js", - "autorest-language-service": "dist/language-service/language-service.js" - }, - "_id": "@microsoft.azure/autorest-core@2.0.4210", - "_from": "file:/root/.autorest/@microsoft.azure_autorest-core@2.0.4210/node_modules/@microsoft.azure/autorest-core", - "_requested": { - "type": "directory", - "where": "/git-restapi", - "raw": "/root/.autorest/@microsoft.azure_autorest-core@2.0.4210/node_modules/@microsoft.azure/autorest-core", - "rawSpec": "/root/.autorest/@microsoft.azure_autorest-core@2.0.4210/node_modules/@microsoft.azure/autorest-core", - "saveSpec": "file:/root/.autorest/@microsoft.azure_autorest-core@2.0.4210/node_modules/@microsoft.azure/autorest-core", - "fetchSpec": "/root/.autorest/@microsoft.azure_autorest-core@2.0.4210/node_modules/@microsoft.azure/autorest-core" - }, - "_spec": "/root/.autorest/@microsoft.azure_autorest-core@2.0.4210/node_modules/@microsoft.azure/autorest-core", - "_where": "/root/.autorest/@microsoft.azure_autorest-core@2.0.4210/node_modules/@microsoft.azure/autorest-core" - }, - "extensionManager": { - "installationPath": "/root/.autorest", - "sharedLock": { - "name": "/root/.autorest", - "exclusiveLock": { - "name": "_root_.autorest.exclusive-lock", - "options": { - "port": 45234, - "host": "2130706813", - "exclusive": true - }, - "pipe": "/tmp/pipe__root_.autorest.exclusive-lock:45234" - }, - "busyLock": { - "name": "_root_.autorest.busy-lock", - "options": { - "port": 37199, - "host": "2130756895", - "exclusive": true - }, - "pipe": "/tmp/pipe__root_.autorest.busy-lock:37199" - }, - "personalLock": { - "name": "_root_.autorest.2078.370327731491.personal-lock", - "options": { - "port": 40790, - "host": "2130726427", - "exclusive": true - }, - "pipe": "/tmp/pipe__root_.autorest.2078.370327731491.personal-lock:40790" - }, - "file": "/tmp/_root_.autorest.lock" - }, - "dotnetPath": "/root/.dotnet" - }, - "installationPath": "/root/.autorest" - }, - { - "resolvedInfo": null, - "packageMetadata": { - "name": "@microsoft.azure/autorest.modeler", - "version": "2.0.21", - "dependencies": { - "dotnet-2.0.0": "^1.3.2" - }, - "optionalDependencies": {}, - "devDependencies": { - "coffee-script": "^1.11.1", - "dotnet-sdk-2.0.0": "^1.1.1", - "gulp": "^3.9.1", - "gulp-filter": "^5.0.0", - "gulp-line-ending-corrector": "^1.0.1", - "iced-coffee-script": "^108.0.11", - "marked": "^0.3.6", - "marked-terminal": "^2.0.0", - "moment": "^2.17.1", - "run-sequence": "*", - "shx": "^0.2.2", - "through2-parallel": "^0.1.3", - "yargs": "^8.0.2", - "yarn": "^1.0.2" - }, - "bundleDependencies": false, - "peerDependencies": {}, - "deprecated": false, - "_resolved": "/root/.autorest/@microsoft.azure_autorest.modeler@2.0.21/node_modules/@microsoft.azure/autorest.modeler", - "_shasum": "3ce7d3939124b31830be15e5de99b9b7768afb90", - "_shrinkwrap": null, - "bin": null, - "_id": "@microsoft.azure/autorest.modeler@2.0.21", - "_from": "file:/root/.autorest/@microsoft.azure_autorest.modeler@2.0.21/node_modules/@microsoft.azure/autorest.modeler", - "_requested": { - "type": "directory", - "where": "/git-restapi", - "raw": "/root/.autorest/@microsoft.azure_autorest.modeler@2.0.21/node_modules/@microsoft.azure/autorest.modeler", - "rawSpec": "/root/.autorest/@microsoft.azure_autorest.modeler@2.0.21/node_modules/@microsoft.azure/autorest.modeler", - "saveSpec": "file:/root/.autorest/@microsoft.azure_autorest.modeler@2.0.21/node_modules/@microsoft.azure/autorest.modeler", - "fetchSpec": "/root/.autorest/@microsoft.azure_autorest.modeler@2.0.21/node_modules/@microsoft.azure/autorest.modeler" - }, - "_spec": "/root/.autorest/@microsoft.azure_autorest.modeler@2.0.21/node_modules/@microsoft.azure/autorest.modeler", - "_where": "/root/.autorest/@microsoft.azure_autorest.modeler@2.0.21/node_modules/@microsoft.azure/autorest.modeler" - }, - "extensionManager": { - "installationPath": "/root/.autorest", - "sharedLock": { - "name": "/root/.autorest", - "exclusiveLock": { - "name": "_root_.autorest.exclusive-lock", - "options": { - "port": 45234, - "host": "2130706813", - "exclusive": true - }, - "pipe": "/tmp/pipe__root_.autorest.exclusive-lock:45234" - }, - "busyLock": { - "name": "_root_.autorest.busy-lock", - "options": { - "port": 37199, - "host": "2130756895", - "exclusive": true - }, - "pipe": "/tmp/pipe__root_.autorest.busy-lock:37199" - }, - "personalLock": { - "name": "_root_.autorest.2078.370327731491.personal-lock", - "options": { - "port": 40790, - "host": "2130726427", - "exclusive": true - }, - "pipe": "/tmp/pipe__root_.autorest.2078.370327731491.personal-lock:40790" - }, - "file": "/tmp/_root_.autorest.lock" - }, - "dotnetPath": "/root/.dotnet" - }, - "installationPath": "/root/.autorest" - }, - { - "resolvedInfo": null, - "packageMetadata": { - "name": "@microsoft.azure/autorest.modeler", - "version": "2.3.38", - "dependencies": { - "dotnet-2.0.0": "^1.4.4" - }, - "optionalDependencies": {}, - "devDependencies": { - "@microsoft.azure/autorest.testserver": "2.3.1", - "autorest": "^2.0.4201", - "coffee-script": "^1.11.1", - "dotnet-sdk-2.0.0": "^1.4.4", - "gulp": "^3.9.1", - "gulp-filter": "^5.0.0", - "gulp-line-ending-corrector": "^1.0.1", - "iced-coffee-script": "^108.0.11", - "marked": "^0.3.6", - "marked-terminal": "^2.0.0", - "moment": "^2.17.1", - "run-sequence": "*", - "shx": "^0.2.2", - "through2-parallel": "^0.1.3", - "yargs": "^8.0.2", - "yarn": "^1.0.2" - }, - "bundleDependencies": false, - "peerDependencies": {}, - "deprecated": false, - "_resolved": "/root/.autorest/@microsoft.azure_autorest.modeler@2.3.38/node_modules/@microsoft.azure/autorest.modeler", - "_shasum": "903bb77932e4ed1b8bc3b25cc39b167143494f6c", - "_shrinkwrap": null, - "bin": null, - "_id": "@microsoft.azure/autorest.modeler@2.3.38", - "_from": "file:/root/.autorest/@microsoft.azure_autorest.modeler@2.3.38/node_modules/@microsoft.azure/autorest.modeler", - "_requested": { - "type": "directory", - "where": "/git-restapi", - "raw": "/root/.autorest/@microsoft.azure_autorest.modeler@2.3.38/node_modules/@microsoft.azure/autorest.modeler", - "rawSpec": "/root/.autorest/@microsoft.azure_autorest.modeler@2.3.38/node_modules/@microsoft.azure/autorest.modeler", - "saveSpec": "file:/root/.autorest/@microsoft.azure_autorest.modeler@2.3.38/node_modules/@microsoft.azure/autorest.modeler", - "fetchSpec": "/root/.autorest/@microsoft.azure_autorest.modeler@2.3.38/node_modules/@microsoft.azure/autorest.modeler" - }, - "_spec": "/root/.autorest/@microsoft.azure_autorest.modeler@2.3.38/node_modules/@microsoft.azure/autorest.modeler", - "_where": "/root/.autorest/@microsoft.azure_autorest.modeler@2.3.38/node_modules/@microsoft.azure/autorest.modeler" - }, - "extensionManager": { - "installationPath": "/root/.autorest", - "sharedLock": { - "name": "/root/.autorest", - "exclusiveLock": { - "name": "_root_.autorest.exclusive-lock", - "options": { - "port": 45234, - "host": "2130706813", - "exclusive": true - }, - "pipe": "/tmp/pipe__root_.autorest.exclusive-lock:45234" - }, - "busyLock": { - "name": "_root_.autorest.busy-lock", - "options": { - "port": 37199, - "host": "2130756895", - "exclusive": true - }, - "pipe": "/tmp/pipe__root_.autorest.busy-lock:37199" - }, - "personalLock": { - "name": "_root_.autorest.2078.370327731491.personal-lock", - "options": { - "port": 40790, - "host": "2130726427", - "exclusive": true - }, - "pipe": "/tmp/pipe__root_.autorest.2078.370327731491.personal-lock:40790" - }, - "file": "/tmp/_root_.autorest.lock" - }, - "dotnetPath": "/root/.dotnet" - }, - "installationPath": "/root/.autorest" - }, - { - "resolvedInfo": null, - "packageMetadata": { - "name": "@microsoft.azure/autorest.python", - "version": "2.1.26", - "dependencies": { - "dotnet-2.0.0": "^1.4.4" - }, - "optionalDependencies": {}, - "devDependencies": { - "@microsoft.azure/autorest.testserver": "2.3.12", - "autorest": "^2.0.4203", - "coffee-script": "^1.11.1", - "dotnet-sdk-2.0.0": "^1.4.4", - "gulp": "^3.9.1", - "gulp-filter": "^5.0.0", - "gulp-line-ending-corrector": "^1.0.1", - "iced-coffee-script": "^108.0.11", - "marked": "^0.3.6", - "marked-terminal": "^2.0.0", - "moment": "^2.17.1", - "run-sequence": "*", - "shx": "^0.2.2", - "through2-parallel": "^0.1.3", - "yargs": "^8.0.2", - "yarn": "^1.0.2" - }, - "bundleDependencies": false, - "peerDependencies": {}, - "deprecated": false, - "_resolved": "/root/.autorest/@microsoft.azure_autorest.python@2.1.26/node_modules/@microsoft.azure/autorest.python", - "_shasum": "fce552a203c4b359b3f21a194a6deacc8f40ae80", - "_shrinkwrap": null, - "bin": null, - "_id": "@microsoft.azure/autorest.python@2.1.26", - "_from": "file:/root/.autorest/@microsoft.azure_autorest.python@2.1.26/node_modules/@microsoft.azure/autorest.python", - "_requested": { - "type": "directory", - "where": "/git-restapi", - "raw": "/root/.autorest/@microsoft.azure_autorest.python@2.1.26/node_modules/@microsoft.azure/autorest.python", - "rawSpec": "/root/.autorest/@microsoft.azure_autorest.python@2.1.26/node_modules/@microsoft.azure/autorest.python", - "saveSpec": "file:/root/.autorest/@microsoft.azure_autorest.python@2.1.26/node_modules/@microsoft.azure/autorest.python", - "fetchSpec": "/root/.autorest/@microsoft.azure_autorest.python@2.1.26/node_modules/@microsoft.azure/autorest.python" - }, - "_spec": "/root/.autorest/@microsoft.azure_autorest.python@2.1.26/node_modules/@microsoft.azure/autorest.python", - "_where": "/root/.autorest/@microsoft.azure_autorest.python@2.1.26/node_modules/@microsoft.azure/autorest.python" - }, - "extensionManager": { - "installationPath": "/root/.autorest", - "sharedLock": { - "name": "/root/.autorest", - "exclusiveLock": { - "name": "_root_.autorest.exclusive-lock", - "options": { - "port": 45234, - "host": "2130706813", - "exclusive": true - }, - "pipe": "/tmp/pipe__root_.autorest.exclusive-lock:45234" - }, - "busyLock": { - "name": "_root_.autorest.busy-lock", - "options": { - "port": 37199, - "host": "2130756895", - "exclusive": true - }, - "pipe": "/tmp/pipe__root_.autorest.busy-lock:37199" - }, - "personalLock": { - "name": "_root_.autorest.2078.370327731491.personal-lock", - "options": { - "port": 40790, - "host": "2130726427", - "exclusive": true - }, - "pipe": "/tmp/pipe__root_.autorest.2078.370327731491.personal-lock:40790" - }, - "file": "/tmp/_root_.autorest.lock" - }, - "dotnetPath": "/root/.dotnet" - }, - "installationPath": "/root/.autorest" - } - ], - "autorest_bootstrap": {} -} \ No newline at end of file From 7f7d7a6664a25759dcdfb18ae5d3b2110292f261 Mon Sep 17 00:00:00 2001 From: Laurent Mazuel Date: Wed, 30 Jan 2019 14:35:24 -0800 Subject: [PATCH 54/54] Skip ADF tests --- .../tests/test_mgmt_datafactory.py | 17 ++++++++--------- 1 file changed, 8 insertions(+), 9 deletions(-) diff --git a/azure-mgmt-datafactory/tests/test_mgmt_datafactory.py b/azure-mgmt-datafactory/tests/test_mgmt_datafactory.py index 103f735912ad..0d64a5a15b9a 100644 --- a/azure-mgmt-datafactory/tests/test_mgmt_datafactory.py +++ b/azure-mgmt-datafactory/tests/test_mgmt_datafactory.py @@ -14,8 +14,7 @@ from devtools_testutils import AzureMgmtTestCase, ResourceGroupPreparer -# import logging -# logging.basicConfig(level=logging.DEBUG) +raise unittest.SkipTest("Skipping all tests") class MgmtAdfTest(AzureMgmtTestCase): @@ -50,7 +49,7 @@ def create_output_dataset(self, df_name, ls_name, ds_name, resource_group): ls_for_ds = LinkedServiceReference(ls_name) ds_ab = AzureBlobDataset(ls_for_ds, folder_path='entitylogs', file_name='OutputBlobName', format=tx_f) return self._adf_client.datasets.create_or_update(resource_group.name, df_name, ds_name, ds_ab) - + def create_pipeline_with_run(self, df_name, p_name, ls_name, dsin_name, dsout_name, act_name, resource_group): ls = self.create_azureblob_linkedservice(df_name, ls_name, resource_group) dsin = self.create_azureblob_dataset(df_name, ls_name, dsin_name, resource_group) @@ -90,7 +89,7 @@ def create_get_metadata_pipeline_with_run(self, df_name, p_name, ls_name, ds_nam "Dummy": "dummy" } ) - + def wait_for_factory(self, df, resource_group): if not self.is_playback(): while df.provisioning_state != 'Succeeded': @@ -105,12 +104,12 @@ def create_copyactivity_blobtoblob(self, act_name, dsin_name, dsout_name, resour return CopyActivity(act_name, inputs=[dsin_ref], outputs=[dsOut_ref], source=bso, sink=bsi) def create_lookupactivity_blob(self, act_name, ds_name): - bso = BlobSource() - ds_ref = DatasetReference(ds_name) + bso = BlobSource() + ds_ref = DatasetReference(ds_name) return LookupActivity(act_name, source=bso, dataset=ds_ref) def create_getmetadataactivity_blob(self, act_name, ds_name): - ds_ref = DatasetReference(ds_name) + ds_ref = DatasetReference(ds_name) return GetMetadataActivity(act_name, field_list = [], dataset=ds_ref) def create_pipeline(self, df_name, act, p_name, p_params, resource_group): @@ -233,7 +232,7 @@ def test_integrationruntime_regeneratekey(self, resource_group, location): df_name = 'testirregeneratekey' ir_name = 'ir1' df = self.create_datafactory(df_name, resource_group, location) - self.wait_for_factory(df, resource_group) + self.wait_for_factory(df, resource_group) ir = self.create_integrationruntime(df_name, ir_name, resource_group) oldkey = self._adf_client.integration_runtimes.list_auth_keys(resource_group.name, df_name, ir_name).auth_key1 newkey = self._adf_client.integration_runtimes.regenerate_auth_key(resource_group.name, df_name, ir_name, key_name='authKey1').auth_key1 @@ -252,7 +251,7 @@ def test_integrationruntime_create(self, resource_group, location): ir_status = self._adf_client.integration_runtimes.get_status(resource_group.name, df_name, ir_name) self.assertTrue(ir_name == ir_status.name) - + self.clean_datafactory(df_name, resource_group) if __name__ == '__main__':