From e26953f72d878cc6f4c386e281620dd86da7ce22 Mon Sep 17 00:00:00 2001 From: SDKAuto Date: Mon, 22 May 2023 22:04:10 +0000 Subject: [PATCH] CodeGen from PR 23204 in Azure/azure-rest-api-specs EventGridv2 TypeSpec Api Preview (#23204) * start typespec * adding eventgrid typespec for api w/ TODOs * update eventgrid typespec with latest eventgrid v2 operations * don't require content-type if there is no body * Update specification/eventgrid/typespec/main.tsp Co-authored-by: JoshLove-msft <54595583+JoshLove-msft@users.noreply.github.com> * Update specification/eventgrid/typespec/main.tsp Co-authored-by: JoshLove-msft <54595583+JoshLove-msft@users.noreply.github.com> * changing naming of cloudevent and added in data_base64 * openapi.json * Update specification/eventgrid/typespec/main.tsp Co-authored-by: Libba Lawrence * Update specification/eventgrid/typespec/main.tsp Co-authored-by: Libba Lawrence * Update specification/eventgrid/typespec/main.tsp Co-authored-by: Libba Lawrence * lockTokens format, updated json, optional? params * address code review comments * name_change * add @internal for python * Update specification/eventgrid/Azure.Messaging.EventGrid/main.tsp Co-authored-by: JoshLove-msft <54595583+JoshLove-msft@users.noreply.github.com> * move @internal to client.tsp (#23538) * rename (#23565) * [EventGrid Typespec] breaking changes with april release of typespec (#23539) * breaking changes with april release of typespec * unknown type * [EG Typespec] Update Release behavior (#23699) * update behavior * just behavior * Add tspconfig and remove AAD auth (#23717) * add tspconfig * add namespace * remove oauth * [EG TypeSpec] Archboard Comments (#23696) * refactoring off of apiview * keep as int * no duration * aligning ack and release * remove behavioral change * ack to release * initial naming changes * Update ReleaseResult doc Co-authored-by: JoshLove-msft <54595583+JoshLove-msft@users.noreply.github.com> * Update AckResult doc Co-authored-by: JoshLove-msft <54595583+JoshLove-msft@users.noreply.github.com> * versioning twice-- remove one instance --------- Co-authored-by: JoshLove-msft <54595583+JoshLove-msft@users.noreply.github.com> * Address couple of stewardship team feedback. These include: 1. Rename CloudEventEvent to simply CloudEvent, 2. Add more description to the operations including the possible erorr codes, 3. Add PublishResult with empty Json object as successful response for the Publish operation, 4. Others. * Add support for missing Reject operation + adding deliveryAttemptCount to BrokerProperties + Adding query parameter for release operation for deliveryDelayInSeconds * Update failedTokens/SuccessfulTokens Description to address code review comments * Update to match service behavior (#23754) * Update to match service behavior * remove locktoken * [EGv2] Editing unused variables (#23917) * event delivery delay not in preview * remove from url comment * [EGv2] Version dependency on Azure.Core (#23936) * verioning fix * spacing mishap? * [EventGrid] Deliveryattempt change (#23960) * deliveryCount 5/1 * small typo * [EventGrid] Remove internal (#23995) * remove internal * remove client.tsp * remove waitWaitTime (#24078) * move location of json file (#24076) * [Egv2] Encode param (#24080) * encode * remove num default on duration --------- Co-authored-by: Laurent Mazuel * [EGv2] Fix pipeline (#24098) * regen off new commit for encode * reference preview tag * ignore word * update readme to have both apis * update with next autorest * change format to int32 --------- Co-authored-by: Ashraf Hamad Co-authored-by: Laurent Mazuel Co-authored-by: JoshLove-msft <54595583+JoshLove-msft@users.noreply.github.com> Co-authored-by: Ashraf Hamad --- sdk/eventgrid/azure-eventgrid/MANIFEST.in | 6 +- sdk/eventgrid/azure-eventgrid/_meta.json | 15 + .../azure-eventgrid/azure/__init__.py | 2 +- .../azure/eventgrid/__init__.py | 25 +- .../azure/eventgrid/_client.py | 78 + .../azure/eventgrid/_configuration.py | 61 + .../azure/eventgrid/_event_mappings.py | 397 ++-- .../azure/eventgrid/_helpers.py | 46 +- .../azure/eventgrid/_messaging_shared.py | 4 +- .../azure/eventgrid/_model_base.py | 706 ++++++ .../azure/eventgrid/_operations/__init__.py | 19 + .../eventgrid/_operations/_operations.py | 640 ++++++ .../azure/eventgrid/_operations/_patch.py | 20 + .../azure-eventgrid/azure/eventgrid/_patch.py | 20 + .../azure/eventgrid/_policies.py | 3 +- .../azure/eventgrid/_publisher_client.py | 21 +- .../azure/eventgrid/_serialization.py | 1996 +++++++++++++++++ .../azure/eventgrid/_vendor.py | 38 + .../azure/eventgrid/_version.py | 11 +- .../azure/eventgrid/aio/__init__.py | 26 +- .../azure/eventgrid/aio/_client.py | 78 + .../azure/eventgrid/aio/_configuration.py | 61 + .../eventgrid/aio/_operations/__init__.py | 19 + .../eventgrid/aio/_operations/_operations.py | 475 ++++ .../azure/eventgrid/aio/_operations/_patch.py | 20 + .../azure/eventgrid/aio/_patch.py | 20 + .../eventgrid/aio/_publisher_client_async.py | 30 +- .../azure/eventgrid/aio/_vendor.py | 26 + .../azure/eventgrid/models/__init__.py | 38 + .../azure/eventgrid/models/_models.py | 462 ++++ .../azure/eventgrid/models/_patch.py | 20 + .../azure-eventgrid/azure/eventgrid/py.typed | 1 + .../sample_authentication_async.py | 9 +- ...le_publish_cloud_event_using_dict_async.py | 28 +- .../sample_publish_cncf_cloud_events_async.py | 23 +- ..._publish_custom_schema_to_a_topic_async.py | 8 +- ...ample_publish_eg_event_using_dict_async.py | 40 +- ...ple_publish_eg_events_to_a_domain_async.py | 42 +- ...mple_publish_eg_events_to_a_topic_async.py | 25 +- ...s_to_a_topic_using_sas_credential_async.py | 24 +- ...nts_using_cloud_events_1.0_schema_async.py | 25 +- .../sample_publish_to_channel_async.py | 32 +- .../consume_cloud_events_from_eventhub.py | 9 +- ...consume_cloud_events_from_storage_queue.py | 13 +- ...eventgrid_events_from_service_bus_queue.py | 8 +- .../EventGridTrigger1/__init__.py | 17 +- ...ish_cloud_events_to_custom_topic_sample.py | 14 +- ...ish_cloud_events_to_domain_topic_sample.py | 16 +- ...sh_custom_schema_events_to_topic_sample.py | 9 +- ...vent_grid_events_to_custom_topic_sample.py | 17 +- ...ish_with_shared_access_signature_sample.py | 16 +- .../sync_samples/sample_authentication.py | 2 +- .../sync_samples/sample_generate_sas.py | 4 +- .../sample_publish_cloud_event_using_dict.py | 24 +- .../sample_publish_cncf_cloud_events.py | 18 +- ...sample_publish_custom_schema_to_a_topic.py | 6 +- .../sample_publish_eg_event_using_dict.py | 40 +- .../sample_publish_eg_events_to_a_domain.py | 38 +- .../sample_publish_eg_events_to_a_topic.py | 20 +- ..._events_to_a_topic_using_sas_credential.py | 20 +- ...sh_events_using_cloud_events_1.0_schema.py | 20 +- sdk/eventgrid/azure-eventgrid/setup.py | 97 +- .../azure-eventgrid/swagger/_constants.py | 51 +- .../swagger/postprocess_eventnames.py | 9 +- sdk/eventgrid/azure-eventgrid/tests/_mocks.py | 96 +- .../azure-eventgrid/tests/conftest.py | 10 +- .../tests/eventgrid_preparer.py | 28 +- .../tests/perfstress_tests/send.py | 41 +- .../tests/test_cloud_event_tracing.py | 50 +- .../azure-eventgrid/tests/test_cncf_events.py | 16 +- .../tests/test_cncf_events_async.py | 15 +- .../tests/test_eg_event_get_bytes.py | 169 +- .../tests/test_eg_publisher_client.py | 238 +- .../tests/test_eg_publisher_client_async.py | 215 +- .../azure-eventgrid/tests/test_exceptions.py | 38 +- .../tests/test_exceptions_async.py | 32 +- .../tests/test_serialization.py | 109 +- .../azure-eventgrid/tsp-location.yaml | 4 + 78 files changed, 5935 insertions(+), 1234 deletions(-) create mode 100644 sdk/eventgrid/azure-eventgrid/_meta.json create mode 100644 sdk/eventgrid/azure-eventgrid/azure/eventgrid/_client.py create mode 100644 sdk/eventgrid/azure-eventgrid/azure/eventgrid/_configuration.py create mode 100644 sdk/eventgrid/azure-eventgrid/azure/eventgrid/_model_base.py create mode 100644 sdk/eventgrid/azure-eventgrid/azure/eventgrid/_operations/__init__.py create mode 100644 sdk/eventgrid/azure-eventgrid/azure/eventgrid/_operations/_operations.py create mode 100644 sdk/eventgrid/azure-eventgrid/azure/eventgrid/_operations/_patch.py create mode 100644 sdk/eventgrid/azure-eventgrid/azure/eventgrid/_patch.py create mode 100644 sdk/eventgrid/azure-eventgrid/azure/eventgrid/_serialization.py create mode 100644 sdk/eventgrid/azure-eventgrid/azure/eventgrid/_vendor.py create mode 100644 sdk/eventgrid/azure-eventgrid/azure/eventgrid/aio/_client.py create mode 100644 sdk/eventgrid/azure-eventgrid/azure/eventgrid/aio/_configuration.py create mode 100644 sdk/eventgrid/azure-eventgrid/azure/eventgrid/aio/_operations/__init__.py create mode 100644 sdk/eventgrid/azure-eventgrid/azure/eventgrid/aio/_operations/_operations.py create mode 100644 sdk/eventgrid/azure-eventgrid/azure/eventgrid/aio/_operations/_patch.py create mode 100644 sdk/eventgrid/azure-eventgrid/azure/eventgrid/aio/_patch.py create mode 100644 sdk/eventgrid/azure-eventgrid/azure/eventgrid/aio/_vendor.py create mode 100644 sdk/eventgrid/azure-eventgrid/azure/eventgrid/models/__init__.py create mode 100644 sdk/eventgrid/azure-eventgrid/azure/eventgrid/models/_models.py create mode 100644 sdk/eventgrid/azure-eventgrid/azure/eventgrid/models/_patch.py create mode 100644 sdk/eventgrid/azure-eventgrid/tsp-location.yaml diff --git a/sdk/eventgrid/azure-eventgrid/MANIFEST.in b/sdk/eventgrid/azure-eventgrid/MANIFEST.in index a4f0c46bcd94..8aee6afa5284 100644 --- a/sdk/eventgrid/azure-eventgrid/MANIFEST.in +++ b/sdk/eventgrid/azure-eventgrid/MANIFEST.in @@ -1,6 +1,6 @@ -recursive-include tests *.py *.yaml -recursive-include samples *.py include *.md include LICENSE -include azure/__init__.py include azure/eventgrid/py.typed +recursive-include tests *.py +recursive-include samples *.py *.md +include azure/__init__.py \ No newline at end of file diff --git a/sdk/eventgrid/azure-eventgrid/_meta.json b/sdk/eventgrid/azure-eventgrid/_meta.json new file mode 100644 index 000000000000..6882bb450bf2 --- /dev/null +++ b/sdk/eventgrid/azure-eventgrid/_meta.json @@ -0,0 +1,15 @@ +{ + "commit": "f6f647a40fbbc43dd6e65c4436448129ab7e1fdd", + "repository_url": "https://github.com/Azure/azure-rest-api-specs", + "typespec_src": "specification/eventgrid/Azure.Messaging.EventGrid", + "@azure-tools/typespec-python": { + "version": "0.8.6", + "resolved": "https://registry.npmjs.org/@azure-tools/typespec-python/-/typespec-python-0.8.6.tgz", + "dependencies": { + "@autorest/python": { + "version": "6.4.15", + "resolved": "https://registry.npmjs.org/@autorest/python/-/python-6.4.15.tgz" + } + } + } +} \ No newline at end of file diff --git a/sdk/eventgrid/azure-eventgrid/azure/__init__.py b/sdk/eventgrid/azure-eventgrid/azure/__init__.py index 0c36c2076ba0..d55ccad1f573 100644 --- a/sdk/eventgrid/azure-eventgrid/azure/__init__.py +++ b/sdk/eventgrid/azure-eventgrid/azure/__init__.py @@ -1 +1 @@ -__path__ = __import__('pkgutil').extend_path(__path__, __name__) # type: ignore +__path__ = __import__("pkgutil").extend_path(__path__, __name__) # type: ignore diff --git a/sdk/eventgrid/azure-eventgrid/azure/eventgrid/__init__.py b/sdk/eventgrid/azure-eventgrid/azure/eventgrid/__init__.py index 1dc3655a13bb..78d20e5468da 100644 --- a/sdk/eventgrid/azure-eventgrid/azure/eventgrid/__init__.py +++ b/sdk/eventgrid/azure-eventgrid/azure/eventgrid/__init__.py @@ -2,18 +2,25 @@ # -------------------------------------------------------------------------- # Copyright (c) Microsoft Corporation. All rights reserved. # Licensed under the MIT License. See License.txt in the project root for license information. +# Code generated by Microsoft (R) Python Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is regenerated. # -------------------------------------------------------------------------- -from ._publisher_client import EventGridPublisherClient -from ._event_mappings import SystemEventNames -from ._helpers import generate_sas -from ._models import EventGridEvent +from ._client import EventGridClient from ._version import VERSION +__version__ = VERSION + +try: + from ._patch import __all__ as _patch_all + from ._patch import * # pylint: disable=unused-wildcard-import +except ImportError: + _patch_all = [] +from ._patch import patch_sdk as _patch_sdk + __all__ = [ - "EventGridPublisherClient", - "EventGridEvent", - "generate_sas", - "SystemEventNames", + "EventGridClient", ] -__version__ = VERSION +__all__.extend([p for p in _patch_all if p not in __all__]) + +_patch_sdk() diff --git a/sdk/eventgrid/azure-eventgrid/azure/eventgrid/_client.py b/sdk/eventgrid/azure-eventgrid/azure/eventgrid/_client.py new file mode 100644 index 000000000000..663d5a0620e1 --- /dev/null +++ b/sdk/eventgrid/azure-eventgrid/azure/eventgrid/_client.py @@ -0,0 +1,78 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for license information. +# Code generated by Microsoft (R) Python Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is regenerated. +# -------------------------------------------------------------------------- + +from copy import deepcopy +from typing import Any + +from azure.core import PipelineClient +from azure.core.credentials import AzureKeyCredential +from azure.core.rest import HttpRequest, HttpResponse + +from ._configuration import EventGridClientConfiguration +from ._operations import EventGridClientOperationsMixin +from ._serialization import Deserializer, Serializer + + +class EventGridClient(EventGridClientOperationsMixin): # pylint: disable=client-accepts-api-version-keyword + """Azure Messaging EventGrid Client. + + :param endpoint: The host name of the namespace, e.g. + namespaceName1.westus-1.eventgrid.azure.net. Required. + :type endpoint: str + :param credential: Credential needed for the client to connect to Azure. Required. + :type credential: ~azure.core.credentials.AzureKeyCredential + :keyword api_version: The API version to use for this operation. Default value is + "2023-06-01-preview". Note that overriding this default value may result in unsupported + behavior. + :paramtype api_version: str + """ + + def __init__(self, endpoint: str, credential: AzureKeyCredential, **kwargs: Any) -> None: + _endpoint = "{endpoint}" + self._config = EventGridClientConfiguration(endpoint=endpoint, credential=credential, **kwargs) + self._client: PipelineClient = PipelineClient(base_url=_endpoint, config=self._config, **kwargs) + + self._serialize = Serializer() + self._deserialize = Deserializer() + self._serialize.client_side_validation = False + + def send_request(self, request: HttpRequest, **kwargs: Any) -> HttpResponse: + """Runs the network request through the client's chained policies. + + >>> from azure.core.rest import HttpRequest + >>> request = HttpRequest("GET", "https://www.example.org/") + + >>> response = client.send_request(request) + + + For more information on this code flow, see https://aka.ms/azsdk/dpcodegen/python/send_request + + :param request: The network request you want to make. Required. + :type request: ~azure.core.rest.HttpRequest + :keyword bool stream: Whether the response payload will be streamed. Defaults to False. + :return: The response of your network call. Does not do error handling on your response. + :rtype: ~azure.core.rest.HttpResponse + """ + + request_copy = deepcopy(request) + path_format_arguments = { + "endpoint": self._serialize.url("self._config.endpoint", self._config.endpoint, "str", skip_quote=True), + } + + request_copy.url = self._client.format_url(request_copy.url, **path_format_arguments) + return self._client.send_request(request_copy, **kwargs) + + def close(self) -> None: + self._client.close() + + def __enter__(self) -> "EventGridClient": + self._client.__enter__() + return self + + def __exit__(self, *exc_details: Any) -> None: + self._client.__exit__(*exc_details) diff --git a/sdk/eventgrid/azure-eventgrid/azure/eventgrid/_configuration.py b/sdk/eventgrid/azure-eventgrid/azure/eventgrid/_configuration.py new file mode 100644 index 000000000000..4f9f49ec55b2 --- /dev/null +++ b/sdk/eventgrid/azure-eventgrid/azure/eventgrid/_configuration.py @@ -0,0 +1,61 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for license information. +# Code generated by Microsoft (R) Python Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is regenerated. +# -------------------------------------------------------------------------- + +from typing import Any + +from azure.core.configuration import Configuration +from azure.core.credentials import AzureKeyCredential +from azure.core.pipeline import policies + +from ._version import VERSION + + +class EventGridClientConfiguration(Configuration): # pylint: disable=too-many-instance-attributes + """Configuration for EventGridClient. + + Note that all parameters used to create this instance are saved as instance + attributes. + + :param endpoint: The host name of the namespace, e.g. + namespaceName1.westus-1.eventgrid.azure.net. Required. + :type endpoint: str + :param credential: Credential needed for the client to connect to Azure. Required. + :type credential: ~azure.core.credentials.AzureKeyCredential + :keyword api_version: The API version to use for this operation. Default value is + "2023-06-01-preview". Note that overriding this default value may result in unsupported + behavior. + :paramtype api_version: str + """ + + def __init__(self, endpoint: str, credential: AzureKeyCredential, **kwargs: Any) -> None: + super(EventGridClientConfiguration, self).__init__(**kwargs) + api_version: str = kwargs.pop("api_version", "2023-06-01-preview") + + if endpoint is None: + raise ValueError("Parameter 'endpoint' must not be None.") + if credential is None: + raise ValueError("Parameter 'credential' must not be None.") + + self.endpoint = endpoint + self.credential = credential + self.api_version = api_version + kwargs.setdefault("sdk_moniker", "eventgrid/{}".format(VERSION)) + self._configure(**kwargs) + + def _configure(self, **kwargs: Any) -> None: + self.user_agent_policy = kwargs.get("user_agent_policy") or policies.UserAgentPolicy(**kwargs) + self.headers_policy = kwargs.get("headers_policy") or policies.HeadersPolicy(**kwargs) + self.proxy_policy = kwargs.get("proxy_policy") or policies.ProxyPolicy(**kwargs) + self.logging_policy = kwargs.get("logging_policy") or policies.NetworkTraceLoggingPolicy(**kwargs) + self.http_logging_policy = kwargs.get("http_logging_policy") or policies.HttpLoggingPolicy(**kwargs) + self.retry_policy = kwargs.get("retry_policy") or policies.RetryPolicy(**kwargs) + self.custom_hook_policy = kwargs.get("custom_hook_policy") or policies.CustomHookPolicy(**kwargs) + self.redirect_policy = kwargs.get("redirect_policy") or policies.RedirectPolicy(**kwargs) + self.authentication_policy = kwargs.get("authentication_policy") + if self.credential and not self.authentication_policy: + self.authentication_policy = policies.AzureKeyCredentialPolicy(self.credential, "SharedAccessKey", **kwargs) diff --git a/sdk/eventgrid/azure-eventgrid/azure/eventgrid/_event_mappings.py b/sdk/eventgrid/azure-eventgrid/azure/eventgrid/_event_mappings.py index 3c5872b4f03a..5795cc94c5a0 100644 --- a/sdk/eventgrid/azure-eventgrid/azure/eventgrid/_event_mappings.py +++ b/sdk/eventgrid/azure-eventgrid/azure/eventgrid/_event_mappings.py @@ -15,380 +15,405 @@ class SystemEventNames(str, Enum, metaclass=CaseInsensitiveEnumMeta): Azure Event Grid. To check the list of recognizable system topics, visit https://docs.microsoft.com/azure/event-grid/system-topics. """ + # These names at the top are 'corrected' aliases of duplicate values that appear below, which are # deprecated but maintained for backwards compatibility. - AcsChatMemberAddedToThreadWithUserEventName = 'Microsoft.Communication.ChatMemberAddedToThreadWithUser' + AcsChatMemberAddedToThreadWithUserEventName = "Microsoft.Communication.ChatMemberAddedToThreadWithUser" - ResourceWriteFailureEventName = 'Microsoft.Resources.ResourceWriteFailure' + ResourceWriteFailureEventName = "Microsoft.Resources.ResourceWriteFailure" - IoTHubDeviceDeletedEventName = 'Microsoft.Devices.DeviceDeleted' + IoTHubDeviceDeletedEventName = "Microsoft.Devices.DeviceDeleted" - IoTHubDeviceDisconnectedEventName = 'Microsoft.Devices.DeviceDisconnected' + IoTHubDeviceDisconnectedEventName = "Microsoft.Devices.DeviceDisconnected" - ResourceDeleteFailureEventName = 'Microsoft.Resources.ResourceDeleteFailure' + ResourceDeleteFailureEventName = "Microsoft.Resources.ResourceDeleteFailure" - ResourceDeleteCancelEventName = 'Microsoft.Resources.ResourceDeleteCancel' + ResourceDeleteCancelEventName = "Microsoft.Resources.ResourceDeleteCancel" - AcsChatThreadParticipantAddedEventName = 'Microsoft.Communication.ChatThreadParticipantAdded' + AcsChatThreadParticipantAddedEventName = "Microsoft.Communication.ChatThreadParticipantAdded" - ResourceDeleteSuccessEventName = 'Microsoft.Resources.ResourceDeleteSuccess' + ResourceDeleteSuccessEventName = "Microsoft.Resources.ResourceDeleteSuccess" - EventGridSubscriptionValidationEventName = 'Microsoft.EventGrid.SubscriptionValidationEvent' + EventGridSubscriptionValidationEventName = "Microsoft.EventGrid.SubscriptionValidationEvent" - ResourceWriteSuccessEventName = 'Microsoft.Resources.ResourceWriteSuccess' + ResourceWriteSuccessEventName = "Microsoft.Resources.ResourceWriteSuccess" - ResourceActionSuccessEventName = 'Microsoft.Resources.ResourceActionSuccess' + ResourceActionSuccessEventName = "Microsoft.Resources.ResourceActionSuccess" - ResourceWriteCancelEventName = 'Microsoft.Resources.ResourceWriteCancel' + ResourceWriteCancelEventName = "Microsoft.Resources.ResourceWriteCancel" - ResourceActionFailureEventName = 'Microsoft.Resources.ResourceActionFailure' + ResourceActionFailureEventName = "Microsoft.Resources.ResourceActionFailure" - AcsChatMemberRemovedFromThreadWithUserEventName = 'Microsoft.Communication.ChatMemberRemovedFromThreadWithUser' + AcsChatMemberRemovedFromThreadWithUserEventName = "Microsoft.Communication.ChatMemberRemovedFromThreadWithUser" - IoTHubDeviceConnectedEventName = 'Microsoft.Devices.DeviceConnected' + IoTHubDeviceConnectedEventName = "Microsoft.Devices.DeviceConnected" - EventGridSubscriptionDeletedEventName = 'Microsoft.EventGrid.SubscriptionDeletedEvent' + EventGridSubscriptionDeletedEventName = "Microsoft.EventGrid.SubscriptionDeletedEvent" - AcsChatThreadParticipantRemovedEventName = 'Microsoft.Communication.ChatThreadParticipantRemoved' + AcsChatThreadParticipantRemovedEventName = "Microsoft.Communication.ChatThreadParticipantRemoved" - ResourceActionCancelEventName = 'Microsoft.Resources.ResourceActionCancel' + ResourceActionCancelEventName = "Microsoft.Resources.ResourceActionCancel" - IoTHubDeviceCreatedEventName = 'Microsoft.Devices.DeviceCreated' + IoTHubDeviceCreatedEventName = "Microsoft.Devices.DeviceCreated" # Aliases end here - AcsChatMessageDeletedEventName = 'Microsoft.Communication.ChatMessageDeleted' + AcsChatMessageDeletedEventName = "Microsoft.Communication.ChatMessageDeleted" - AcsChatMessageDeletedInThreadEventName = 'Microsoft.Communication.ChatMessageDeletedInThread' + AcsChatMessageDeletedInThreadEventName = "Microsoft.Communication.ChatMessageDeletedInThread" - AcsChatMessageEditedEventName = 'Microsoft.Communication.ChatMessageEdited' + AcsChatMessageEditedEventName = "Microsoft.Communication.ChatMessageEdited" - AcsChatMessageEditedInThreadEventName = 'Microsoft.Communication.ChatMessageEditedInThread' + AcsChatMessageEditedInThreadEventName = "Microsoft.Communication.ChatMessageEditedInThread" - AcsChatMessageReceivedEventName = 'Microsoft.Communication.ChatMessageReceived' + AcsChatMessageReceivedEventName = "Microsoft.Communication.ChatMessageReceived" - AcsChatMessageReceivedInThreadEventName = 'Microsoft.Communication.ChatMessageReceivedInThread' + AcsChatMessageReceivedInThreadEventName = "Microsoft.Communication.ChatMessageReceivedInThread" - AcsChatParticipantAddedToThreadEventName = 'Microsoft.Communication.ChatThreadParticipantAdded' + AcsChatParticipantAddedToThreadEventName = "Microsoft.Communication.ChatThreadParticipantAdded" - AcsChatParticipantAddedToThreadWithUserEventName = 'Microsoft.Communication.ChatParticipantAddedToThreadWithUser' + AcsChatParticipantAddedToThreadWithUserEventName = "Microsoft.Communication.ChatParticipantAddedToThreadWithUser" - AcsChatParticipantRemovedFromThreadEventName = 'Microsoft.Communication.ChatThreadParticipantRemoved' + AcsChatParticipantRemovedFromThreadEventName = "Microsoft.Communication.ChatThreadParticipantRemoved" - AcsChatParticipantRemovedFromThreadWithUserEventName = 'Microsoft.Communication.ChatParticipantRemovedFromThreadWithUser' + AcsChatParticipantRemovedFromThreadWithUserEventName = ( + "Microsoft.Communication.ChatParticipantRemovedFromThreadWithUser" + ) - AcsChatThreadCreatedEventName = 'Microsoft.Communication.ChatThreadCreated' + AcsChatThreadCreatedEventName = "Microsoft.Communication.ChatThreadCreated" - AcsChatThreadCreatedWithUserEventName = 'Microsoft.Communication.ChatThreadCreatedWithUser' + AcsChatThreadCreatedWithUserEventName = "Microsoft.Communication.ChatThreadCreatedWithUser" - AcsChatThreadDeletedEventName = 'Microsoft.Communication.ChatThreadDeleted' + AcsChatThreadDeletedEventName = "Microsoft.Communication.ChatThreadDeleted" - AcsChatThreadPropertiesUpdatedEventName = 'Microsoft.Communication.ChatThreadPropertiesUpdated' + AcsChatThreadPropertiesUpdatedEventName = "Microsoft.Communication.ChatThreadPropertiesUpdated" - AcsChatThreadPropertiesUpdatedPerUserEventName = 'Microsoft.Communication.ChatThreadPropertiesUpdatedPerUser' + AcsChatThreadPropertiesUpdatedPerUserEventName = "Microsoft.Communication.ChatThreadPropertiesUpdatedPerUser" - AcsChatThreadWithUserDeletedEventName = 'Microsoft.Communication.ChatThreadWithUserDeleted' + AcsChatThreadWithUserDeletedEventName = "Microsoft.Communication.ChatThreadWithUserDeleted" - AcsEmailDeliveryReportReceivedEventName = 'Microsoft.Communication.EmailDeliveryReportReceived' + AcsEmailDeliveryReportReceivedEventName = "Microsoft.Communication.EmailDeliveryReportReceived" - AcsEmailEngagementTrackingReportReceivedEventName = 'Microsoft.Communication.EmailEngagementTrackingReportReceived' + AcsEmailEngagementTrackingReportReceivedEventName = "Microsoft.Communication.EmailEngagementTrackingReportReceived" - AcsIncomingCallEventName = 'Microsoft.Communication.IncomingCall' + AcsIncomingCallEventName = "Microsoft.Communication.IncomingCall" - AcsRecordingFileStatusUpdatedEventName = 'Microsoft.Communication.RecordingFileStatusUpdated' + AcsRecordingFileStatusUpdatedEventName = "Microsoft.Communication.RecordingFileStatusUpdated" - AcsSmsDeliveryReportReceivedEventName = 'Microsoft.Communication.SMSDeliveryReportReceived' + AcsSmsDeliveryReportReceivedEventName = "Microsoft.Communication.SMSDeliveryReportReceived" - AcsSmsReceivedEventName = 'Microsoft.Communication.SMSReceived' + AcsSmsReceivedEventName = "Microsoft.Communication.SMSReceived" - AcsUserDisconnectedEventName = 'Microsoft.Communication.UserDisconnected' + AcsUserDisconnectedEventName = "Microsoft.Communication.UserDisconnected" - ApiManagementApiCreatedEventName = 'Microsoft.ApiManagement.APICreated' + ApiManagementApiCreatedEventName = "Microsoft.ApiManagement.APICreated" - ApiManagementApiDeletedEventName = 'Microsoft.ApiManagement.APIDeleted' + ApiManagementApiDeletedEventName = "Microsoft.ApiManagement.APIDeleted" - ApiManagementApiReleaseCreatedEventName = 'Microsoft.ApiManagement.APIReleaseCreated' + ApiManagementApiReleaseCreatedEventName = "Microsoft.ApiManagement.APIReleaseCreated" - ApiManagementApiReleaseDeletedEventName = 'Microsoft.ApiManagement.APIReleaseDeleted' + ApiManagementApiReleaseDeletedEventName = "Microsoft.ApiManagement.APIReleaseDeleted" - ApiManagementApiReleaseUpdatedEventName = 'Microsoft.ApiManagement.APIReleaseUpdated' + ApiManagementApiReleaseUpdatedEventName = "Microsoft.ApiManagement.APIReleaseUpdated" - ApiManagementApiUpdatedEventName = 'Microsoft.ApiManagement.APIUpdated' + ApiManagementApiUpdatedEventName = "Microsoft.ApiManagement.APIUpdated" - ApiManagementGatewayApiAddedEventName = 'Microsoft.ApiManagement.GatewayAPIAdded' + ApiManagementGatewayApiAddedEventName = "Microsoft.ApiManagement.GatewayAPIAdded" - ApiManagementGatewayApiRemovedEventName = 'Microsoft.ApiManagement.GatewayAPIRemoved' + ApiManagementGatewayApiRemovedEventName = "Microsoft.ApiManagement.GatewayAPIRemoved" - ApiManagementGatewayCertificateAuthorityCreatedEventName = 'Microsoft.ApiManagement.GatewayCertificateAuthorityCreated' + ApiManagementGatewayCertificateAuthorityCreatedEventName = ( + "Microsoft.ApiManagement.GatewayCertificateAuthorityCreated" + ) - ApiManagementGatewayCertificateAuthorityDeletedEventName = 'Microsoft.ApiManagement.GatewayCertificateAuthorityDeleted' + ApiManagementGatewayCertificateAuthorityDeletedEventName = ( + "Microsoft.ApiManagement.GatewayCertificateAuthorityDeleted" + ) - ApiManagementGatewayCertificateAuthorityUpdatedEventName = 'Microsoft.ApiManagement.GatewayCertificateAuthorityUpdated' + ApiManagementGatewayCertificateAuthorityUpdatedEventName = ( + "Microsoft.ApiManagement.GatewayCertificateAuthorityUpdated" + ) - ApiManagementGatewayCreatedEventName = 'Microsoft.ApiManagement.GatewayCreated' + ApiManagementGatewayCreatedEventName = "Microsoft.ApiManagement.GatewayCreated" - ApiManagementGatewayDeletedEventName = 'Microsoft.ApiManagement.GatewayDeleted' + ApiManagementGatewayDeletedEventName = "Microsoft.ApiManagement.GatewayDeleted" - ApiManagementGatewayHostnameConfigurationCreatedEventName = 'Microsoft.ApiManagement.GatewayHostnameConfigurationCreated' + ApiManagementGatewayHostnameConfigurationCreatedEventName = ( + "Microsoft.ApiManagement.GatewayHostnameConfigurationCreated" + ) - ApiManagementGatewayHostnameConfigurationDeletedEventName = 'Microsoft.ApiManagement.GatewayHostnameConfigurationDeleted' + ApiManagementGatewayHostnameConfigurationDeletedEventName = ( + "Microsoft.ApiManagement.GatewayHostnameConfigurationDeleted" + ) - ApiManagementGatewayHostnameConfigurationUpdatedEventName = 'Microsoft.ApiManagement.GatewayHostnameConfigurationUpdated' + ApiManagementGatewayHostnameConfigurationUpdatedEventName = ( + "Microsoft.ApiManagement.GatewayHostnameConfigurationUpdated" + ) - ApiManagementGatewayUpdatedEventName = 'Microsoft.ApiManagement.GatewayUpdated' + ApiManagementGatewayUpdatedEventName = "Microsoft.ApiManagement.GatewayUpdated" - ApiManagementProductCreatedEventName = 'Microsoft.ApiManagement.ProductCreated' + ApiManagementProductCreatedEventName = "Microsoft.ApiManagement.ProductCreated" - ApiManagementProductDeletedEventName = 'Microsoft.ApiManagement.ProductDeleted' + ApiManagementProductDeletedEventName = "Microsoft.ApiManagement.ProductDeleted" - ApiManagementProductUpdatedEventName = 'Microsoft.ApiManagement.ProductUpdated' + ApiManagementProductUpdatedEventName = "Microsoft.ApiManagement.ProductUpdated" - ApiManagementSubscriptionCreatedEventName = 'Microsoft.ApiManagement.SubscriptionCreated' + ApiManagementSubscriptionCreatedEventName = "Microsoft.ApiManagement.SubscriptionCreated" - ApiManagementSubscriptionDeletedEventName = 'Microsoft.ApiManagement.SubscriptionDeleted' + ApiManagementSubscriptionDeletedEventName = "Microsoft.ApiManagement.SubscriptionDeleted" - ApiManagementSubscriptionUpdatedEventName = 'Microsoft.ApiManagement.SubscriptionUpdated' + ApiManagementSubscriptionUpdatedEventName = "Microsoft.ApiManagement.SubscriptionUpdated" - ApiManagementUserCreatedEventName = 'Microsoft.ApiManagement.UserCreated' + ApiManagementUserCreatedEventName = "Microsoft.ApiManagement.UserCreated" - ApiManagementUserDeletedEventName = 'Microsoft.ApiManagement.UserDeleted' + ApiManagementUserDeletedEventName = "Microsoft.ApiManagement.UserDeleted" - ApiManagementUserUpdatedEventName = 'Microsoft.ApiManagement.UserUpdated' + ApiManagementUserUpdatedEventName = "Microsoft.ApiManagement.UserUpdated" - AppConfigurationKeyValueDeletedEventName = 'Microsoft.AppConfiguration.KeyValueDeleted' + AppConfigurationKeyValueDeletedEventName = "Microsoft.AppConfiguration.KeyValueDeleted" - AppConfigurationKeyValueModifiedEventName = 'Microsoft.AppConfiguration.KeyValueModified' + AppConfigurationKeyValueModifiedEventName = "Microsoft.AppConfiguration.KeyValueModified" - ContainerRegistryChartDeletedEventName = 'Microsoft.ContainerRegistry.ChartDeleted' + ContainerRegistryChartDeletedEventName = "Microsoft.ContainerRegistry.ChartDeleted" - ContainerRegistryChartPushedEventName = 'Microsoft.ContainerRegistry.ChartPushed' + ContainerRegistryChartPushedEventName = "Microsoft.ContainerRegistry.ChartPushed" - ContainerRegistryImageDeletedEventName = 'Microsoft.ContainerRegistry.ImageDeleted' + ContainerRegistryImageDeletedEventName = "Microsoft.ContainerRegistry.ImageDeleted" - ContainerRegistryImagePushedEventName = 'Microsoft.ContainerRegistry.ImagePushed' + ContainerRegistryImagePushedEventName = "Microsoft.ContainerRegistry.ImagePushed" - ContainerServiceNewKubernetesVersionAvailableEventName = 'Microsoft.ContainerService.NewKubernetesVersionAvailable' + ContainerServiceNewKubernetesVersionAvailableEventName = "Microsoft.ContainerService.NewKubernetesVersionAvailable" - DataBoxCopyCompletedEventName = 'Microsoft.DataBox.CopyCompleted' + DataBoxCopyCompletedEventName = "Microsoft.DataBox.CopyCompleted" - DataBoxCopyStartedEventName = 'Microsoft.DataBox.CopyStarted' + DataBoxCopyStartedEventName = "Microsoft.DataBox.CopyStarted" - DataBoxOrderCompletedEventName = 'Microsoft.DataBox.OrderCompleted' + DataBoxOrderCompletedEventName = "Microsoft.DataBox.OrderCompleted" - EventHubCaptureFileCreatedEventName = 'Microsoft.EventHub.CaptureFileCreated' + EventHubCaptureFileCreatedEventName = "Microsoft.EventHub.CaptureFileCreated" # spell-checker:ignore Dicom - HealthcareDicomImageCreatedEventName = 'Microsoft.HealthcareApis.DicomImageCreated' + HealthcareDicomImageCreatedEventName = "Microsoft.HealthcareApis.DicomImageCreated" - HealthcareDicomImageDeletedEventName = 'Microsoft.HealthcareApis.DicomImageDeleted' + HealthcareDicomImageDeletedEventName = "Microsoft.HealthcareApis.DicomImageDeleted" - HealthcareDicomImageUpdatedEventName = 'Microsoft.HealthcareApis.DicomImageUpdated' + HealthcareDicomImageUpdatedEventName = "Microsoft.HealthcareApis.DicomImageUpdated" # spell-checker:ignore Fhir - HealthcareFhirResourceCreatedEventName = 'Microsoft.HealthcareApis.FhirResourceCreated' + HealthcareFhirResourceCreatedEventName = "Microsoft.HealthcareApis.FhirResourceCreated" - HealthcareFhirResourceDeletedEventName = 'Microsoft.HealthcareApis.FhirResourceDeleted' + HealthcareFhirResourceDeletedEventName = "Microsoft.HealthcareApis.FhirResourceDeleted" - HealthcareFhirResourceUpdatedEventName = 'Microsoft.HealthcareApis.FhirResourceUpdated' + HealthcareFhirResourceUpdatedEventName = "Microsoft.HealthcareApis.FhirResourceUpdated" - IotHubDeviceConnectedEventName = 'Microsoft.Devices.DeviceConnected' + IotHubDeviceConnectedEventName = "Microsoft.Devices.DeviceConnected" - IotHubDeviceCreatedEventName = 'Microsoft.Devices.DeviceCreated' + IotHubDeviceCreatedEventName = "Microsoft.Devices.DeviceCreated" - IotHubDeviceDeletedEventName = 'Microsoft.Devices.DeviceDeleted' + IotHubDeviceDeletedEventName = "Microsoft.Devices.DeviceDeleted" - IotHubDeviceDisconnectedEventName = 'Microsoft.Devices.DeviceDisconnected' + IotHubDeviceDisconnectedEventName = "Microsoft.Devices.DeviceDisconnected" - IotHubDeviceTelemetryEventName = 'Microsoft.Devices.DeviceTelemetry' + IotHubDeviceTelemetryEventName = "Microsoft.Devices.DeviceTelemetry" - KeyVaultCertificateExpiredEventName = 'Microsoft.KeyVault.CertificateExpired' + KeyVaultCertificateExpiredEventName = "Microsoft.KeyVault.CertificateExpired" - KeyVaultCertificateNearExpiryEventName = 'Microsoft.KeyVault.CertificateNearExpiry' + KeyVaultCertificateNearExpiryEventName = "Microsoft.KeyVault.CertificateNearExpiry" - KeyVaultCertificateNewVersionCreatedEventName = 'Microsoft.KeyVault.CertificateNewVersionCreated' + KeyVaultCertificateNewVersionCreatedEventName = "Microsoft.KeyVault.CertificateNewVersionCreated" - KeyVaultKeyExpiredEventName = 'Microsoft.KeyVault.KeyExpired' + KeyVaultKeyExpiredEventName = "Microsoft.KeyVault.KeyExpired" - KeyVaultKeyNearExpiryEventName = 'Microsoft.KeyVault.KeyNearExpiry' + KeyVaultKeyNearExpiryEventName = "Microsoft.KeyVault.KeyNearExpiry" - KeyVaultKeyNewVersionCreatedEventName = 'Microsoft.KeyVault.KeyNewVersionCreated' + KeyVaultKeyNewVersionCreatedEventName = "Microsoft.KeyVault.KeyNewVersionCreated" - KeyVaultSecretExpiredEventName = 'Microsoft.KeyVault.SecretExpired' + KeyVaultSecretExpiredEventName = "Microsoft.KeyVault.SecretExpired" - KeyVaultSecretNearExpiryEventName = 'Microsoft.KeyVault.SecretNearExpiry' + KeyVaultSecretNearExpiryEventName = "Microsoft.KeyVault.SecretNearExpiry" - KeyVaultSecretNewVersionCreatedEventName = 'Microsoft.KeyVault.SecretNewVersionCreated' + KeyVaultSecretNewVersionCreatedEventName = "Microsoft.KeyVault.SecretNewVersionCreated" - KeyVaultVaultAccessPolicyChangedEventName = 'Microsoft.KeyVault.VaultAccessPolicyChanged' + KeyVaultVaultAccessPolicyChangedEventName = "Microsoft.KeyVault.VaultAccessPolicyChanged" - MachineLearningServicesDatasetDriftDetectedEventName = 'Microsoft.MachineLearningServices.DatasetDriftDetected' + MachineLearningServicesDatasetDriftDetectedEventName = "Microsoft.MachineLearningServices.DatasetDriftDetected" - MachineLearningServicesModelDeployedEventName = 'Microsoft.MachineLearningServices.ModelDeployed' + MachineLearningServicesModelDeployedEventName = "Microsoft.MachineLearningServices.ModelDeployed" - MachineLearningServicesModelRegisteredEventName = 'Microsoft.MachineLearningServices.ModelRegistered' + MachineLearningServicesModelRegisteredEventName = "Microsoft.MachineLearningServices.ModelRegistered" - MachineLearningServicesRunCompletedEventName = 'Microsoft.MachineLearningServices.RunCompleted' + MachineLearningServicesRunCompletedEventName = "Microsoft.MachineLearningServices.RunCompleted" - MachineLearningServicesRunStatusChangedEventName = 'Microsoft.MachineLearningServices.RunStatusChanged' + MachineLearningServicesRunStatusChangedEventName = "Microsoft.MachineLearningServices.RunStatusChanged" - MapsGeofenceEnteredEventName = 'Microsoft.Maps.GeofenceEntered' + MapsGeofenceEnteredEventName = "Microsoft.Maps.GeofenceEntered" - MapsGeofenceExitedEventName = 'Microsoft.Maps.GeofenceExited' + MapsGeofenceExitedEventName = "Microsoft.Maps.GeofenceExited" - MapsGeofenceResultEventName = 'Microsoft.Maps.GeofenceResult' + MapsGeofenceResultEventName = "Microsoft.Maps.GeofenceResult" - MediaJobCanceledEventName = 'Microsoft.Media.JobCanceled' + MediaJobCanceledEventName = "Microsoft.Media.JobCanceled" - MediaJobCancelingEventName = 'Microsoft.Media.JobCanceling' + MediaJobCancelingEventName = "Microsoft.Media.JobCanceling" # spell-checker:ignore Errored - MediaJobErroredEventName = 'Microsoft.Media.JobErrored' + MediaJobErroredEventName = "Microsoft.Media.JobErrored" - MediaJobFinishedEventName = 'Microsoft.Media.JobFinished' + MediaJobFinishedEventName = "Microsoft.Media.JobFinished" - MediaJobOutputCanceledEventName = 'Microsoft.Media.JobOutputCanceled' + MediaJobOutputCanceledEventName = "Microsoft.Media.JobOutputCanceled" - MediaJobOutputCancelingEventName = 'Microsoft.Media.JobOutputCanceling' + MediaJobOutputCancelingEventName = "Microsoft.Media.JobOutputCanceling" - MediaJobOutputErroredEventName = 'Microsoft.Media.JobOutputErrored' + MediaJobOutputErroredEventName = "Microsoft.Media.JobOutputErrored" - MediaJobOutputFinishedEventName = 'Microsoft.Media.JobOutputFinished' + MediaJobOutputFinishedEventName = "Microsoft.Media.JobOutputFinished" - MediaJobOutputProcessingEventName = 'Microsoft.Media.JobOutputProcessing' + MediaJobOutputProcessingEventName = "Microsoft.Media.JobOutputProcessing" - MediaJobOutputProgressEventName = 'Microsoft.Media.JobOutputProgress' + MediaJobOutputProgressEventName = "Microsoft.Media.JobOutputProgress" - MediaJobOutputScheduledEventName = 'Microsoft.Media.JobOutputScheduled' + MediaJobOutputScheduledEventName = "Microsoft.Media.JobOutputScheduled" - MediaJobOutputStateChangeEventName = 'Microsoft.Media.JobOutputStateChange' + MediaJobOutputStateChangeEventName = "Microsoft.Media.JobOutputStateChange" - MediaJobProcessingEventName = 'Microsoft.Media.JobProcessing' + MediaJobProcessingEventName = "Microsoft.Media.JobProcessing" - MediaJobScheduledEventName = 'Microsoft.Media.JobScheduled' + MediaJobScheduledEventName = "Microsoft.Media.JobScheduled" - MediaJobStateChangeEventName = 'Microsoft.Media.JobStateChange' + MediaJobStateChangeEventName = "Microsoft.Media.JobStateChange" - MediaLiveEventChannelArchiveHeartbeatEventName = 'Microsoft.Media.LiveEventChannelArchiveHeartbeat' + MediaLiveEventChannelArchiveHeartbeatEventName = "Microsoft.Media.LiveEventChannelArchiveHeartbeat" - MediaLiveEventConnectionRejectedEventName = 'Microsoft.Media.LiveEventConnectionRejected' + MediaLiveEventConnectionRejectedEventName = "Microsoft.Media.LiveEventConnectionRejected" - MediaLiveEventEncoderConnectedEventName = 'Microsoft.Media.LiveEventEncoderConnected' + MediaLiveEventEncoderConnectedEventName = "Microsoft.Media.LiveEventEncoderConnected" - MediaLiveEventEncoderDisconnectedEventName = 'Microsoft.Media.LiveEventEncoderDisconnected' + MediaLiveEventEncoderDisconnectedEventName = "Microsoft.Media.LiveEventEncoderDisconnected" - MediaLiveEventIncomingDataChunkDroppedEventName = 'Microsoft.Media.LiveEventIncomingDataChunkDropped' + MediaLiveEventIncomingDataChunkDroppedEventName = "Microsoft.Media.LiveEventIncomingDataChunkDropped" - MediaLiveEventIncomingStreamReceivedEventName = 'Microsoft.Media.LiveEventIncomingStreamReceived' + MediaLiveEventIncomingStreamReceivedEventName = "Microsoft.Media.LiveEventIncomingStreamReceived" - MediaLiveEventIncomingStreamsOutOfSyncEventName = 'Microsoft.Media.LiveEventIncomingStreamsOutOfSync' + MediaLiveEventIncomingStreamsOutOfSyncEventName = "Microsoft.Media.LiveEventIncomingStreamsOutOfSync" - MediaLiveEventIncomingVideoStreamsOutOfSyncEventName = 'Microsoft.Media.LiveEventIncomingVideoStreamsOutOfSync' + MediaLiveEventIncomingVideoStreamsOutOfSyncEventName = "Microsoft.Media.LiveEventIncomingVideoStreamsOutOfSync" - MediaLiveEventIngestHeartbeatEventName = 'Microsoft.Media.LiveEventIngestHeartbeat' + MediaLiveEventIngestHeartbeatEventName = "Microsoft.Media.LiveEventIngestHeartbeat" - MediaLiveEventTrackDiscontinuityDetectedEventName = 'Microsoft.Media.LiveEventTrackDiscontinuityDetected' + MediaLiveEventTrackDiscontinuityDetectedEventName = "Microsoft.Media.LiveEventTrackDiscontinuityDetected" - PolicyInsightsPolicyStateChangedEventName = 'Microsoft.PolicyInsights.PolicyStateChanged' + PolicyInsightsPolicyStateChangedEventName = "Microsoft.PolicyInsights.PolicyStateChanged" - PolicyInsightsPolicyStateCreatedEventName = 'Microsoft.PolicyInsights.PolicyStateCreated' + PolicyInsightsPolicyStateCreatedEventName = "Microsoft.PolicyInsights.PolicyStateCreated" - PolicyInsightsPolicyStateDeletedEventName = 'Microsoft.PolicyInsights.PolicyStateDeleted' + PolicyInsightsPolicyStateDeletedEventName = "Microsoft.PolicyInsights.PolicyStateDeleted" - RedisExportRDBCompletedEventName = 'Microsoft.Cache.ExportRDBCompleted' + RedisExportRDBCompletedEventName = "Microsoft.Cache.ExportRDBCompleted" - RedisImportRDBCompletedEventName = 'Microsoft.Cache.ImportRDBCompleted' + RedisImportRDBCompletedEventName = "Microsoft.Cache.ImportRDBCompleted" - RedisPatchingCompletedEventName = 'Microsoft.Cache.PatchingCompleted' + RedisPatchingCompletedEventName = "Microsoft.Cache.PatchingCompleted" - RedisScalingCompletedEventName = 'Microsoft.Cache.ScalingCompleted' + RedisScalingCompletedEventName = "Microsoft.Cache.ScalingCompleted" - ResourceActionCancelName = 'Microsoft.Resources.ResourceActionCancel' + ResourceActionCancelName = "Microsoft.Resources.ResourceActionCancel" - ResourceActionFailureName = 'Microsoft.Resources.ResourceActionFailure' + ResourceActionFailureName = "Microsoft.Resources.ResourceActionFailure" - ResourceActionSuccessName = 'Microsoft.Resources.ResourceActionSuccess' + ResourceActionSuccessName = "Microsoft.Resources.ResourceActionSuccess" - ResourceDeleteCancelName = 'Microsoft.Resources.ResourceDeleteCancel' + ResourceDeleteCancelName = "Microsoft.Resources.ResourceDeleteCancel" - ResourceDeleteFailureName = 'Microsoft.Resources.ResourceDeleteFailure' + ResourceDeleteFailureName = "Microsoft.Resources.ResourceDeleteFailure" - ResourceDeleteSuccessName = 'Microsoft.Resources.ResourceDeleteSuccess' + ResourceDeleteSuccessName = "Microsoft.Resources.ResourceDeleteSuccess" - ResourceWriteCancelName = 'Microsoft.Resources.ResourceWriteCancel' + ResourceWriteCancelName = "Microsoft.Resources.ResourceWriteCancel" - ResourceWriteFailureName = 'Microsoft.Resources.ResourceWriteFailure' + ResourceWriteFailureName = "Microsoft.Resources.ResourceWriteFailure" - ResourceWriteSuccessName = 'Microsoft.Resources.ResourceWriteSuccess' + ResourceWriteSuccessName = "Microsoft.Resources.ResourceWriteSuccess" - ServiceBusActiveMessagesAvailablePeriodicNotificationsEventName = 'Microsoft.ServiceBus.ActiveMessagesAvailablePeriodicNotifications' + ServiceBusActiveMessagesAvailablePeriodicNotificationsEventName = ( + "Microsoft.ServiceBus.ActiveMessagesAvailablePeriodicNotifications" + ) - ServiceBusActiveMessagesAvailableWithNoListenersEventName = 'Microsoft.ServiceBus.ActiveMessagesAvailableWithNoListeners' + ServiceBusActiveMessagesAvailableWithNoListenersEventName = ( + "Microsoft.ServiceBus.ActiveMessagesAvailableWithNoListeners" + ) - ServiceBusDeadletterMessagesAvailablePeriodicNotificationsEventName = 'Microsoft.ServiceBus.DeadletterMessagesAvailablePeriodicNotifications' + ServiceBusDeadletterMessagesAvailablePeriodicNotificationsEventName = ( + "Microsoft.ServiceBus.DeadletterMessagesAvailablePeriodicNotifications" + ) - ServiceBusDeadletterMessagesAvailableWithNoListenersEventName = 'Microsoft.ServiceBus.DeadletterMessagesAvailableWithNoListeners' + ServiceBusDeadletterMessagesAvailableWithNoListenersEventName = ( + "Microsoft.ServiceBus.DeadletterMessagesAvailableWithNoListeners" + ) - SignalRServiceClientConnectionConnectedEventName = 'Microsoft.SignalRService.ClientConnectionConnected' + SignalRServiceClientConnectionConnectedEventName = "Microsoft.SignalRService.ClientConnectionConnected" - SignalRServiceClientConnectionDisconnectedEventName = 'Microsoft.SignalRService.ClientConnectionDisconnected' + SignalRServiceClientConnectionDisconnectedEventName = "Microsoft.SignalRService.ClientConnectionDisconnected" - StorageAsyncOperationInitiatedEventName = 'Microsoft.Storage.AsyncOperationInitiated' + StorageAsyncOperationInitiatedEventName = "Microsoft.Storage.AsyncOperationInitiated" - StorageBlobCreatedEventName = 'Microsoft.Storage.BlobCreated' + StorageBlobCreatedEventName = "Microsoft.Storage.BlobCreated" - StorageBlobDeletedEventName = 'Microsoft.Storage.BlobDeleted' + StorageBlobDeletedEventName = "Microsoft.Storage.BlobDeleted" - StorageBlobInventoryPolicyCompletedEventName = 'Microsoft.Storage.BlobInventoryPolicyCompleted' + StorageBlobInventoryPolicyCompletedEventName = "Microsoft.Storage.BlobInventoryPolicyCompleted" - StorageBlobRenamedEventName = 'Microsoft.Storage.BlobRenamed' + StorageBlobRenamedEventName = "Microsoft.Storage.BlobRenamed" - StorageBlobTierChangedEventName = 'Microsoft.Storage.BlobTierChanged' + StorageBlobTierChangedEventName = "Microsoft.Storage.BlobTierChanged" - StorageDirectoryCreatedEventName = 'Microsoft.Storage.DirectoryCreated' + StorageDirectoryCreatedEventName = "Microsoft.Storage.DirectoryCreated" - StorageDirectoryDeletedEventName = 'Microsoft.Storage.DirectoryDeleted' + StorageDirectoryDeletedEventName = "Microsoft.Storage.DirectoryDeleted" - StorageDirectoryRenamedEventName = 'Microsoft.Storage.DirectoryRenamed' + StorageDirectoryRenamedEventName = "Microsoft.Storage.DirectoryRenamed" - StorageLifecyclePolicyCompletedEventName = 'Microsoft.Storage.LifecyclePolicyCompleted' + StorageLifecyclePolicyCompletedEventName = "Microsoft.Storage.LifecyclePolicyCompleted" - StorageTaskCompletedEventName = 'Microsoft.Storage.StorageTaskCompleted' + StorageTaskCompletedEventName = "Microsoft.Storage.StorageTaskCompleted" - StorageTaskQueuedEventName = 'Microsoft.Storage.StorageTaskQueued' + StorageTaskQueuedEventName = "Microsoft.Storage.StorageTaskQueued" - SubscriptionDeletedEventName = 'Microsoft.EventGrid.SubscriptionDeletedEvent' + SubscriptionDeletedEventName = "Microsoft.EventGrid.SubscriptionDeletedEvent" - SubscriptionValidationEventName = 'Microsoft.EventGrid.SubscriptionValidationEvent' + SubscriptionValidationEventName = "Microsoft.EventGrid.SubscriptionValidationEvent" - WebAppServicePlanUpdatedEventName = 'Microsoft.Web.AppServicePlanUpdated' + WebAppServicePlanUpdatedEventName = "Microsoft.Web.AppServicePlanUpdated" - WebAppUpdatedEventName = 'Microsoft.Web.AppUpdated' + WebAppUpdatedEventName = "Microsoft.Web.AppUpdated" - WebBackupOperationCompletedEventName = 'Microsoft.Web.BackupOperationCompleted' + WebBackupOperationCompletedEventName = "Microsoft.Web.BackupOperationCompleted" - WebBackupOperationFailedEventName = 'Microsoft.Web.BackupOperationFailed' + WebBackupOperationFailedEventName = "Microsoft.Web.BackupOperationFailed" - WebBackupOperationStartedEventName = 'Microsoft.Web.BackupOperationStarted' + WebBackupOperationStartedEventName = "Microsoft.Web.BackupOperationStarted" - WebRestoreOperationCompletedEventName = 'Microsoft.Web.RestoreOperationCompleted' + WebRestoreOperationCompletedEventName = "Microsoft.Web.RestoreOperationCompleted" - WebRestoreOperationFailedEventName = 'Microsoft.Web.RestoreOperationFailed' + WebRestoreOperationFailedEventName = "Microsoft.Web.RestoreOperationFailed" - WebRestoreOperationStartedEventName = 'Microsoft.Web.RestoreOperationStarted' + WebRestoreOperationStartedEventName = "Microsoft.Web.RestoreOperationStarted" - WebSlotSwapCompletedEventName = 'Microsoft.Web.SlotSwapCompleted' + WebSlotSwapCompletedEventName = "Microsoft.Web.SlotSwapCompleted" - WebSlotSwapFailedEventName = 'Microsoft.Web.SlotSwapFailed' + WebSlotSwapFailedEventName = "Microsoft.Web.SlotSwapFailed" - WebSlotSwapStartedEventName = 'Microsoft.Web.SlotSwapStarted' + WebSlotSwapStartedEventName = "Microsoft.Web.SlotSwapStarted" - WebSlotSwapWithPreviewCancelledEventName = 'Microsoft.Web.SlotSwapWithPreviewCancelled' + WebSlotSwapWithPreviewCancelledEventName = "Microsoft.Web.SlotSwapWithPreviewCancelled" - WebSlotSwapWithPreviewStartedEventName = 'Microsoft.Web.SlotSwapWithPreviewStarted' + WebSlotSwapWithPreviewStartedEventName = "Microsoft.Web.SlotSwapWithPreviewStarted" - ContainerRegistryArtifactEventName = 'Microsoft.AppConfiguration.KeyValueModified' + ContainerRegistryArtifactEventName = "Microsoft.AppConfiguration.KeyValueModified" - KeyVaultAccessPolicyChangedEventName = 'Microsoft.KeyVault.VaultAccessPolicyChanged' + KeyVaultAccessPolicyChangedEventName = "Microsoft.KeyVault.VaultAccessPolicyChanged" - ContainerRegistryEventName = 'Microsoft.ContainerRegistry.ChartPushed' + ContainerRegistryEventName = "Microsoft.ContainerRegistry.ChartPushed" - ServiceBusDeadletterMessagesAvailableWithNoListenerEventName = 'Microsoft.ServiceBus.DeadletterMessagesAvailableWithNoListeners' + ServiceBusDeadletterMessagesAvailableWithNoListenerEventName = ( + "Microsoft.ServiceBus.DeadletterMessagesAvailableWithNoListeners" + ) diff --git a/sdk/eventgrid/azure-eventgrid/azure/eventgrid/_helpers.py b/sdk/eventgrid/azure-eventgrid/azure/eventgrid/_helpers.py index c5743dda1106..f54c77cd9735 100644 --- a/sdk/eventgrid/azure-eventgrid/azure/eventgrid/_helpers.py +++ b/sdk/eventgrid/azure-eventgrid/azure/eventgrid/_helpers.py @@ -28,6 +28,7 @@ if TYPE_CHECKING: from datetime import datetime + def generate_sas(endpoint, shared_access_key, expiration_date_utc, **kwargs): # type: (str, str, datetime, Any) -> str """Helper method to generate shared access signature given hostname, key, and expiration date. @@ -48,19 +49,16 @@ def generate_sas(endpoint, shared_access_key, expiration_date_utc, **kwargs): :dedent: 0 :caption: Generate a shared access signature. """ - full_endpoint = "{}?apiVersion={}".format( - endpoint, kwargs.get("api_version", constants.DEFAULT_API_VERSION) - ) + full_endpoint = "{}?apiVersion={}".format(endpoint, kwargs.get("api_version", constants.DEFAULT_API_VERSION)) encoded_resource = quote(full_endpoint, safe=constants.SAFE_ENCODE) encoded_expiration_utc = quote(str(expiration_date_utc), safe=constants.SAFE_ENCODE) unsigned_sas = "r={}&e={}".format(encoded_resource, encoded_expiration_utc) - signature = quote( - _generate_hmac(shared_access_key, unsigned_sas), safe=constants.SAFE_ENCODE - ) + signature = quote(_generate_hmac(shared_access_key, unsigned_sas), safe=constants.SAFE_ENCODE) signed_sas = "{}&s={}".format(unsigned_sas, signature) return signed_sas + def _generate_hmac(key, message): decoded_key = base64.b64decode(key) bytes_message = message.encode("ascii") @@ -73,18 +71,11 @@ def _get_authentication_policy(credential, bearer_token_policy=BearerTokenCreden if credential is None: raise ValueError("Parameter 'self._credential' must not be None.") if hasattr(credential, "get_token"): - return bearer_token_policy( - credential, - constants.DEFAULT_EVENTGRID_SCOPE - ) + return bearer_token_policy(credential, constants.DEFAULT_EVENTGRID_SCOPE) if isinstance(credential, AzureKeyCredential): - return AzureKeyCredentialPolicy( - credential=credential, name=constants.EVENTGRID_KEY_HEADER - ) + return AzureKeyCredentialPolicy(credential=credential, name=constants.EVENTGRID_KEY_HEADER) if isinstance(credential, AzureSasCredential): - return EventGridSasCredentialPolicy( - credential=credential, name=constants.EVENTGRID_TOKEN_HEADER - ) + return EventGridSasCredentialPolicy(credential=credential, name=constants.EVENTGRID_TOKEN_HEADER) raise ValueError( "The provided credential should be an instance of a TokenCredential, AzureSasCredential or AzureKeyCredential" ) @@ -98,6 +89,7 @@ def _is_cloud_event(event): except TypeError: return False + def _is_eventgrid_event(event): # type: (Any) -> bool required = ("subject", "eventType", "data", "dataVersion", "id", "eventTime") @@ -119,6 +111,7 @@ def _eventgrid_data_typecheck(event): "https://docs.microsoft.com/en-us/azure/event-grid/event-schema" ) + def _cloud_event_to_generated(cloud_event, **kwargs): if isinstance(cloud_event.data, bytes): data_base64 = cloud_event.data @@ -141,17 +134,19 @@ def _cloud_event_to_generated(cloud_event, **kwargs): **kwargs ) + def _from_cncf_events(event): """This takes in a CNCF cloudevent and returns a dictionary. If cloud events library is not installed, the event is returned back. """ try: from cloudevents.http import to_json + return json.loads(to_json(event)) except (AttributeError, ImportError): # means this is not a CNCF event return event - except Exception as err: # pylint: disable=broad-except + except Exception as err: # pylint: disable=broad-except msg = """Failed to serialize the event. Please ensure your CloudEvents is correctly formatted (https://pypi.org/project/cloudevents/)""" raise_with_traceback(ValueError, msg, err) @@ -160,26 +155,21 @@ def _from_cncf_events(event): def _build_request(endpoint, content_type, events, *, channel_name=None): serialize = Serializer() header_parameters = {} # type: Dict[str, Any] - header_parameters['Content-Type'] = serialize.header("content_type", content_type, 'str') + header_parameters["Content-Type"] = serialize.header("content_type", content_type, "str") if channel_name: - header_parameters['aeg-channel-name'] = channel_name + header_parameters["aeg-channel-name"] = channel_name query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = serialize.query("api_version", "2018-01-01", 'str') + query_parameters["api-version"] = serialize.query("api_version", "2018-01-01", "str") - body = serialize.body(events, '[object]') + body = serialize.body(events, "[object]") if body is None: data = None else: data = json.dumps(body) - header_parameters['Content-Length'] = str(len(data)) + header_parameters["Content-Length"] = str(len(data)) - request = HttpRequest( - method="POST", - url=endpoint, - headers=header_parameters, - data=data - ) + request = HttpRequest(method="POST", url=endpoint, headers=header_parameters, data=data) request.format_parameters(query_parameters) return request diff --git a/sdk/eventgrid/azure-eventgrid/azure/eventgrid/_messaging_shared.py b/sdk/eventgrid/azure-eventgrid/azure/eventgrid/_messaging_shared.py index bc9307d00580..802cd0cdf529 100644 --- a/sdk/eventgrid/azure-eventgrid/azure/eventgrid/_messaging_shared.py +++ b/sdk/eventgrid/azure-eventgrid/azure/eventgrid/_messaging_shared.py @@ -11,10 +11,10 @@ # ========================================================================== - import json from azure.core.exceptions import raise_with_traceback + def _get_json_content(obj): """Event mixin to have methods that are common to different Event types like CloudEvent, EventGridEvent etc. @@ -34,7 +34,7 @@ def _get_json_content(obj): return json.loads(next(obj.body)) except ValueError as err: raise_with_traceback(ValueError, msg, err) - except: # pylint: disable=bare-except + except: # pylint: disable=bare-except try: return json.loads(obj) except ValueError as err: diff --git a/sdk/eventgrid/azure-eventgrid/azure/eventgrid/_model_base.py b/sdk/eventgrid/azure-eventgrid/azure/eventgrid/_model_base.py new file mode 100644 index 000000000000..3a2f72e93fda --- /dev/null +++ b/sdk/eventgrid/azure-eventgrid/azure/eventgrid/_model_base.py @@ -0,0 +1,706 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# -------------------------------------------------------------------------- +# pylint: disable=protected-access, arguments-differ, signature-differs, broad-except +# pyright: reportGeneralTypeIssues=false + +import functools +import sys +import logging +import base64 +import re +import copy +import typing +from datetime import datetime, date, time, timedelta, timezone +from json import JSONEncoder +import isodate +from azure.core.exceptions import DeserializationError +from azure.core import CaseInsensitiveEnumMeta +from azure.core.pipeline import PipelineResponse +from azure.core.serialization import _Null # pylint: disable=protected-access + +if sys.version_info >= (3, 9): + from collections.abc import MutableMapping +else: + from typing import MutableMapping + +_LOGGER = logging.getLogger(__name__) + +__all__ = ["AzureJSONEncoder", "Model", "rest_field", "rest_discriminator"] + + +TZ_UTC = timezone.utc + + +def _timedelta_as_isostr(td: timedelta) -> str: + """Converts a datetime.timedelta object into an ISO 8601 formatted string, e.g. 'P4DT12H30M05S' + + Function adapted from the Tin Can Python project: https://github.com/RusticiSoftware/TinCanPython + + :param timedelta td: The timedelta to convert + :rtype: str + :return: ISO8601 version of this timedelta + """ + + # Split seconds to larger units + seconds = td.total_seconds() + minutes, seconds = divmod(seconds, 60) + hours, minutes = divmod(minutes, 60) + days, hours = divmod(hours, 24) + + days, hours, minutes = list(map(int, (days, hours, minutes))) + seconds = round(seconds, 6) + + # Build date + date_str = "" + if days: + date_str = "%sD" % days + + if hours or minutes or seconds: + # Build time + time_str = "T" + + # Hours + bigger_exists = date_str or hours + if bigger_exists: + time_str += "{:02}H".format(hours) + + # Minutes + bigger_exists = bigger_exists or minutes + if bigger_exists: + time_str += "{:02}M".format(minutes) + + # Seconds + try: + if seconds.is_integer(): + seconds_string = "{:02}".format(int(seconds)) + else: + # 9 chars long w/ leading 0, 6 digits after decimal + seconds_string = "%09.6f" % seconds + # Remove trailing zeros + seconds_string = seconds_string.rstrip("0") + except AttributeError: # int.is_integer() raises + seconds_string = "{:02}".format(seconds) + + time_str += "{}S".format(seconds_string) + else: + time_str = "" + + return "P" + date_str + time_str + + +def _datetime_as_isostr(dt: typing.Union[datetime, date, time, timedelta]) -> str: + """Converts a datetime.(datetime|date|time|timedelta) object into an ISO 8601 formatted string + + :param timedelta dt: The date object to convert + :rtype: str + :return: ISO8601 version of this datetime + """ + # First try datetime.datetime + if hasattr(dt, "year") and hasattr(dt, "hour"): + dt = typing.cast(datetime, dt) + # astimezone() fails for naive times in Python 2.7, so make make sure dt is aware (tzinfo is set) + if not dt.tzinfo: + iso_formatted = dt.replace(tzinfo=TZ_UTC).isoformat() + else: + iso_formatted = dt.astimezone(TZ_UTC).isoformat() + # Replace the trailing "+00:00" UTC offset with "Z" (RFC 3339: https://www.ietf.org/rfc/rfc3339.txt) + return iso_formatted.replace("+00:00", "Z") + # Next try datetime.date or datetime.time + try: + dt = typing.cast(typing.Union[date, time], dt) + return dt.isoformat() + # Last, try datetime.timedelta + except AttributeError: + dt = typing.cast(timedelta, dt) + return _timedelta_as_isostr(dt) + + +def _serialize_bytes(o) -> str: + return base64.b64encode(o).decode() + + +def _serialize_datetime(o): + if hasattr(o, "year") and hasattr(o, "hour"): + # astimezone() fails for naive times in Python 2.7, so make make sure o is aware (tzinfo is set) + if not o.tzinfo: + iso_formatted = o.replace(tzinfo=TZ_UTC).isoformat() + else: + iso_formatted = o.astimezone(TZ_UTC).isoformat() + # Replace the trailing "+00:00" UTC offset with "Z" (RFC 3339: https://www.ietf.org/rfc/rfc3339.txt) + return iso_formatted.replace("+00:00", "Z") + # Next try datetime.date or datetime.time + return o.isoformat() + + +def _is_readonly(p): + try: + return p._readonly # pylint: disable=protected-access + except AttributeError: + return False + + +class AzureJSONEncoder(JSONEncoder): + """A JSON encoder that's capable of serializing datetime objects and bytes.""" + + def default(self, o): # pylint: disable=too-many-return-statements + if _is_model(o): + readonly_props = [ + p._rest_name for p in o._attr_to_rest_field.values() if _is_readonly(p) + ] # pylint: disable=protected-access + return {k: v for k, v in o.items() if k not in readonly_props} + if isinstance(o, (bytes, bytearray)): + return base64.b64encode(o).decode() + if isinstance(o, _Null): + return None + try: + return super(AzureJSONEncoder, self).default(o) + except TypeError: + if isinstance(o, (bytes, bytearray)): + return _serialize_bytes(o) + try: + # First try datetime.datetime + return _serialize_datetime(o) + except AttributeError: + pass + # Last, try datetime.timedelta + try: + return _timedelta_as_isostr(o) + except AttributeError: + # This will be raised when it hits value.total_seconds in the method above + pass + return super(AzureJSONEncoder, self).default(o) + + +_VALID_DATE = re.compile(r"\d{4}[-]\d{2}[-]\d{2}T\d{2}:\d{2}:\d{2}" + r"\.?\d*Z?[-+]?[\d{2}]?:?[\d{2}]?") + + +def _deserialize_datetime(attr: typing.Union[str, datetime]) -> datetime: + """Deserialize ISO-8601 formatted string into Datetime object. + + :param str attr: response string to be deserialized. + :rtype: ~datetime.datetime + :returns: The datetime object from that input + """ + if isinstance(attr, datetime): + # i'm already deserialized + return attr + attr = attr.upper() + match = _VALID_DATE.match(attr) + if not match: + raise ValueError("Invalid datetime string: " + attr) + + check_decimal = attr.split(".") + if len(check_decimal) > 1: + decimal_str = "" + for digit in check_decimal[1]: + if digit.isdigit(): + decimal_str += digit + else: + break + if len(decimal_str) > 6: + attr = attr.replace(decimal_str, decimal_str[0:6]) + + date_obj = isodate.parse_datetime(attr) + test_utc = date_obj.utctimetuple() + if test_utc.tm_year > 9999 or test_utc.tm_year < 1: + raise OverflowError("Hit max or min date") + return date_obj + + +def _deserialize_date(attr: typing.Union[str, date]) -> date: + """Deserialize ISO-8601 formatted string into Date object. + :param str attr: response string to be deserialized. + :rtype: date + :returns: The date object from that input + """ + # This must NOT use defaultmonth/defaultday. Using None ensure this raises an exception. + if isinstance(attr, date): + return attr + return isodate.parse_date(attr, defaultmonth=None, defaultday=None) + + +def _deserialize_time(attr: typing.Union[str, time]) -> time: + """Deserialize ISO-8601 formatted string into time object. + + :param str attr: response string to be deserialized. + :rtype: datetime.time + :returns: The time object from that input + """ + if isinstance(attr, time): + return attr + return isodate.parse_time(attr) + + +def deserialize_bytes(attr): + if isinstance(attr, (bytes, bytearray)): + return attr + return bytes(base64.b64decode(attr)) + + +def deserialize_duration(attr): + if isinstance(attr, timedelta): + return attr + return isodate.parse_duration(attr) + + +_DESERIALIZE_MAPPING = { + datetime: _deserialize_datetime, + date: _deserialize_date, + time: _deserialize_time, + bytes: deserialize_bytes, + timedelta: deserialize_duration, + typing.Any: lambda x: x, +} + + +def _get_model(module_name: str, model_name: str): + models = {k: v for k, v in sys.modules[module_name].__dict__.items() if isinstance(v, type)} + module_end = module_name.rsplit(".", 1)[0] + module = sys.modules[module_end] + models.update({k: v for k, v in module.__dict__.items() if isinstance(v, type)}) + if isinstance(model_name, str): + model_name = model_name.split(".")[-1] + if model_name not in models: + return model_name + return models[model_name] + + +_UNSET = object() + + +class _MyMutableMapping(MutableMapping[str, typing.Any]): # pylint: disable=unsubscriptable-object + def __init__(self, data: typing.Dict[str, typing.Any]) -> None: + self._data = copy.deepcopy(data) + + def __contains__(self, key: typing.Any) -> bool: + return key in self._data + + def __getitem__(self, key: str) -> typing.Any: + return self._data.__getitem__(key) + + def __setitem__(self, key: str, value: typing.Any) -> None: + self._data.__setitem__(key, value) + + def __delitem__(self, key: str) -> None: + self._data.__delitem__(key) + + def __iter__(self) -> typing.Iterator[typing.Any]: + return self._data.__iter__() + + def __len__(self) -> int: + return self._data.__len__() + + def __ne__(self, other: typing.Any) -> bool: + return not self.__eq__(other) + + def keys(self) -> typing.KeysView[str]: + return self._data.keys() + + def values(self) -> typing.ValuesView[typing.Any]: + return self._data.values() + + def items(self) -> typing.ItemsView[str, typing.Any]: + return self._data.items() + + def get(self, key: str, default: typing.Any = None) -> typing.Any: + try: + return self[key] + except KeyError: + return default + + @typing.overload # type: ignore + def pop(self, key: str) -> typing.Any: # pylint: disable=no-member + ... + + @typing.overload + def pop(self, key: str, default: typing.Any) -> typing.Any: + ... + + def pop(self, key: str, default: typing.Any = _UNSET) -> typing.Any: + if default is _UNSET: + return self._data.pop(key) + return self._data.pop(key, default) + + def popitem(self) -> typing.Tuple[str, typing.Any]: + return self._data.popitem() + + def clear(self) -> None: + self._data.clear() + + def update(self, *args: typing.Any, **kwargs: typing.Any) -> None: + self._data.update(*args, **kwargs) + + @typing.overload # type: ignore + def setdefault(self, key: str) -> typing.Any: + ... + + @typing.overload + def setdefault(self, key: str, default: typing.Any) -> typing.Any: + ... + + def setdefault(self, key: str, default: typing.Any = _UNSET) -> typing.Any: + if default is _UNSET: + return self._data.setdefault(key) + return self._data.setdefault(key, default) + + def __eq__(self, other: typing.Any) -> bool: + try: + other_model = self.__class__(other) + except Exception: + return False + return self._data == other_model._data + + def __repr__(self) -> str: + return str(self._data) + + +def _is_model(obj: typing.Any) -> bool: + return getattr(obj, "_is_model", False) + + +def _serialize(o): + if isinstance(o, (bytes, bytearray)): + return _serialize_bytes(o) + try: + # First try datetime.datetime + return _serialize_datetime(o) + except AttributeError: + pass + # Last, try datetime.timedelta + try: + return _timedelta_as_isostr(o) + except AttributeError: + # This will be raised when it hits value.total_seconds in the method above + pass + return o + + +def _get_rest_field( + attr_to_rest_field: typing.Dict[str, "_RestField"], rest_name: str +) -> typing.Optional["_RestField"]: + try: + return next(rf for rf in attr_to_rest_field.values() if rf._rest_name == rest_name) + except StopIteration: + return None + + +def _create_value(rf: typing.Optional["_RestField"], value: typing.Any) -> typing.Any: + return _deserialize(rf._type, value) if (rf and rf._is_model) else _serialize(value) + + +class Model(_MyMutableMapping): + _is_model = True + + def __init__(self, *args: typing.Any, **kwargs: typing.Any) -> None: + class_name = self.__class__.__name__ + if len(args) > 1: + raise TypeError(f"{class_name}.__init__() takes 2 positional arguments but {len(args) + 1} were given") + dict_to_pass = { + rest_field._rest_name: rest_field._default + for rest_field in self._attr_to_rest_field.values() + if rest_field._default is not _UNSET + } + if args: + dict_to_pass.update( + {k: _create_value(_get_rest_field(self._attr_to_rest_field, k), v) for k, v in args[0].items()} + ) + else: + non_attr_kwargs = [k for k in kwargs if k not in self._attr_to_rest_field] + if non_attr_kwargs: + # actual type errors only throw the first wrong keyword arg they see, so following that. + raise TypeError(f"{class_name}.__init__() got an unexpected keyword argument '{non_attr_kwargs[0]}'") + dict_to_pass.update( + {self._attr_to_rest_field[k]._rest_name: _serialize(v) for k, v in kwargs.items() if v is not None} + ) + super().__init__(dict_to_pass) + + def copy(self) -> "Model": + return Model(self.__dict__) + + def __new__(cls, *args: typing.Any, **kwargs: typing.Any) -> "Model": # pylint: disable=unused-argument + # we know the last three classes in mro are going to be 'Model', 'dict', and 'object' + mros = cls.__mro__[:-3][::-1] # ignore model, dict, and object parents, and reverse the mro order + attr_to_rest_field: typing.Dict[str, _RestField] = { # map attribute name to rest_field property + k: v for mro_class in mros for k, v in mro_class.__dict__.items() if k[0] != "_" and hasattr(v, "_type") + } + annotations = { + k: v + for mro_class in mros + if hasattr(mro_class, "__annotations__") # pylint: disable=no-member + for k, v in mro_class.__annotations__.items() # pylint: disable=no-member + } + for attr, rf in attr_to_rest_field.items(): + rf._module = cls.__module__ + if not rf._type: + rf._type = rf._get_deserialize_callable_from_annotation(annotations.get(attr, None)) + if not rf._rest_name_input: + rf._rest_name_input = attr + cls._attr_to_rest_field: typing.Dict[str, _RestField] = dict(attr_to_rest_field.items()) + + return super().__new__(cls) # pylint: disable=no-value-for-parameter + + def __init_subclass__(cls, discriminator: typing.Optional[str] = None) -> None: + for base in cls.__bases__: + if hasattr(base, "__mapping__"): # pylint: disable=no-member + base.__mapping__[discriminator or cls.__name__] = cls # type: ignore # pylint: disable=no-member + + @classmethod + def _get_discriminator(cls) -> typing.Optional[str]: + for v in cls.__dict__.values(): + if isinstance(v, _RestField) and v._is_discriminator: # pylint: disable=protected-access + return v._rest_name # pylint: disable=protected-access + return None + + @classmethod + def _deserialize(cls, data): + if not hasattr(cls, "__mapping__"): # pylint: disable=no-member + return cls(data) + discriminator = cls._get_discriminator() + mapped_cls = cls.__mapping__.get(data.get(discriminator), cls) # pylint: disable=no-member + if mapped_cls == cls: + return cls(data) + return mapped_cls._deserialize(data) # pylint: disable=protected-access + + +def _get_deserialize_callable_from_annotation( # pylint: disable=too-many-return-statements, too-many-statements + annotation: typing.Any, module: typing.Optional[str], rf: typing.Optional["_RestField"] = None +) -> typing.Optional[typing.Callable[[typing.Any], typing.Any]]: + if not annotation or annotation in [int, float]: + return None + + try: + if module and _is_model(_get_model(module, annotation)): + if rf: + rf._is_model = True + + def _deserialize_model(model_deserializer: typing.Optional[typing.Callable], obj): + if _is_model(obj): + return obj + return _deserialize(model_deserializer, obj) + + return functools.partial(_deserialize_model, _get_model(module, annotation)) + except Exception: + pass + + # is it a literal? + try: + if sys.version_info >= (3, 8): + from typing import Literal # pylint: disable=no-name-in-module, ungrouped-imports + else: + from typing_extensions import Literal # type: ignore # pylint: disable=ungrouped-imports + + if annotation.__origin__ == Literal: + return None + except AttributeError: + pass + + if getattr(annotation, "__origin__", None) is typing.Union: + + def _deserialize_with_union(union_annotation, obj): + for t in union_annotation.__args__: + try: + return _deserialize(t, obj, module) + except DeserializationError: + pass + raise DeserializationError() + + return functools.partial(_deserialize_with_union, annotation) + + # is it optional? + try: + # right now, assuming we don't have unions, since we're getting rid of the only + # union we used to have in msrest models, which was union of str and enum + if any(a for a in annotation.__args__ if a == type(None)): + + if_obj_deserializer = _get_deserialize_callable_from_annotation( + next(a for a in annotation.__args__ if a != type(None)), module, rf + ) + + def _deserialize_with_optional(if_obj_deserializer: typing.Optional[typing.Callable], obj): + if obj is None: + return obj + return _deserialize_with_callable(if_obj_deserializer, obj) + + return functools.partial(_deserialize_with_optional, if_obj_deserializer) + except AttributeError: + pass + + # is it a forward ref / in quotes? + if isinstance(annotation, (str, typing.ForwardRef)): + try: + model_name = annotation.__forward_arg__ # type: ignore + except AttributeError: + model_name = annotation + if module is not None: + annotation = _get_model(module, model_name) + + try: + if annotation._name == "Dict": + key_deserializer = _get_deserialize_callable_from_annotation(annotation.__args__[0], module, rf) + value_deserializer = _get_deserialize_callable_from_annotation(annotation.__args__[1], module, rf) + + def _deserialize_dict( + key_deserializer: typing.Optional[typing.Callable], + value_deserializer: typing.Optional[typing.Callable], + obj: typing.Dict[typing.Any, typing.Any], + ): + if obj is None: + return obj + return { + _deserialize(key_deserializer, k, module): _deserialize(value_deserializer, v, module) + for k, v in obj.items() + } + + return functools.partial( + _deserialize_dict, + key_deserializer, + value_deserializer, + ) + except (AttributeError, IndexError): + pass + try: + if annotation._name in ["List", "Set", "Tuple", "Sequence"]: + if len(annotation.__args__) > 1: + + def _deserialize_multiple_sequence( + entry_deserializers: typing.List[typing.Optional[typing.Callable]], obj + ): + if obj is None: + return obj + return type(obj)( + _deserialize(deserializer, entry, module) + for entry, deserializer in zip(obj, entry_deserializers) + ) + + entry_deserializers = [ + _get_deserialize_callable_from_annotation(dt, module, rf) for dt in annotation.__args__ + ] + return functools.partial(_deserialize_multiple_sequence, entry_deserializers) + deserializer = _get_deserialize_callable_from_annotation(annotation.__args__[0], module, rf) + + def _deserialize_sequence( + deserializer: typing.Optional[typing.Callable], + obj, + ): + if obj is None: + return obj + return type(obj)(_deserialize(deserializer, entry, module) for entry in obj) + + return functools.partial(_deserialize_sequence, deserializer) + except (TypeError, IndexError, AttributeError, SyntaxError): + pass + + def _deserialize_default( + annotation, + deserializer_from_mapping, + obj, + ): + if obj is None: + return obj + try: + return _deserialize_with_callable(annotation, obj) + except Exception: + pass + return _deserialize_with_callable(deserializer_from_mapping, obj) + + return functools.partial(_deserialize_default, annotation, _DESERIALIZE_MAPPING.get(annotation)) + + +def _deserialize_with_callable( + deserializer: typing.Optional[typing.Callable[[typing.Any], typing.Any]], value: typing.Any +): + try: + if value is None: + return None + if deserializer is None: + return value + if isinstance(deserializer, CaseInsensitiveEnumMeta): + try: + return deserializer(value) + except ValueError: + # for unknown value, return raw value + return value + if isinstance(deserializer, type) and issubclass(deserializer, Model): + return deserializer._deserialize(value) + return typing.cast(typing.Callable[[typing.Any], typing.Any], deserializer)(value) + except Exception as e: + raise DeserializationError() from e + + +def _deserialize(deserializer: typing.Any, value: typing.Any, module: typing.Optional[str] = None) -> typing.Any: + if isinstance(value, PipelineResponse): + value = value.http_response.json() + deserializer = _get_deserialize_callable_from_annotation(deserializer, module) + return _deserialize_with_callable(deserializer, value) + + +class _RestField: + def __init__( + self, + *, + name: typing.Optional[str] = None, + type: typing.Optional[typing.Callable] = None, # pylint: disable=redefined-builtin + is_discriminator: bool = False, + readonly: bool = False, + default: typing.Any = _UNSET, + ): + self._type = type + self._rest_name_input = name + self._module: typing.Optional[str] = None + self._is_discriminator = is_discriminator + self._readonly = readonly + self._is_model = False + self._default = default + + @property + def _rest_name(self) -> str: + if self._rest_name_input is None: + raise ValueError("Rest name was never set") + return self._rest_name_input + + def __get__(self, obj: Model, type=None): # pylint: disable=redefined-builtin + # by this point, type and rest_name will have a value bc we default + # them in __new__ of the Model class + item = obj.get(self._rest_name) + if item is None: + return item + return _deserialize(self._type, _serialize(item)) + + def __set__(self, obj: Model, value) -> None: + if value is None: + # we want to wipe out entries if users set attr to None + try: + obj.__delitem__(self._rest_name) + except KeyError: + pass + return + if self._is_model and not _is_model(value): + obj.__setitem__(self._rest_name, _deserialize(self._type, value)) + obj.__setitem__(self._rest_name, _serialize(value)) + + def _get_deserialize_callable_from_annotation( + self, annotation: typing.Any + ) -> typing.Optional[typing.Callable[[typing.Any], typing.Any]]: + return _get_deserialize_callable_from_annotation(annotation, self._module, self) + + +def rest_field( + *, + name: typing.Optional[str] = None, + type: typing.Optional[typing.Callable] = None, # pylint: disable=redefined-builtin + readonly: bool = False, + default: typing.Any = _UNSET, +) -> typing.Any: + return _RestField(name=name, type=type, readonly=readonly, default=default) + + +def rest_discriminator( + *, + name: typing.Optional[str] = None, + type: typing.Optional[typing.Callable] = None, # pylint: disable=redefined-builtin +) -> typing.Any: + return _RestField(name=name, type=type, is_discriminator=True) diff --git a/sdk/eventgrid/azure-eventgrid/azure/eventgrid/_operations/__init__.py b/sdk/eventgrid/azure-eventgrid/azure/eventgrid/_operations/__init__.py new file mode 100644 index 000000000000..51bf5a7f3dcf --- /dev/null +++ b/sdk/eventgrid/azure-eventgrid/azure/eventgrid/_operations/__init__.py @@ -0,0 +1,19 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for license information. +# Code generated by Microsoft (R) Python Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is regenerated. +# -------------------------------------------------------------------------- + +from ._operations import EventGridClientOperationsMixin + +from ._patch import __all__ as _patch_all +from ._patch import * # pylint: disable=unused-wildcard-import +from ._patch import patch_sdk as _patch_sdk + +__all__ = [ + "EventGridClientOperationsMixin", +] +__all__.extend([p for p in _patch_all if p not in __all__]) +_patch_sdk() diff --git a/sdk/eventgrid/azure-eventgrid/azure/eventgrid/_operations/_operations.py b/sdk/eventgrid/azure-eventgrid/azure/eventgrid/_operations/_operations.py new file mode 100644 index 000000000000..21b51383c7ed --- /dev/null +++ b/sdk/eventgrid/azure-eventgrid/azure/eventgrid/_operations/_operations.py @@ -0,0 +1,640 @@ +# pylint: disable=too-many-lines +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for license information. +# Code generated by Microsoft (R) Python Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is regenerated. +# -------------------------------------------------------------------------- +import datetime +import json +from typing import Any, Callable, Dict, List, Optional, TypeVar + +from azure.core.exceptions import ( + ClientAuthenticationError, + HttpResponseError, + ResourceExistsError, + ResourceNotFoundError, + ResourceNotModifiedError, + map_error, +) +from azure.core.pipeline import PipelineResponse +from azure.core.pipeline.transport import HttpResponse +from azure.core.rest import HttpRequest +from azure.core.tracing.decorator import distributed_trace +from azure.core.utils import case_insensitive_dict + +from .. import models as _models +from .._model_base import AzureJSONEncoder, _deserialize +from .._serialization import Serializer +from .._vendor import EventGridClientMixinABC, _format_url_section + +T = TypeVar("T") +ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]] + +_SERIALIZER = Serializer() +_SERIALIZER.client_side_validation = False + + +def build_event_grid_publish_cloud_event_request(topic_name: str, **kwargs: Any) -> HttpRequest: + _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) + _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) + + content_type: str = kwargs.pop("content_type") + api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2023-06-01-preview")) + # Construct URL + _url = "/topics/{topicName}:publish" + path_format_arguments = { + "topicName": _SERIALIZER.url("topic_name", topic_name, "str"), + } + + _url: str = _format_url_section(_url, **path_format_arguments) # type: ignore + + # Construct parameters + _params["api-version"] = _SERIALIZER.query("api_version", api_version, "str") + + # Construct headers + _headers["content-type"] = _SERIALIZER.header("content_type", content_type, "str") + + return HttpRequest(method="POST", url=_url, params=_params, headers=_headers, **kwargs) + + +def build_event_grid_publish_cloud_events_request(topic_name: str, **kwargs: Any) -> HttpRequest: + _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) + _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) + + content_type: str = kwargs.pop("content_type") + api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2023-06-01-preview")) + # Construct URL + _url = "/topics/{topicName}:publish" + path_format_arguments = { + "topicName": _SERIALIZER.url("topic_name", topic_name, "str"), + } + + _url: str = _format_url_section(_url, **path_format_arguments) # type: ignore + + # Construct parameters + _params["api-version"] = _SERIALIZER.query("api_version", api_version, "str") + + # Construct headers + _headers["content-type"] = _SERIALIZER.header("content_type", content_type, "str") + + return HttpRequest(method="POST", url=_url, params=_params, headers=_headers, **kwargs) + + +def build_event_grid_receive_cloud_events_request( + topic_name: str, + event_subscription_name: str, + *, + max_events: Optional[int] = None, + max_wait_time: Optional[datetime.timedelta] = None, + **kwargs: Any +) -> HttpRequest: + _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) + _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) + + api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2023-06-01-preview")) + accept = _headers.pop("Accept", "application/json") + + # Construct URL + _url = "/topics/{topicName}/eventsubscriptions/{eventSubscriptionName}:receive" + path_format_arguments = { + "topicName": _SERIALIZER.url("topic_name", topic_name, "str"), + "eventSubscriptionName": _SERIALIZER.url("event_subscription_name", event_subscription_name, "str"), + } + + _url: str = _format_url_section(_url, **path_format_arguments) # type: ignore + + # Construct parameters + _params["api-version"] = _SERIALIZER.query("api_version", api_version, "str") + if max_events is not None: + _params["maxEvents"] = _SERIALIZER.query("max_events", max_events, "int") + if max_wait_time is not None: + _params["maxWaitTime"] = _SERIALIZER.query("max_wait_time", max_wait_time, "duration") + + # Construct headers + _headers["Accept"] = _SERIALIZER.header("accept", accept, "str") + + return HttpRequest(method="POST", url=_url, params=_params, headers=_headers, **kwargs) + + +def build_event_grid_acknowledge_cloud_events_request( + topic_name: str, event_subscription_name: str, **kwargs: Any +) -> HttpRequest: + _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) + _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) + + content_type: str = kwargs.pop("content_type") + api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2023-06-01-preview")) + accept = _headers.pop("Accept", "application/json") + + # Construct URL + _url = "/topics/{topicName}/eventsubscriptions/{eventSubscriptionName}:acknowledge" + path_format_arguments = { + "topicName": _SERIALIZER.url("topic_name", topic_name, "str"), + "eventSubscriptionName": _SERIALIZER.url("event_subscription_name", event_subscription_name, "str"), + } + + _url: str = _format_url_section(_url, **path_format_arguments) # type: ignore + + # Construct parameters + _params["api-version"] = _SERIALIZER.query("api_version", api_version, "str") + + # Construct headers + _headers["content-type"] = _SERIALIZER.header("content_type", content_type, "str") + _headers["Accept"] = _SERIALIZER.header("accept", accept, "str") + + return HttpRequest(method="POST", url=_url, params=_params, headers=_headers, **kwargs) + + +def build_event_grid_release_cloud_events_request( + topic_name: str, event_subscription_name: str, **kwargs: Any +) -> HttpRequest: + _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) + _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) + + content_type: str = kwargs.pop("content_type") + api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2023-06-01-preview")) + accept = _headers.pop("Accept", "application/json") + + # Construct URL + _url = "/topics/{topicName}/eventsubscriptions/{eventSubscriptionName}:release" + path_format_arguments = { + "topicName": _SERIALIZER.url("topic_name", topic_name, "str"), + "eventSubscriptionName": _SERIALIZER.url("event_subscription_name", event_subscription_name, "str"), + } + + _url: str = _format_url_section(_url, **path_format_arguments) # type: ignore + + # Construct parameters + _params["api-version"] = _SERIALIZER.query("api_version", api_version, "str") + + # Construct headers + _headers["content-type"] = _SERIALIZER.header("content_type", content_type, "str") + _headers["Accept"] = _SERIALIZER.header("accept", accept, "str") + + return HttpRequest(method="POST", url=_url, params=_params, headers=_headers, **kwargs) + + +def build_event_grid_reject_cloud_events_request( + topic_name: str, event_subscription_name: str, **kwargs: Any +) -> HttpRequest: + _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) + _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) + + content_type: str = kwargs.pop("content_type") + api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2023-06-01-preview")) + accept = _headers.pop("Accept", "application/json") + + # Construct URL + _url = "/topics/{topicName}/eventsubscriptions/{eventSubscriptionName}:reject" + path_format_arguments = { + "topicName": _SERIALIZER.url("topic_name", topic_name, "str"), + "eventSubscriptionName": _SERIALIZER.url("event_subscription_name", event_subscription_name, "str"), + } + + _url: str = _format_url_section(_url, **path_format_arguments) # type: ignore + + # Construct parameters + _params["api-version"] = _SERIALIZER.query("api_version", api_version, "str") + + # Construct headers + _headers["content-type"] = _SERIALIZER.header("content_type", content_type, "str") + _headers["Accept"] = _SERIALIZER.header("accept", accept, "str") + + return HttpRequest(method="POST", url=_url, params=_params, headers=_headers, **kwargs) + + +class EventGridClientOperationsMixin(EventGridClientMixinABC): + @distributed_trace + def publish_cloud_event( # pylint: disable=inconsistent-return-statements + self, topic_name: str, event: _models.CloudEvent, **kwargs: Any + ) -> None: + """Publish Single Cloud Event to namespace topic. In case of success, the server responds with an + HTTP 200 status code with an empty JSON object in response. Otherwise, the server can return + various error codes. For example, 401: which indicates authorization failure, 403: which + indicates quota exceeded or message is too large, 410: which indicates that specific topic is + not found, 400: for bad request, and 500: for internal server error. + + :param topic_name: Topic Name. Required. + :type topic_name: str + :param event: Single Cloud Event being published. Required. + :type event: ~azure.eventgrid.models.CloudEvent + :keyword content_type: content type. Default value is "application/cloudevents+json; + charset=utf-8". + :paramtype content_type: str + :keyword bool stream: Whether to stream the response of this operation. Defaults to False. You + will have to context manage the returned stream. + :return: None + :rtype: None + :raises ~azure.core.exceptions.HttpResponseError: + """ + error_map = { + 401: ClientAuthenticationError, + 404: ResourceNotFoundError, + 409: ResourceExistsError, + 304: ResourceNotModifiedError, + } + error_map.update(kwargs.pop("error_map", {}) or {}) + + _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) + _params = kwargs.pop("params", {}) or {} + + content_type: str = kwargs.pop( + "content_type", _headers.pop("content-type", "application/cloudevents+json; charset=utf-8") + ) + cls: ClsType[None] = kwargs.pop("cls", None) + + _content = json.dumps(event, cls=AzureJSONEncoder) # type: ignore + + request = build_event_grid_publish_cloud_event_request( + topic_name=topic_name, + content_type=content_type, + api_version=self._config.api_version, + content=_content, + headers=_headers, + params=_params, + ) + path_format_arguments = { + "endpoint": self._serialize.url("self._config.endpoint", self._config.endpoint, "str", skip_quote=True), + } + request.url = self._client.format_url(request.url, **path_format_arguments) + + _stream = kwargs.pop("stream", False) + pipeline_response: PipelineResponse = self._client._pipeline.run( # pylint: disable=protected-access + request, stream=_stream, **kwargs + ) + + response = pipeline_response.http_response + + if response.status_code not in [200]: + map_error(status_code=response.status_code, response=response, error_map=error_map) + raise HttpResponseError(response=response) + + if cls: + return cls(pipeline_response, None, {}) + + @distributed_trace + def publish_cloud_events( # pylint: disable=inconsistent-return-statements + self, topic_name: str, events: List[_models.CloudEvent], **kwargs: Any + ) -> None: + """Publish Batch Cloud Event to namespace topic. In case of success, the server responds with an + HTTP 200 status code with an empty JSON object in response. Otherwise, the server can return + various error codes. For example, 401: which indicates authorization failure, 403: which + indicates quota exceeded or message is too large, 410: which indicates that specific topic is + not found, 400: for bad request, and 500: for internal server error. + + :param topic_name: Topic Name. Required. + :type topic_name: str + :param events: Array of Cloud Events being published. Required. + :type events: list[~azure.eventgrid.models.CloudEvent] + :keyword content_type: content type. Default value is "application/cloudevents-batch+json; + charset=utf-8". + :paramtype content_type: str + :keyword bool stream: Whether to stream the response of this operation. Defaults to False. You + will have to context manage the returned stream. + :return: None + :rtype: None + :raises ~azure.core.exceptions.HttpResponseError: + """ + error_map = { + 401: ClientAuthenticationError, + 404: ResourceNotFoundError, + 409: ResourceExistsError, + 304: ResourceNotModifiedError, + } + error_map.update(kwargs.pop("error_map", {}) or {}) + + _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) + _params = kwargs.pop("params", {}) or {} + + content_type: str = kwargs.pop( + "content_type", _headers.pop("content-type", "application/cloudevents-batch+json; charset=utf-8") + ) + cls: ClsType[None] = kwargs.pop("cls", None) + + _content = json.dumps(events, cls=AzureJSONEncoder) # type: ignore + + request = build_event_grid_publish_cloud_events_request( + topic_name=topic_name, + content_type=content_type, + api_version=self._config.api_version, + content=_content, + headers=_headers, + params=_params, + ) + path_format_arguments = { + "endpoint": self._serialize.url("self._config.endpoint", self._config.endpoint, "str", skip_quote=True), + } + request.url = self._client.format_url(request.url, **path_format_arguments) + + _stream = kwargs.pop("stream", False) + pipeline_response: PipelineResponse = self._client._pipeline.run( # pylint: disable=protected-access + request, stream=_stream, **kwargs + ) + + response = pipeline_response.http_response + + if response.status_code not in [200]: + map_error(status_code=response.status_code, response=response, error_map=error_map) + raise HttpResponseError(response=response) + + if cls: + return cls(pipeline_response, None, {}) + + @distributed_trace + def receive_cloud_events( + self, + topic_name: str, + event_subscription_name: str, + *, + max_events: Optional[int] = None, + max_wait_time: Optional[datetime.timedelta] = None, + **kwargs: Any + ) -> _models.ReceiveResult: + """Receive Batch of Cloud Events from the Event Subscription. + + :param topic_name: Topic Name. Required. + :type topic_name: str + :param event_subscription_name: Event Subscription Name. Required. + :type event_subscription_name: str + :keyword max_events: Max Events count to be received. Minimum value is 1, while maximum value + is 100 events. If not specified, the default value is 1. Default value is None. + :paramtype max_events: int + :keyword max_wait_time: Max wait time value for receive operation in Seconds. It is the time in + seconds that the server approximately waits for the availability of an event and responds to + the request. If an event is available, the broker responds immediately to the client. Minimum + value is 10 seconds, while maximum value is 120 seconds. If not specified, the default value is + 60 seconds. Default value is None. + :paramtype max_wait_time: ~datetime.timedelta + :keyword bool stream: Whether to stream the response of this operation. Defaults to False. You + will have to context manage the returned stream. + :return: ReceiveResult. The ReceiveResult is compatible with MutableMapping + :rtype: ~azure.eventgrid.models.ReceiveResult + :raises ~azure.core.exceptions.HttpResponseError: + """ + error_map = { + 401: ClientAuthenticationError, + 404: ResourceNotFoundError, + 409: ResourceExistsError, + 304: ResourceNotModifiedError, + } + error_map.update(kwargs.pop("error_map", {}) or {}) + + _headers = kwargs.pop("headers", {}) or {} + _params = kwargs.pop("params", {}) or {} + + cls: ClsType[_models.ReceiveResult] = kwargs.pop("cls", None) + + request = build_event_grid_receive_cloud_events_request( + topic_name=topic_name, + event_subscription_name=event_subscription_name, + max_events=max_events, + max_wait_time=max_wait_time, + api_version=self._config.api_version, + headers=_headers, + params=_params, + ) + path_format_arguments = { + "endpoint": self._serialize.url("self._config.endpoint", self._config.endpoint, "str", skip_quote=True), + } + request.url = self._client.format_url(request.url, **path_format_arguments) + + _stream = kwargs.pop("stream", False) + pipeline_response: PipelineResponse = self._client._pipeline.run( # pylint: disable=protected-access + request, stream=_stream, **kwargs + ) + + response = pipeline_response.http_response + + if response.status_code not in [200]: + map_error(status_code=response.status_code, response=response, error_map=error_map) + raise HttpResponseError(response=response) + + if _stream: + deserialized = response.iter_bytes() + else: + deserialized = _deserialize(_models.ReceiveResult, response.json()) + + if cls: + return cls(pipeline_response, deserialized, {}) # type: ignore + + return deserialized # type: ignore + + @distributed_trace + def acknowledge_cloud_events( + self, topic_name: str, event_subscription_name: str, lock_tokens: _models.AcknowledgeOptions, **kwargs: Any + ) -> _models.AcknowledgeResult: + """Acknowledge batch of Cloud Events. The server responds with an HTTP 200 status code if at least + one event is successfully acknowledged. The response body will include the set of successfully + acknowledged lockTokens, along with other failed lockTokens with their corresponding error + information. Successfully acknowledged events will no longer be available to any consumer. + + :param topic_name: Topic Name. Required. + :type topic_name: str + :param event_subscription_name: Event Subscription Name. Required. + :type event_subscription_name: str + :param lock_tokens: AcknowledgeOptions. Required. + :type lock_tokens: ~azure.eventgrid.models.AcknowledgeOptions + :keyword content_type: content type. Default value is "application/json; charset=utf-8". + :paramtype content_type: str + :keyword bool stream: Whether to stream the response of this operation. Defaults to False. You + will have to context manage the returned stream. + :return: AcknowledgeResult. The AcknowledgeResult is compatible with MutableMapping + :rtype: ~azure.eventgrid.models.AcknowledgeResult + :raises ~azure.core.exceptions.HttpResponseError: + """ + error_map = { + 401: ClientAuthenticationError, + 404: ResourceNotFoundError, + 409: ResourceExistsError, + 304: ResourceNotModifiedError, + } + error_map.update(kwargs.pop("error_map", {}) or {}) + + _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) + _params = kwargs.pop("params", {}) or {} + + content_type: str = kwargs.pop("content_type", _headers.pop("content-type", "application/json; charset=utf-8")) + cls: ClsType[_models.AcknowledgeResult] = kwargs.pop("cls", None) + + _content = json.dumps(lock_tokens, cls=AzureJSONEncoder) # type: ignore + + request = build_event_grid_acknowledge_cloud_events_request( + topic_name=topic_name, + event_subscription_name=event_subscription_name, + content_type=content_type, + api_version=self._config.api_version, + content=_content, + headers=_headers, + params=_params, + ) + path_format_arguments = { + "endpoint": self._serialize.url("self._config.endpoint", self._config.endpoint, "str", skip_quote=True), + } + request.url = self._client.format_url(request.url, **path_format_arguments) + + _stream = kwargs.pop("stream", False) + pipeline_response: PipelineResponse = self._client._pipeline.run( # pylint: disable=protected-access + request, stream=_stream, **kwargs + ) + + response = pipeline_response.http_response + + if response.status_code not in [200]: + map_error(status_code=response.status_code, response=response, error_map=error_map) + raise HttpResponseError(response=response) + + if _stream: + deserialized = response.iter_bytes() + else: + deserialized = _deserialize(_models.AcknowledgeResult, response.json()) + + if cls: + return cls(pipeline_response, deserialized, {}) # type: ignore + + return deserialized # type: ignore + + @distributed_trace + def release_cloud_events( + self, topic_name: str, event_subscription_name: str, lock_tokens: _models.ReleaseOptions, **kwargs: Any + ) -> _models.ReleaseResult: + """Release batch of Cloud Events. The server responds with an HTTP 200 status code if at least one + event is successfully released. The response body will include the set of successfully released + lockTokens, along with other failed lockTokens with their corresponding error information. + + :param topic_name: Topic Name. Required. + :type topic_name: str + :param event_subscription_name: Event Subscription Name. Required. + :type event_subscription_name: str + :param lock_tokens: ReleaseOptions. Required. + :type lock_tokens: ~azure.eventgrid.models.ReleaseOptions + :keyword content_type: content type. Default value is "application/json; charset=utf-8". + :paramtype content_type: str + :keyword bool stream: Whether to stream the response of this operation. Defaults to False. You + will have to context manage the returned stream. + :return: ReleaseResult. The ReleaseResult is compatible with MutableMapping + :rtype: ~azure.eventgrid.models.ReleaseResult + :raises ~azure.core.exceptions.HttpResponseError: + """ + error_map = { + 401: ClientAuthenticationError, + 404: ResourceNotFoundError, + 409: ResourceExistsError, + 304: ResourceNotModifiedError, + } + error_map.update(kwargs.pop("error_map", {}) or {}) + + _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) + _params = kwargs.pop("params", {}) or {} + + content_type: str = kwargs.pop("content_type", _headers.pop("content-type", "application/json; charset=utf-8")) + cls: ClsType[_models.ReleaseResult] = kwargs.pop("cls", None) + + _content = json.dumps(lock_tokens, cls=AzureJSONEncoder) # type: ignore + + request = build_event_grid_release_cloud_events_request( + topic_name=topic_name, + event_subscription_name=event_subscription_name, + content_type=content_type, + api_version=self._config.api_version, + content=_content, + headers=_headers, + params=_params, + ) + path_format_arguments = { + "endpoint": self._serialize.url("self._config.endpoint", self._config.endpoint, "str", skip_quote=True), + } + request.url = self._client.format_url(request.url, **path_format_arguments) + + _stream = kwargs.pop("stream", False) + pipeline_response: PipelineResponse = self._client._pipeline.run( # pylint: disable=protected-access + request, stream=_stream, **kwargs + ) + + response = pipeline_response.http_response + + if response.status_code not in [200]: + map_error(status_code=response.status_code, response=response, error_map=error_map) + raise HttpResponseError(response=response) + + if _stream: + deserialized = response.iter_bytes() + else: + deserialized = _deserialize(_models.ReleaseResult, response.json()) + + if cls: + return cls(pipeline_response, deserialized, {}) # type: ignore + + return deserialized # type: ignore + + @distributed_trace + def reject_cloud_events( + self, topic_name: str, event_subscription_name: str, lock_tokens: _models.RejectOptions, **kwargs: Any + ) -> _models.RejectResult: + """Reject batch of Cloud Events. + + :param topic_name: Topic Name. Required. + :type topic_name: str + :param event_subscription_name: Event Subscription Name. Required. + :type event_subscription_name: str + :param lock_tokens: RejectOptions. Required. + :type lock_tokens: ~azure.eventgrid.models.RejectOptions + :keyword content_type: content type. Default value is "application/json; charset=utf-8". + :paramtype content_type: str + :keyword bool stream: Whether to stream the response of this operation. Defaults to False. You + will have to context manage the returned stream. + :return: RejectResult. The RejectResult is compatible with MutableMapping + :rtype: ~azure.eventgrid.models.RejectResult + :raises ~azure.core.exceptions.HttpResponseError: + """ + error_map = { + 401: ClientAuthenticationError, + 404: ResourceNotFoundError, + 409: ResourceExistsError, + 304: ResourceNotModifiedError, + } + error_map.update(kwargs.pop("error_map", {}) or {}) + + _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) + _params = kwargs.pop("params", {}) or {} + + content_type: str = kwargs.pop("content_type", _headers.pop("content-type", "application/json; charset=utf-8")) + cls: ClsType[_models.RejectResult] = kwargs.pop("cls", None) + + _content = json.dumps(lock_tokens, cls=AzureJSONEncoder) # type: ignore + + request = build_event_grid_reject_cloud_events_request( + topic_name=topic_name, + event_subscription_name=event_subscription_name, + content_type=content_type, + api_version=self._config.api_version, + content=_content, + headers=_headers, + params=_params, + ) + path_format_arguments = { + "endpoint": self._serialize.url("self._config.endpoint", self._config.endpoint, "str", skip_quote=True), + } + request.url = self._client.format_url(request.url, **path_format_arguments) + + _stream = kwargs.pop("stream", False) + pipeline_response: PipelineResponse = self._client._pipeline.run( # pylint: disable=protected-access + request, stream=_stream, **kwargs + ) + + response = pipeline_response.http_response + + if response.status_code not in [200]: + map_error(status_code=response.status_code, response=response, error_map=error_map) + raise HttpResponseError(response=response) + + if _stream: + deserialized = response.iter_bytes() + else: + deserialized = _deserialize(_models.RejectResult, response.json()) + + if cls: + return cls(pipeline_response, deserialized, {}) # type: ignore + + return deserialized # type: ignore diff --git a/sdk/eventgrid/azure-eventgrid/azure/eventgrid/_operations/_patch.py b/sdk/eventgrid/azure-eventgrid/azure/eventgrid/_operations/_patch.py new file mode 100644 index 000000000000..f7dd32510333 --- /dev/null +++ b/sdk/eventgrid/azure-eventgrid/azure/eventgrid/_operations/_patch.py @@ -0,0 +1,20 @@ +# ------------------------------------ +# Copyright (c) Microsoft Corporation. +# Licensed under the MIT License. +# ------------------------------------ +"""Customize generated code here. + +Follow our quickstart for examples: https://aka.ms/azsdk/python/dpcodegen/python/customize +""" +from typing import List + +__all__: List[str] = [] # Add all objects you want publicly available to users at this package level + + +def patch_sdk(): + """Do not remove from this file. + + `patch_sdk` is a last resort escape hatch that allows you to do customizations + you can't accomplish using the techniques described in + https://aka.ms/azsdk/python/dpcodegen/python/customize + """ diff --git a/sdk/eventgrid/azure-eventgrid/azure/eventgrid/_patch.py b/sdk/eventgrid/azure-eventgrid/azure/eventgrid/_patch.py new file mode 100644 index 000000000000..f7dd32510333 --- /dev/null +++ b/sdk/eventgrid/azure-eventgrid/azure/eventgrid/_patch.py @@ -0,0 +1,20 @@ +# ------------------------------------ +# Copyright (c) Microsoft Corporation. +# Licensed under the MIT License. +# ------------------------------------ +"""Customize generated code here. + +Follow our quickstart for examples: https://aka.ms/azsdk/python/dpcodegen/python/customize +""" +from typing import List + +__all__: List[str] = [] # Add all objects you want publicly available to users at this package level + + +def patch_sdk(): + """Do not remove from this file. + + `patch_sdk` is a last resort escape hatch that allows you to do customizations + you can't accomplish using the techniques described in + https://aka.ms/azsdk/python/dpcodegen/python/customize + """ diff --git a/sdk/eventgrid/azure-eventgrid/azure/eventgrid/_policies.py b/sdk/eventgrid/azure-eventgrid/azure/eventgrid/_policies.py index 13a19d60bbcd..8ba95cb7c649 100644 --- a/sdk/eventgrid/azure-eventgrid/azure/eventgrid/_policies.py +++ b/sdk/eventgrid/azure-eventgrid/azure/eventgrid/_policies.py @@ -36,8 +36,7 @@ def on_request(self, request): return if ( - request.http_request.headers["content-type"] - == CloudEventDistributedTracingPolicy._CONTENT_TYPE + request.http_request.headers["content-type"] == CloudEventDistributedTracingPolicy._CONTENT_TYPE and traceparent is not None ): diff --git a/sdk/eventgrid/azure-eventgrid/azure/eventgrid/_publisher_client.py b/sdk/eventgrid/azure-eventgrid/azure/eventgrid/_publisher_client.py index a69e3ec35abc..5ac30d31f412 100644 --- a/sdk/eventgrid/azure-eventgrid/azure/eventgrid/_publisher_client.py +++ b/sdk/eventgrid/azure-eventgrid/azure/eventgrid/_publisher_client.py @@ -26,7 +26,7 @@ HttpResponseError, ResourceNotFoundError, ResourceExistsError, - map_error + map_error, ) from azure.core.messaging import CloudEvent @@ -66,7 +66,7 @@ ListEventType = Union[List[CloudEvent], List[EventGridEvent], List[Dict]] -class EventGridPublisherClient(object): # pylint: disable=client-accepts-api-version-keyword +class EventGridPublisherClient(object): # pylint: disable=client-accepts-api-version-keyword """EventGridPublisherClient publishes events to an EventGrid topic or domain. It can be used to publish either an EventGridEvent, a CloudEvent or a Custom Schema. @@ -124,13 +124,7 @@ def _policies(credential, **kwargs): return policies @distributed_trace - def send( - self, - events: SendType, - *, - channel_name: Optional[str] = None, - **kwargs: Any - ) -> None: + def send(self, events: SendType, *, channel_name: Optional[str] = None, **kwargs: Any) -> None: """Sends events to a topic or a domain specified during the client initialization. A single instance or a list of dictionaries, CloudEvents or EventGridEvents are accepted. @@ -203,15 +197,10 @@ def send( content_type = kwargs.pop("content_type", "application/json; charset=utf-8") if isinstance(events[0], CloudEvent) or _is_cloud_event(events[0]): try: - events = [ - _cloud_event_to_generated(e, **kwargs) - for e in events # pylint: disable=protected-access - ] + events = [_cloud_event_to_generated(e, **kwargs) for e in events] # pylint: disable=protected-access except AttributeError: ## this is either a dictionary or a CNCF cloud event - events = [ - _from_cncf_events(e) for e in events - ] + events = [_from_cncf_events(e) for e in events] content_type = "application/cloudevents-batch+json; charset=utf-8" elif isinstance(events[0], EventGridEvent) or _is_eventgrid_event(events[0]): for event in events: diff --git a/sdk/eventgrid/azure-eventgrid/azure/eventgrid/_serialization.py b/sdk/eventgrid/azure-eventgrid/azure/eventgrid/_serialization.py new file mode 100644 index 000000000000..842ae727fbbc --- /dev/null +++ b/sdk/eventgrid/azure-eventgrid/azure/eventgrid/_serialization.py @@ -0,0 +1,1996 @@ +# -------------------------------------------------------------------------- +# +# Copyright (c) Microsoft Corporation. All rights reserved. +# +# The MIT License (MIT) +# +# Permission is hereby granted, free of charge, to any person obtaining a copy +# of this software and associated documentation files (the ""Software""), to +# deal in the Software without restriction, including without limitation the +# rights to use, copy, modify, merge, publish, distribute, sublicense, and/or +# sell copies of the Software, and to permit persons to whom the Software is +# furnished to do so, subject to the following conditions: +# +# The above copyright notice and this permission notice shall be included in +# all copies or substantial portions of the Software. +# +# THE SOFTWARE IS PROVIDED *AS IS*, WITHOUT WARRANTY OF ANY KIND, EXPRESS OR +# IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, +# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE +# AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER +# LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING +# FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS +# IN THE SOFTWARE. +# +# -------------------------------------------------------------------------- + +# pylint: skip-file +# pyright: reportUnnecessaryTypeIgnoreComment=false + +from base64 import b64decode, b64encode +import calendar +import datetime +import decimal +import email +from enum import Enum +import json +import logging +import re +import sys +import codecs +from typing import ( + Dict, + Any, + cast, + Optional, + Union, + AnyStr, + IO, + Mapping, + Callable, + TypeVar, + MutableMapping, + Type, + List, + Mapping, +) + +try: + from urllib import quote # type: ignore +except ImportError: + from urllib.parse import quote +import xml.etree.ElementTree as ET + +import isodate # type: ignore + +from azure.core.exceptions import DeserializationError, SerializationError, raise_with_traceback +from azure.core.serialization import NULL as AzureCoreNull + +_BOM = codecs.BOM_UTF8.decode(encoding="utf-8") + +ModelType = TypeVar("ModelType", bound="Model") +JSON = MutableMapping[str, Any] + + +class RawDeserializer: + + # Accept "text" because we're open minded people... + JSON_REGEXP = re.compile(r"^(application|text)/([a-z+.]+\+)?json$") + + # Name used in context + CONTEXT_NAME = "deserialized_data" + + @classmethod + def deserialize_from_text(cls, data: Optional[Union[AnyStr, IO]], content_type: Optional[str] = None) -> Any: + """Decode data according to content-type. + + Accept a stream of data as well, but will be load at once in memory for now. + + If no content-type, will return the string version (not bytes, not stream) + + :param data: Input, could be bytes or stream (will be decoded with UTF8) or text + :type data: str or bytes or IO + :param str content_type: The content type. + """ + if hasattr(data, "read"): + # Assume a stream + data = cast(IO, data).read() + + if isinstance(data, bytes): + data_as_str = data.decode(encoding="utf-8-sig") + else: + # Explain to mypy the correct type. + data_as_str = cast(str, data) + + # Remove Byte Order Mark if present in string + data_as_str = data_as_str.lstrip(_BOM) + + if content_type is None: + return data + + if cls.JSON_REGEXP.match(content_type): + try: + return json.loads(data_as_str) + except ValueError as err: + raise DeserializationError("JSON is invalid: {}".format(err), err) + elif "xml" in (content_type or []): + try: + + try: + if isinstance(data, unicode): # type: ignore + # If I'm Python 2.7 and unicode XML will scream if I try a "fromstring" on unicode string + data_as_str = data_as_str.encode(encoding="utf-8") # type: ignore + except NameError: + pass + + return ET.fromstring(data_as_str) # nosec + except ET.ParseError: + # It might be because the server has an issue, and returned JSON with + # content-type XML.... + # So let's try a JSON load, and if it's still broken + # let's flow the initial exception + def _json_attemp(data): + try: + return True, json.loads(data) + except ValueError: + return False, None # Don't care about this one + + success, json_result = _json_attemp(data) + if success: + return json_result + # If i'm here, it's not JSON, it's not XML, let's scream + # and raise the last context in this block (the XML exception) + # The function hack is because Py2.7 messes up with exception + # context otherwise. + _LOGGER.critical("Wasn't XML not JSON, failing") + raise_with_traceback(DeserializationError, "XML is invalid") + raise DeserializationError("Cannot deserialize content-type: {}".format(content_type)) + + @classmethod + def deserialize_from_http_generics(cls, body_bytes: Optional[Union[AnyStr, IO]], headers: Mapping) -> Any: + """Deserialize from HTTP response. + + Use bytes and headers to NOT use any requests/aiohttp or whatever + specific implementation. + Headers will tested for "content-type" + """ + # Try to use content-type from headers if available + content_type = None + if "content-type" in headers: + content_type = headers["content-type"].split(";")[0].strip().lower() + # Ouch, this server did not declare what it sent... + # Let's guess it's JSON... + # Also, since Autorest was considering that an empty body was a valid JSON, + # need that test as well.... + else: + content_type = "application/json" + + if body_bytes: + return cls.deserialize_from_text(body_bytes, content_type) + return None + + +try: + basestring # type: ignore + unicode_str = unicode # type: ignore +except NameError: + basestring = str + unicode_str = str + +_LOGGER = logging.getLogger(__name__) + +try: + _long_type = long # type: ignore +except NameError: + _long_type = int + + +class UTC(datetime.tzinfo): + """Time Zone info for handling UTC""" + + def utcoffset(self, dt): + """UTF offset for UTC is 0.""" + return datetime.timedelta(0) + + def tzname(self, dt): + """Timestamp representation.""" + return "Z" + + def dst(self, dt): + """No daylight saving for UTC.""" + return datetime.timedelta(hours=1) + + +try: + from datetime import timezone as _FixedOffset # type: ignore +except ImportError: # Python 2.7 + + class _FixedOffset(datetime.tzinfo): # type: ignore + """Fixed offset in minutes east from UTC. + Copy/pasted from Python doc + :param datetime.timedelta offset: offset in timedelta format + """ + + def __init__(self, offset): + self.__offset = offset + + def utcoffset(self, dt): + return self.__offset + + def tzname(self, dt): + return str(self.__offset.total_seconds() / 3600) + + def __repr__(self): + return "".format(self.tzname(None)) + + def dst(self, dt): + return datetime.timedelta(0) + + def __getinitargs__(self): + return (self.__offset,) + + +try: + from datetime import timezone + + TZ_UTC = timezone.utc +except ImportError: + TZ_UTC = UTC() # type: ignore + +_FLATTEN = re.compile(r"(? None: + self.additional_properties: Dict[str, Any] = {} + for k in kwargs: + if k not in self._attribute_map: + _LOGGER.warning("%s is not a known attribute of class %s and will be ignored", k, self.__class__) + elif k in self._validation and self._validation[k].get("readonly", False): + _LOGGER.warning("Readonly attribute %s will be ignored in class %s", k, self.__class__) + else: + setattr(self, k, kwargs[k]) + + def __eq__(self, other: Any) -> bool: + """Compare objects by comparing all attributes.""" + if isinstance(other, self.__class__): + return self.__dict__ == other.__dict__ + return False + + def __ne__(self, other: Any) -> bool: + """Compare objects by comparing all attributes.""" + return not self.__eq__(other) + + def __str__(self) -> str: + return str(self.__dict__) + + @classmethod + def enable_additional_properties_sending(cls) -> None: + cls._attribute_map["additional_properties"] = {"key": "", "type": "{object}"} + + @classmethod + def is_xml_model(cls) -> bool: + try: + cls._xml_map # type: ignore + except AttributeError: + return False + return True + + @classmethod + def _create_xml_node(cls): + """Create XML node.""" + try: + xml_map = cls._xml_map # type: ignore + except AttributeError: + xml_map = {} + + return _create_xml_node(xml_map.get("name", cls.__name__), xml_map.get("prefix", None), xml_map.get("ns", None)) + + def serialize(self, keep_readonly: bool = False, **kwargs: Any) -> JSON: + """Return the JSON that would be sent to azure from this model. + + This is an alias to `as_dict(full_restapi_key_transformer, keep_readonly=False)`. + + If you want XML serialization, you can pass the kwargs is_xml=True. + + :param bool keep_readonly: If you want to serialize the readonly attributes + :returns: A dict JSON compatible object + :rtype: dict + """ + serializer = Serializer(self._infer_class_models()) + return serializer._serialize(self, keep_readonly=keep_readonly, **kwargs) + + def as_dict( + self, + keep_readonly: bool = True, + key_transformer: Callable[[str, Dict[str, Any], Any], Any] = attribute_transformer, + **kwargs: Any + ) -> JSON: + """Return a dict that can be serialized using json.dump. + + Advanced usage might optionally use a callback as parameter: + + .. code::python + + def my_key_transformer(key, attr_desc, value): + return key + + Key is the attribute name used in Python. Attr_desc + is a dict of metadata. Currently contains 'type' with the + msrest type and 'key' with the RestAPI encoded key. + Value is the current value in this object. + + The string returned will be used to serialize the key. + If the return type is a list, this is considered hierarchical + result dict. + + See the three examples in this file: + + - attribute_transformer + - full_restapi_key_transformer + - last_restapi_key_transformer + + If you want XML serialization, you can pass the kwargs is_xml=True. + + :param function key_transformer: A key transformer function. + :returns: A dict JSON compatible object + :rtype: dict + """ + serializer = Serializer(self._infer_class_models()) + return serializer._serialize(self, key_transformer=key_transformer, keep_readonly=keep_readonly, **kwargs) + + @classmethod + def _infer_class_models(cls): + try: + str_models = cls.__module__.rsplit(".", 1)[0] + models = sys.modules[str_models] + client_models = {k: v for k, v in models.__dict__.items() if isinstance(v, type)} + if cls.__name__ not in client_models: + raise ValueError("Not Autorest generated code") + except Exception: + # Assume it's not Autorest generated (tests?). Add ourselves as dependencies. + client_models = {cls.__name__: cls} + return client_models + + @classmethod + def deserialize(cls: Type[ModelType], data: Any, content_type: Optional[str] = None) -> ModelType: + """Parse a str using the RestAPI syntax and return a model. + + :param str data: A str using RestAPI structure. JSON by default. + :param str content_type: JSON by default, set application/xml if XML. + :returns: An instance of this model + :raises: DeserializationError if something went wrong + """ + deserializer = Deserializer(cls._infer_class_models()) + return deserializer(cls.__name__, data, content_type=content_type) + + @classmethod + def from_dict( + cls: Type[ModelType], + data: Any, + key_extractors: Optional[Callable[[str, Dict[str, Any], Any], Any]] = None, + content_type: Optional[str] = None, + ) -> ModelType: + """Parse a dict using given key extractor return a model. + + By default consider key + extractors (rest_key_case_insensitive_extractor, attribute_key_case_insensitive_extractor + and last_rest_key_case_insensitive_extractor) + + :param dict data: A dict using RestAPI structure + :param str content_type: JSON by default, set application/xml if XML. + :returns: An instance of this model + :raises: DeserializationError if something went wrong + """ + deserializer = Deserializer(cls._infer_class_models()) + deserializer.key_extractors = ( # type: ignore + [ # type: ignore + attribute_key_case_insensitive_extractor, + rest_key_case_insensitive_extractor, + last_rest_key_case_insensitive_extractor, + ] + if key_extractors is None + else key_extractors + ) + return deserializer(cls.__name__, data, content_type=content_type) + + @classmethod + def _flatten_subtype(cls, key, objects): + if "_subtype_map" not in cls.__dict__: + return {} + result = dict(cls._subtype_map[key]) + for valuetype in cls._subtype_map[key].values(): + result.update(objects[valuetype]._flatten_subtype(key, objects)) + return result + + @classmethod + def _classify(cls, response, objects): + """Check the class _subtype_map for any child classes. + We want to ignore any inherited _subtype_maps. + Remove the polymorphic key from the initial data. + """ + for subtype_key in cls.__dict__.get("_subtype_map", {}).keys(): + subtype_value = None + + if not isinstance(response, ET.Element): + rest_api_response_key = cls._get_rest_key_parts(subtype_key)[-1] + subtype_value = response.pop(rest_api_response_key, None) or response.pop(subtype_key, None) + else: + subtype_value = xml_key_extractor(subtype_key, cls._attribute_map[subtype_key], response) + if subtype_value: + # Try to match base class. Can be class name only + # (bug to fix in Autorest to support x-ms-discriminator-name) + if cls.__name__ == subtype_value: + return cls + flatten_mapping_type = cls._flatten_subtype(subtype_key, objects) + try: + return objects[flatten_mapping_type[subtype_value]] # type: ignore + except KeyError: + _LOGGER.warning( + "Subtype value %s has no mapping, use base class %s.", + subtype_value, + cls.__name__, + ) + break + else: + _LOGGER.warning("Discriminator %s is absent or null, use base class %s.", subtype_key, cls.__name__) + break + return cls + + @classmethod + def _get_rest_key_parts(cls, attr_key): + """Get the RestAPI key of this attr, split it and decode part + :param str attr_key: Attribute key must be in attribute_map. + :returns: A list of RestAPI part + :rtype: list + """ + rest_split_key = _FLATTEN.split(cls._attribute_map[attr_key]["key"]) + return [_decode_attribute_map_key(key_part) for key_part in rest_split_key] + + +def _decode_attribute_map_key(key): + """This decode a key in an _attribute_map to the actual key we want to look at + inside the received data. + + :param str key: A key string from the generated code + """ + return key.replace("\\.", ".") + + +class Serializer(object): + """Request object model serializer.""" + + basic_types = {str: "str", int: "int", bool: "bool", float: "float"} + + _xml_basic_types_serializers = {"bool": lambda x: str(x).lower()} + days = {0: "Mon", 1: "Tue", 2: "Wed", 3: "Thu", 4: "Fri", 5: "Sat", 6: "Sun"} + months = { + 1: "Jan", + 2: "Feb", + 3: "Mar", + 4: "Apr", + 5: "May", + 6: "Jun", + 7: "Jul", + 8: "Aug", + 9: "Sep", + 10: "Oct", + 11: "Nov", + 12: "Dec", + } + validation = { + "min_length": lambda x, y: len(x) < y, + "max_length": lambda x, y: len(x) > y, + "minimum": lambda x, y: x < y, + "maximum": lambda x, y: x > y, + "minimum_ex": lambda x, y: x <= y, + "maximum_ex": lambda x, y: x >= y, + "min_items": lambda x, y: len(x) < y, + "max_items": lambda x, y: len(x) > y, + "pattern": lambda x, y: not re.match(y, x, re.UNICODE), + "unique": lambda x, y: len(x) != len(set(x)), + "multiple": lambda x, y: x % y != 0, + } + + def __init__(self, classes: Optional[Mapping[str, Type[ModelType]]] = None): + self.serialize_type = { + "iso-8601": Serializer.serialize_iso, + "rfc-1123": Serializer.serialize_rfc, + "unix-time": Serializer.serialize_unix, + "duration": Serializer.serialize_duration, + "date": Serializer.serialize_date, + "time": Serializer.serialize_time, + "decimal": Serializer.serialize_decimal, + "long": Serializer.serialize_long, + "bytearray": Serializer.serialize_bytearray, + "base64": Serializer.serialize_base64, + "object": self.serialize_object, + "[]": self.serialize_iter, + "{}": self.serialize_dict, + } + self.dependencies: Dict[str, Type[ModelType]] = dict(classes) if classes else {} + self.key_transformer = full_restapi_key_transformer + self.client_side_validation = True + + def _serialize(self, target_obj, data_type=None, **kwargs): + """Serialize data into a string according to type. + + :param target_obj: The data to be serialized. + :param str data_type: The type to be serialized from. + :rtype: str, dict + :raises: SerializationError if serialization fails. + """ + key_transformer = kwargs.get("key_transformer", self.key_transformer) + keep_readonly = kwargs.get("keep_readonly", False) + if target_obj is None: + return None + + attr_name = None + class_name = target_obj.__class__.__name__ + + if data_type: + return self.serialize_data(target_obj, data_type, **kwargs) + + if not hasattr(target_obj, "_attribute_map"): + data_type = type(target_obj).__name__ + if data_type in self.basic_types.values(): + return self.serialize_data(target_obj, data_type, **kwargs) + + # Force "is_xml" kwargs if we detect a XML model + try: + is_xml_model_serialization = kwargs["is_xml"] + except KeyError: + is_xml_model_serialization = kwargs.setdefault("is_xml", target_obj.is_xml_model()) + + serialized = {} + if is_xml_model_serialization: + serialized = target_obj._create_xml_node() + try: + attributes = target_obj._attribute_map + for attr, attr_desc in attributes.items(): + attr_name = attr + if not keep_readonly and target_obj._validation.get(attr_name, {}).get("readonly", False): + continue + + if attr_name == "additional_properties" and attr_desc["key"] == "": + if target_obj.additional_properties is not None: + serialized.update(target_obj.additional_properties) + continue + try: + + orig_attr = getattr(target_obj, attr) + if is_xml_model_serialization: + pass # Don't provide "transformer" for XML for now. Keep "orig_attr" + else: # JSON + keys, orig_attr = key_transformer(attr, attr_desc.copy(), orig_attr) + keys = keys if isinstance(keys, list) else [keys] + + kwargs["serialization_ctxt"] = attr_desc + new_attr = self.serialize_data(orig_attr, attr_desc["type"], **kwargs) + + if is_xml_model_serialization: + xml_desc = attr_desc.get("xml", {}) + xml_name = xml_desc.get("name", attr_desc["key"]) + xml_prefix = xml_desc.get("prefix", None) + xml_ns = xml_desc.get("ns", None) + if xml_desc.get("attr", False): + if xml_ns: + ET.register_namespace(xml_prefix, xml_ns) + xml_name = "{{{}}}{}".format(xml_ns, xml_name) + serialized.set(xml_name, new_attr) # type: ignore + continue + if xml_desc.get("text", False): + serialized.text = new_attr # type: ignore + continue + if isinstance(new_attr, list): + serialized.extend(new_attr) # type: ignore + elif isinstance(new_attr, ET.Element): + # If the down XML has no XML/Name, we MUST replace the tag with the local tag. But keeping the namespaces. + if "name" not in getattr(orig_attr, "_xml_map", {}): + splitted_tag = new_attr.tag.split("}") + if len(splitted_tag) == 2: # Namespace + new_attr.tag = "}".join([splitted_tag[0], xml_name]) + else: + new_attr.tag = xml_name + serialized.append(new_attr) # type: ignore + else: # That's a basic type + # Integrate namespace if necessary + local_node = _create_xml_node(xml_name, xml_prefix, xml_ns) + local_node.text = unicode_str(new_attr) + serialized.append(local_node) # type: ignore + else: # JSON + for k in reversed(keys): # type: ignore + new_attr = {k: new_attr} + + _new_attr = new_attr + _serialized = serialized + for k in keys: # type: ignore + if k not in _serialized: + _serialized.update(_new_attr) # type: ignore + _new_attr = _new_attr[k] # type: ignore + _serialized = _serialized[k] + except ValueError: + continue + + except (AttributeError, KeyError, TypeError) as err: + msg = "Attribute {} in object {} cannot be serialized.\n{}".format(attr_name, class_name, str(target_obj)) + raise_with_traceback(SerializationError, msg, err) + else: + return serialized + + def body(self, data, data_type, **kwargs): + """Serialize data intended for a request body. + + :param data: The data to be serialized. + :param str data_type: The type to be serialized from. + :rtype: dict + :raises: SerializationError if serialization fails. + :raises: ValueError if data is None + """ + + # Just in case this is a dict + internal_data_type_str = data_type.strip("[]{}") + internal_data_type = self.dependencies.get(internal_data_type_str, None) + try: + is_xml_model_serialization = kwargs["is_xml"] + except KeyError: + if internal_data_type and issubclass(internal_data_type, Model): + is_xml_model_serialization = kwargs.setdefault("is_xml", internal_data_type.is_xml_model()) + else: + is_xml_model_serialization = False + if internal_data_type and not isinstance(internal_data_type, Enum): + try: + deserializer = Deserializer(self.dependencies) + # Since it's on serialization, it's almost sure that format is not JSON REST + # We're not able to deal with additional properties for now. + deserializer.additional_properties_detection = False + if is_xml_model_serialization: + deserializer.key_extractors = [ # type: ignore + attribute_key_case_insensitive_extractor, + ] + else: + deserializer.key_extractors = [ + rest_key_case_insensitive_extractor, + attribute_key_case_insensitive_extractor, + last_rest_key_case_insensitive_extractor, + ] + data = deserializer._deserialize(data_type, data) + except DeserializationError as err: + raise_with_traceback(SerializationError, "Unable to build a model: " + str(err), err) + + return self._serialize(data, data_type, **kwargs) + + def url(self, name, data, data_type, **kwargs): + """Serialize data intended for a URL path. + + :param data: The data to be serialized. + :param str data_type: The type to be serialized from. + :rtype: str + :raises: TypeError if serialization fails. + :raises: ValueError if data is None + """ + try: + output = self.serialize_data(data, data_type, **kwargs) + if data_type == "bool": + output = json.dumps(output) + + if kwargs.get("skip_quote") is True: + output = str(output) + else: + output = quote(str(output), safe="") + except SerializationError: + raise TypeError("{} must be type {}.".format(name, data_type)) + else: + return output + + def query(self, name, data, data_type, **kwargs): + """Serialize data intended for a URL query. + + :param data: The data to be serialized. + :param str data_type: The type to be serialized from. + :rtype: str + :raises: TypeError if serialization fails. + :raises: ValueError if data is None + """ + try: + # Treat the list aside, since we don't want to encode the div separator + if data_type.startswith("["): + internal_data_type = data_type[1:-1] + data = [self.serialize_data(d, internal_data_type, **kwargs) if d is not None else "" for d in data] + if not kwargs.get("skip_quote", False): + data = [quote(str(d), safe="") for d in data] + return str(self.serialize_iter(data, internal_data_type, **kwargs)) + + # Not a list, regular serialization + output = self.serialize_data(data, data_type, **kwargs) + if data_type == "bool": + output = json.dumps(output) + if kwargs.get("skip_quote") is True: + output = str(output) + else: + output = quote(str(output), safe="") + except SerializationError: + raise TypeError("{} must be type {}.".format(name, data_type)) + else: + return str(output) + + def header(self, name, data, data_type, **kwargs): + """Serialize data intended for a request header. + + :param data: The data to be serialized. + :param str data_type: The type to be serialized from. + :rtype: str + :raises: TypeError if serialization fails. + :raises: ValueError if data is None + """ + try: + if data_type in ["[str]"]: + data = ["" if d is None else d for d in data] + + output = self.serialize_data(data, data_type, **kwargs) + if data_type == "bool": + output = json.dumps(output) + except SerializationError: + raise TypeError("{} must be type {}.".format(name, data_type)) + else: + return str(output) + + def serialize_data(self, data, data_type, **kwargs): + """Serialize generic data according to supplied data type. + + :param data: The data to be serialized. + :param str data_type: The type to be serialized from. + :param bool required: Whether it's essential that the data not be + empty or None + :raises: AttributeError if required data is None. + :raises: ValueError if data is None + :raises: SerializationError if serialization fails. + """ + if data is None: + raise ValueError("No value for given attribute") + + try: + if data is AzureCoreNull: + return None + if data_type in self.basic_types.values(): + return self.serialize_basic(data, data_type, **kwargs) + + elif data_type in self.serialize_type: + return self.serialize_type[data_type](data, **kwargs) + + # If dependencies is empty, try with current data class + # It has to be a subclass of Enum anyway + enum_type = self.dependencies.get(data_type, data.__class__) + if issubclass(enum_type, Enum): + return Serializer.serialize_enum(data, enum_obj=enum_type) + + iter_type = data_type[0] + data_type[-1] + if iter_type in self.serialize_type: + return self.serialize_type[iter_type](data, data_type[1:-1], **kwargs) + + except (ValueError, TypeError) as err: + msg = "Unable to serialize value: {!r} as type: {!r}." + raise_with_traceback(SerializationError, msg.format(data, data_type), err) + else: + return self._serialize(data, **kwargs) + + @classmethod + def _get_custom_serializers(cls, data_type, **kwargs): + custom_serializer = kwargs.get("basic_types_serializers", {}).get(data_type) + if custom_serializer: + return custom_serializer + if kwargs.get("is_xml", False): + return cls._xml_basic_types_serializers.get(data_type) + + @classmethod + def serialize_basic(cls, data, data_type, **kwargs): + """Serialize basic builting data type. + Serializes objects to str, int, float or bool. + + Possible kwargs: + - basic_types_serializers dict[str, callable] : If set, use the callable as serializer + - is_xml bool : If set, use xml_basic_types_serializers + + :param data: Object to be serialized. + :param str data_type: Type of object in the iterable. + """ + custom_serializer = cls._get_custom_serializers(data_type, **kwargs) + if custom_serializer: + return custom_serializer(data) + if data_type == "str": + return cls.serialize_unicode(data) + return eval(data_type)(data) # nosec + + @classmethod + def serialize_unicode(cls, data): + """Special handling for serializing unicode strings in Py2. + Encode to UTF-8 if unicode, otherwise handle as a str. + + :param data: Object to be serialized. + :rtype: str + """ + try: # If I received an enum, return its value + return data.value + except AttributeError: + pass + + try: + if isinstance(data, unicode): # type: ignore + # Don't change it, JSON and XML ElementTree are totally able + # to serialize correctly u'' strings + return data + except NameError: + return str(data) + else: + return str(data) + + def serialize_iter(self, data, iter_type, div=None, **kwargs): + """Serialize iterable. + + Supported kwargs: + - serialization_ctxt dict : The current entry of _attribute_map, or same format. + serialization_ctxt['type'] should be same as data_type. + - is_xml bool : If set, serialize as XML + + :param list attr: Object to be serialized. + :param str iter_type: Type of object in the iterable. + :param bool required: Whether the objects in the iterable must + not be None or empty. + :param str div: If set, this str will be used to combine the elements + in the iterable into a combined string. Default is 'None'. + :rtype: list, str + """ + if isinstance(data, str): + raise SerializationError("Refuse str type as a valid iter type.") + + serialization_ctxt = kwargs.get("serialization_ctxt", {}) + is_xml = kwargs.get("is_xml", False) + + serialized = [] + for d in data: + try: + serialized.append(self.serialize_data(d, iter_type, **kwargs)) + except ValueError: + serialized.append(None) + + if div: + serialized = ["" if s is None else str(s) for s in serialized] + serialized = div.join(serialized) + + if "xml" in serialization_ctxt or is_xml: + # XML serialization is more complicated + xml_desc = serialization_ctxt.get("xml", {}) + xml_name = xml_desc.get("name") + if not xml_name: + xml_name = serialization_ctxt["key"] + + # Create a wrap node if necessary (use the fact that Element and list have "append") + is_wrapped = xml_desc.get("wrapped", False) + node_name = xml_desc.get("itemsName", xml_name) + if is_wrapped: + final_result = _create_xml_node(xml_name, xml_desc.get("prefix", None), xml_desc.get("ns", None)) + else: + final_result = [] + # All list elements to "local_node" + for el in serialized: + if isinstance(el, ET.Element): + el_node = el + else: + el_node = _create_xml_node(node_name, xml_desc.get("prefix", None), xml_desc.get("ns", None)) + if el is not None: # Otherwise it writes "None" :-p + el_node.text = str(el) + final_result.append(el_node) + return final_result + return serialized + + def serialize_dict(self, attr, dict_type, **kwargs): + """Serialize a dictionary of objects. + + :param dict attr: Object to be serialized. + :param str dict_type: Type of object in the dictionary. + :param bool required: Whether the objects in the dictionary must + not be None or empty. + :rtype: dict + """ + serialization_ctxt = kwargs.get("serialization_ctxt", {}) + serialized = {} + for key, value in attr.items(): + try: + serialized[self.serialize_unicode(key)] = self.serialize_data(value, dict_type, **kwargs) + except ValueError: + serialized[self.serialize_unicode(key)] = None + + if "xml" in serialization_ctxt: + # XML serialization is more complicated + xml_desc = serialization_ctxt["xml"] + xml_name = xml_desc["name"] + + final_result = _create_xml_node(xml_name, xml_desc.get("prefix", None), xml_desc.get("ns", None)) + for key, value in serialized.items(): + ET.SubElement(final_result, key).text = value + return final_result + + return serialized + + def serialize_object(self, attr, **kwargs): + """Serialize a generic object. + This will be handled as a dictionary. If object passed in is not + a basic type (str, int, float, dict, list) it will simply be + cast to str. + + :param dict attr: Object to be serialized. + :rtype: dict or str + """ + if attr is None: + return None + if isinstance(attr, ET.Element): + return attr + obj_type = type(attr) + if obj_type in self.basic_types: + return self.serialize_basic(attr, self.basic_types[obj_type], **kwargs) + if obj_type is _long_type: + return self.serialize_long(attr) + if obj_type is unicode_str: + return self.serialize_unicode(attr) + if obj_type is datetime.datetime: + return self.serialize_iso(attr) + if obj_type is datetime.date: + return self.serialize_date(attr) + if obj_type is datetime.time: + return self.serialize_time(attr) + if obj_type is datetime.timedelta: + return self.serialize_duration(attr) + if obj_type is decimal.Decimal: + return self.serialize_decimal(attr) + + # If it's a model or I know this dependency, serialize as a Model + elif obj_type in self.dependencies.values() or isinstance(attr, Model): + return self._serialize(attr) + + if obj_type == dict: + serialized = {} + for key, value in attr.items(): + try: + serialized[self.serialize_unicode(key)] = self.serialize_object(value, **kwargs) + except ValueError: + serialized[self.serialize_unicode(key)] = None + return serialized + + if obj_type == list: + serialized = [] + for obj in attr: + try: + serialized.append(self.serialize_object(obj, **kwargs)) + except ValueError: + pass + return serialized + return str(attr) + + @staticmethod + def serialize_enum(attr, enum_obj=None): + try: + result = attr.value + except AttributeError: + result = attr + try: + enum_obj(result) # type: ignore + return result + except ValueError: + for enum_value in enum_obj: # type: ignore + if enum_value.value.lower() == str(attr).lower(): + return enum_value.value + error = "{!r} is not valid value for enum {!r}" + raise SerializationError(error.format(attr, enum_obj)) + + @staticmethod + def serialize_bytearray(attr, **kwargs): + """Serialize bytearray into base-64 string. + + :param attr: Object to be serialized. + :rtype: str + """ + return b64encode(attr).decode() + + @staticmethod + def serialize_base64(attr, **kwargs): + """Serialize str into base-64 string. + + :param attr: Object to be serialized. + :rtype: str + """ + encoded = b64encode(attr).decode("ascii") + return encoded.strip("=").replace("+", "-").replace("/", "_") + + @staticmethod + def serialize_decimal(attr, **kwargs): + """Serialize Decimal object to float. + + :param attr: Object to be serialized. + :rtype: float + """ + return float(attr) + + @staticmethod + def serialize_long(attr, **kwargs): + """Serialize long (Py2) or int (Py3). + + :param attr: Object to be serialized. + :rtype: int/long + """ + return _long_type(attr) + + @staticmethod + def serialize_date(attr, **kwargs): + """Serialize Date object into ISO-8601 formatted string. + + :param Date attr: Object to be serialized. + :rtype: str + """ + if isinstance(attr, str): + attr = isodate.parse_date(attr) + t = "{:04}-{:02}-{:02}".format(attr.year, attr.month, attr.day) + return t + + @staticmethod + def serialize_time(attr, **kwargs): + """Serialize Time object into ISO-8601 formatted string. + + :param datetime.time attr: Object to be serialized. + :rtype: str + """ + if isinstance(attr, str): + attr = isodate.parse_time(attr) + t = "{:02}:{:02}:{:02}".format(attr.hour, attr.minute, attr.second) + if attr.microsecond: + t += ".{:02}".format(attr.microsecond) + return t + + @staticmethod + def serialize_duration(attr, **kwargs): + """Serialize TimeDelta object into ISO-8601 formatted string. + + :param TimeDelta attr: Object to be serialized. + :rtype: str + """ + if isinstance(attr, str): + attr = isodate.parse_duration(attr) + return isodate.duration_isoformat(attr) + + @staticmethod + def serialize_rfc(attr, **kwargs): + """Serialize Datetime object into RFC-1123 formatted string. + + :param Datetime attr: Object to be serialized. + :rtype: str + :raises: TypeError if format invalid. + """ + try: + if not attr.tzinfo: + _LOGGER.warning("Datetime with no tzinfo will be considered UTC.") + utc = attr.utctimetuple() + except AttributeError: + raise TypeError("RFC1123 object must be valid Datetime object.") + + return "{}, {:02} {} {:04} {:02}:{:02}:{:02} GMT".format( + Serializer.days[utc.tm_wday], + utc.tm_mday, + Serializer.months[utc.tm_mon], + utc.tm_year, + utc.tm_hour, + utc.tm_min, + utc.tm_sec, + ) + + @staticmethod + def serialize_iso(attr, **kwargs): + """Serialize Datetime object into ISO-8601 formatted string. + + :param Datetime attr: Object to be serialized. + :rtype: str + :raises: SerializationError if format invalid. + """ + if isinstance(attr, str): + attr = isodate.parse_datetime(attr) + try: + if not attr.tzinfo: + _LOGGER.warning("Datetime with no tzinfo will be considered UTC.") + utc = attr.utctimetuple() + if utc.tm_year > 9999 or utc.tm_year < 1: + raise OverflowError("Hit max or min date") + + microseconds = str(attr.microsecond).rjust(6, "0").rstrip("0").ljust(3, "0") + if microseconds: + microseconds = "." + microseconds + date = "{:04}-{:02}-{:02}T{:02}:{:02}:{:02}".format( + utc.tm_year, utc.tm_mon, utc.tm_mday, utc.tm_hour, utc.tm_min, utc.tm_sec + ) + return date + microseconds + "Z" + except (ValueError, OverflowError) as err: + msg = "Unable to serialize datetime object." + raise_with_traceback(SerializationError, msg, err) + except AttributeError as err: + msg = "ISO-8601 object must be valid Datetime object." + raise_with_traceback(TypeError, msg, err) + + @staticmethod + def serialize_unix(attr, **kwargs): + """Serialize Datetime object into IntTime format. + This is represented as seconds. + + :param Datetime attr: Object to be serialized. + :rtype: int + :raises: SerializationError if format invalid + """ + if isinstance(attr, int): + return attr + try: + if not attr.tzinfo: + _LOGGER.warning("Datetime with no tzinfo will be considered UTC.") + return int(calendar.timegm(attr.utctimetuple())) + except AttributeError: + raise TypeError("Unix time object must be valid Datetime object.") + + +def rest_key_extractor(attr, attr_desc, data): + key = attr_desc["key"] + working_data = data + + while "." in key: + # Need the cast, as for some reasons "split" is typed as list[str | Any] + dict_keys = cast(List[str], _FLATTEN.split(key)) + if len(dict_keys) == 1: + key = _decode_attribute_map_key(dict_keys[0]) + break + working_key = _decode_attribute_map_key(dict_keys[0]) + working_data = working_data.get(working_key, data) + if working_data is None: + # If at any point while following flatten JSON path see None, it means + # that all properties under are None as well + # https://github.com/Azure/msrest-for-python/issues/197 + return None + key = ".".join(dict_keys[1:]) + + return working_data.get(key) + + +def rest_key_case_insensitive_extractor(attr, attr_desc, data): + key = attr_desc["key"] + working_data = data + + while "." in key: + dict_keys = _FLATTEN.split(key) + if len(dict_keys) == 1: + key = _decode_attribute_map_key(dict_keys[0]) + break + working_key = _decode_attribute_map_key(dict_keys[0]) + working_data = attribute_key_case_insensitive_extractor(working_key, None, working_data) + if working_data is None: + # If at any point while following flatten JSON path see None, it means + # that all properties under are None as well + # https://github.com/Azure/msrest-for-python/issues/197 + return None + key = ".".join(dict_keys[1:]) + + if working_data: + return attribute_key_case_insensitive_extractor(key, None, working_data) + + +def last_rest_key_extractor(attr, attr_desc, data): + """Extract the attribute in "data" based on the last part of the JSON path key.""" + key = attr_desc["key"] + dict_keys = _FLATTEN.split(key) + return attribute_key_extractor(dict_keys[-1], None, data) + + +def last_rest_key_case_insensitive_extractor(attr, attr_desc, data): + """Extract the attribute in "data" based on the last part of the JSON path key. + + This is the case insensitive version of "last_rest_key_extractor" + """ + key = attr_desc["key"] + dict_keys = _FLATTEN.split(key) + return attribute_key_case_insensitive_extractor(dict_keys[-1], None, data) + + +def attribute_key_extractor(attr, _, data): + return data.get(attr) + + +def attribute_key_case_insensitive_extractor(attr, _, data): + found_key = None + lower_attr = attr.lower() + for key in data: + if lower_attr == key.lower(): + found_key = key + break + + return data.get(found_key) + + +def _extract_name_from_internal_type(internal_type): + """Given an internal type XML description, extract correct XML name with namespace. + + :param dict internal_type: An model type + :rtype: tuple + :returns: A tuple XML name + namespace dict + """ + internal_type_xml_map = getattr(internal_type, "_xml_map", {}) + xml_name = internal_type_xml_map.get("name", internal_type.__name__) + xml_ns = internal_type_xml_map.get("ns", None) + if xml_ns: + xml_name = "{{{}}}{}".format(xml_ns, xml_name) + return xml_name + + +def xml_key_extractor(attr, attr_desc, data): + if isinstance(data, dict): + return None + + # Test if this model is XML ready first + if not isinstance(data, ET.Element): + return None + + xml_desc = attr_desc.get("xml", {}) + xml_name = xml_desc.get("name", attr_desc["key"]) + + # Look for a children + is_iter_type = attr_desc["type"].startswith("[") + is_wrapped = xml_desc.get("wrapped", False) + internal_type = attr_desc.get("internalType", None) + internal_type_xml_map = getattr(internal_type, "_xml_map", {}) + + # Integrate namespace if necessary + xml_ns = xml_desc.get("ns", internal_type_xml_map.get("ns", None)) + if xml_ns: + xml_name = "{{{}}}{}".format(xml_ns, xml_name) + + # If it's an attribute, that's simple + if xml_desc.get("attr", False): + return data.get(xml_name) + + # If it's x-ms-text, that's simple too + if xml_desc.get("text", False): + return data.text + + # Scenario where I take the local name: + # - Wrapped node + # - Internal type is an enum (considered basic types) + # - Internal type has no XML/Name node + if is_wrapped or (internal_type and (issubclass(internal_type, Enum) or "name" not in internal_type_xml_map)): + children = data.findall(xml_name) + # If internal type has a local name and it's not a list, I use that name + elif not is_iter_type and internal_type and "name" in internal_type_xml_map: + xml_name = _extract_name_from_internal_type(internal_type) + children = data.findall(xml_name) + # That's an array + else: + if internal_type: # Complex type, ignore itemsName and use the complex type name + items_name = _extract_name_from_internal_type(internal_type) + else: + items_name = xml_desc.get("itemsName", xml_name) + children = data.findall(items_name) + + if len(children) == 0: + if is_iter_type: + if is_wrapped: + return None # is_wrapped no node, we want None + else: + return [] # not wrapped, assume empty list + return None # Assume it's not there, maybe an optional node. + + # If is_iter_type and not wrapped, return all found children + if is_iter_type: + if not is_wrapped: + return children + else: # Iter and wrapped, should have found one node only (the wrap one) + if len(children) != 1: + raise DeserializationError( + "Tried to deserialize an array not wrapped, and found several nodes '{}'. Maybe you should declare this array as wrapped?".format( + xml_name + ) + ) + return list(children[0]) # Might be empty list and that's ok. + + # Here it's not a itertype, we should have found one element only or empty + if len(children) > 1: + raise DeserializationError("Find several XML '{}' where it was not expected".format(xml_name)) + return children[0] + + +class Deserializer(object): + """Response object model deserializer. + + :param dict classes: Class type dictionary for deserializing complex types. + :ivar list key_extractors: Ordered list of extractors to be used by this deserializer. + """ + + basic_types = {str: "str", int: "int", bool: "bool", float: "float"} + + valid_date = re.compile(r"\d{4}[-]\d{2}[-]\d{2}T\d{2}:\d{2}:\d{2}" r"\.?\d*Z?[-+]?[\d{2}]?:?[\d{2}]?") + + def __init__(self, classes: Optional[Mapping[str, Type[ModelType]]] = None): + self.deserialize_type = { + "iso-8601": Deserializer.deserialize_iso, + "rfc-1123": Deserializer.deserialize_rfc, + "unix-time": Deserializer.deserialize_unix, + "duration": Deserializer.deserialize_duration, + "date": Deserializer.deserialize_date, + "time": Deserializer.deserialize_time, + "decimal": Deserializer.deserialize_decimal, + "long": Deserializer.deserialize_long, + "bytearray": Deserializer.deserialize_bytearray, + "base64": Deserializer.deserialize_base64, + "object": self.deserialize_object, + "[]": self.deserialize_iter, + "{}": self.deserialize_dict, + } + self.deserialize_expected_types = { + "duration": (isodate.Duration, datetime.timedelta), + "iso-8601": (datetime.datetime), + } + self.dependencies: Dict[str, Type[ModelType]] = dict(classes) if classes else {} + self.key_extractors = [rest_key_extractor, xml_key_extractor] + # Additional properties only works if the "rest_key_extractor" is used to + # extract the keys. Making it to work whatever the key extractor is too much + # complicated, with no real scenario for now. + # So adding a flag to disable additional properties detection. This flag should be + # used if your expect the deserialization to NOT come from a JSON REST syntax. + # Otherwise, result are unexpected + self.additional_properties_detection = True + + def __call__(self, target_obj, response_data, content_type=None): + """Call the deserializer to process a REST response. + + :param str target_obj: Target data type to deserialize to. + :param requests.Response response_data: REST response object. + :param str content_type: Swagger "produces" if available. + :raises: DeserializationError if deserialization fails. + :return: Deserialized object. + """ + data = self._unpack_content(response_data, content_type) + return self._deserialize(target_obj, data) + + def _deserialize(self, target_obj, data): + """Call the deserializer on a model. + + Data needs to be already deserialized as JSON or XML ElementTree + + :param str target_obj: Target data type to deserialize to. + :param object data: Object to deserialize. + :raises: DeserializationError if deserialization fails. + :return: Deserialized object. + """ + # This is already a model, go recursive just in case + if hasattr(data, "_attribute_map"): + constants = [name for name, config in getattr(data, "_validation", {}).items() if config.get("constant")] + try: + for attr, mapconfig in data._attribute_map.items(): + if attr in constants: + continue + value = getattr(data, attr) + if value is None: + continue + local_type = mapconfig["type"] + internal_data_type = local_type.strip("[]{}") + if internal_data_type not in self.dependencies or isinstance(internal_data_type, Enum): + continue + setattr(data, attr, self._deserialize(local_type, value)) + return data + except AttributeError: + return + + response, class_name = self._classify_target(target_obj, data) + + if isinstance(response, basestring): + return self.deserialize_data(data, response) + elif isinstance(response, type) and issubclass(response, Enum): + return self.deserialize_enum(data, response) + + if data is None: + return data + try: + attributes = response._attribute_map # type: ignore + d_attrs = {} + for attr, attr_desc in attributes.items(): + # Check empty string. If it's not empty, someone has a real "additionalProperties"... + if attr == "additional_properties" and attr_desc["key"] == "": + continue + raw_value = None + # Enhance attr_desc with some dynamic data + attr_desc = attr_desc.copy() # Do a copy, do not change the real one + internal_data_type = attr_desc["type"].strip("[]{}") + if internal_data_type in self.dependencies: + attr_desc["internalType"] = self.dependencies[internal_data_type] + + for key_extractor in self.key_extractors: + found_value = key_extractor(attr, attr_desc, data) + if found_value is not None: + if raw_value is not None and raw_value != found_value: + msg = ( + "Ignoring extracted value '%s' from %s for key '%s'" + " (duplicate extraction, follow extractors order)" + ) + _LOGGER.warning(msg, found_value, key_extractor, attr) + continue + raw_value = found_value + + value = self.deserialize_data(raw_value, attr_desc["type"]) + d_attrs[attr] = value + except (AttributeError, TypeError, KeyError) as err: + msg = "Unable to deserialize to object: " + class_name # type: ignore + raise_with_traceback(DeserializationError, msg, err) + else: + additional_properties = self._build_additional_properties(attributes, data) + return self._instantiate_model(response, d_attrs, additional_properties) + + def _build_additional_properties(self, attribute_map, data): + if not self.additional_properties_detection: + return None + if "additional_properties" in attribute_map and attribute_map.get("additional_properties", {}).get("key") != "": + # Check empty string. If it's not empty, someone has a real "additionalProperties" + return None + if isinstance(data, ET.Element): + data = {el.tag: el.text for el in data} + + known_keys = { + _decode_attribute_map_key(_FLATTEN.split(desc["key"])[0]) + for desc in attribute_map.values() + if desc["key"] != "" + } + present_keys = set(data.keys()) + missing_keys = present_keys - known_keys + return {key: data[key] for key in missing_keys} + + def _classify_target(self, target, data): + """Check to see whether the deserialization target object can + be classified into a subclass. + Once classification has been determined, initialize object. + + :param str target: The target object type to deserialize to. + :param str/dict data: The response data to deserialize. + """ + if target is None: + return None, None + + if isinstance(target, basestring): + try: + target = self.dependencies[target] + except KeyError: + return target, target + + try: + target = target._classify(data, self.dependencies) + except AttributeError: + pass # Target is not a Model, no classify + return target, target.__class__.__name__ # type: ignore + + def failsafe_deserialize(self, target_obj, data, content_type=None): + """Ignores any errors encountered in deserialization, + and falls back to not deserializing the object. Recommended + for use in error deserialization, as we want to return the + HttpResponseError to users, and not have them deal with + a deserialization error. + + :param str target_obj: The target object type to deserialize to. + :param str/dict data: The response data to deserialize. + :param str content_type: Swagger "produces" if available. + """ + try: + return self(target_obj, data, content_type=content_type) + except: + _LOGGER.debug( + "Ran into a deserialization error. Ignoring since this is failsafe deserialization", exc_info=True + ) + return None + + @staticmethod + def _unpack_content(raw_data, content_type=None): + """Extract the correct structure for deserialization. + + If raw_data is a PipelineResponse, try to extract the result of RawDeserializer. + if we can't, raise. Your Pipeline should have a RawDeserializer. + + If not a pipeline response and raw_data is bytes or string, use content-type + to decode it. If no content-type, try JSON. + + If raw_data is something else, bypass all logic and return it directly. + + :param raw_data: Data to be processed. + :param content_type: How to parse if raw_data is a string/bytes. + :raises JSONDecodeError: If JSON is requested and parsing is impossible. + :raises UnicodeDecodeError: If bytes is not UTF8 + """ + # Assume this is enough to detect a Pipeline Response without importing it + context = getattr(raw_data, "context", {}) + if context: + if RawDeserializer.CONTEXT_NAME in context: + return context[RawDeserializer.CONTEXT_NAME] + raise ValueError("This pipeline didn't have the RawDeserializer policy; can't deserialize") + + # Assume this is enough to recognize universal_http.ClientResponse without importing it + if hasattr(raw_data, "body"): + return RawDeserializer.deserialize_from_http_generics(raw_data.text(), raw_data.headers) + + # Assume this enough to recognize requests.Response without importing it. + if hasattr(raw_data, "_content_consumed"): + return RawDeserializer.deserialize_from_http_generics(raw_data.text, raw_data.headers) + + if isinstance(raw_data, (basestring, bytes)) or hasattr(raw_data, "read"): + return RawDeserializer.deserialize_from_text(raw_data, content_type) # type: ignore + return raw_data + + def _instantiate_model(self, response, attrs, additional_properties=None): + """Instantiate a response model passing in deserialized args. + + :param response: The response model class. + :param d_attrs: The deserialized response attributes. + """ + if callable(response): + subtype = getattr(response, "_subtype_map", {}) + try: + readonly = [k for k, v in response._validation.items() if v.get("readonly")] + const = [k for k, v in response._validation.items() if v.get("constant")] + kwargs = {k: v for k, v in attrs.items() if k not in subtype and k not in readonly + const} + response_obj = response(**kwargs) + for attr in readonly: + setattr(response_obj, attr, attrs.get(attr)) + if additional_properties: + response_obj.additional_properties = additional_properties + return response_obj + except TypeError as err: + msg = "Unable to deserialize {} into model {}. ".format(kwargs, response) # type: ignore + raise DeserializationError(msg + str(err)) + else: + try: + for attr, value in attrs.items(): + setattr(response, attr, value) + return response + except Exception as exp: + msg = "Unable to populate response model. " + msg += "Type: {}, Error: {}".format(type(response), exp) + raise DeserializationError(msg) + + def deserialize_data(self, data, data_type): + """Process data for deserialization according to data type. + + :param str data: The response string to be deserialized. + :param str data_type: The type to deserialize to. + :raises: DeserializationError if deserialization fails. + :return: Deserialized object. + """ + if data is None: + return data + + try: + if not data_type: + return data + if data_type in self.basic_types.values(): + return self.deserialize_basic(data, data_type) + if data_type in self.deserialize_type: + if isinstance(data, self.deserialize_expected_types.get(data_type, tuple())): + return data + + is_a_text_parsing_type = lambda x: x not in ["object", "[]", r"{}"] + if isinstance(data, ET.Element) and is_a_text_parsing_type(data_type) and not data.text: + return None + data_val = self.deserialize_type[data_type](data) + return data_val + + iter_type = data_type[0] + data_type[-1] + if iter_type in self.deserialize_type: + return self.deserialize_type[iter_type](data, data_type[1:-1]) + + obj_type = self.dependencies[data_type] + if issubclass(obj_type, Enum): + if isinstance(data, ET.Element): + data = data.text + return self.deserialize_enum(data, obj_type) + + except (ValueError, TypeError, AttributeError) as err: + msg = "Unable to deserialize response data." + msg += " Data: {}, {}".format(data, data_type) + raise_with_traceback(DeserializationError, msg, err) + else: + return self._deserialize(obj_type, data) + + def deserialize_iter(self, attr, iter_type): + """Deserialize an iterable. + + :param list attr: Iterable to be deserialized. + :param str iter_type: The type of object in the iterable. + :rtype: list + """ + if attr is None: + return None + if isinstance(attr, ET.Element): # If I receive an element here, get the children + attr = list(attr) + if not isinstance(attr, (list, set)): + raise DeserializationError("Cannot deserialize as [{}] an object of type {}".format(iter_type, type(attr))) + return [self.deserialize_data(a, iter_type) for a in attr] + + def deserialize_dict(self, attr, dict_type): + """Deserialize a dictionary. + + :param dict/list attr: Dictionary to be deserialized. Also accepts + a list of key, value pairs. + :param str dict_type: The object type of the items in the dictionary. + :rtype: dict + """ + if isinstance(attr, list): + return {x["key"]: self.deserialize_data(x["value"], dict_type) for x in attr} + + if isinstance(attr, ET.Element): + # Transform value into {"Key": "value"} + attr = {el.tag: el.text for el in attr} + return {k: self.deserialize_data(v, dict_type) for k, v in attr.items()} + + def deserialize_object(self, attr, **kwargs): + """Deserialize a generic object. + This will be handled as a dictionary. + + :param dict attr: Dictionary to be deserialized. + :rtype: dict + :raises: TypeError if non-builtin datatype encountered. + """ + if attr is None: + return None + if isinstance(attr, ET.Element): + # Do no recurse on XML, just return the tree as-is + return attr + if isinstance(attr, basestring): + return self.deserialize_basic(attr, "str") + obj_type = type(attr) + if obj_type in self.basic_types: + return self.deserialize_basic(attr, self.basic_types[obj_type]) + if obj_type is _long_type: + return self.deserialize_long(attr) + + if obj_type == dict: + deserialized = {} + for key, value in attr.items(): + try: + deserialized[key] = self.deserialize_object(value, **kwargs) + except ValueError: + deserialized[key] = None + return deserialized + + if obj_type == list: + deserialized = [] + for obj in attr: + try: + deserialized.append(self.deserialize_object(obj, **kwargs)) + except ValueError: + pass + return deserialized + + else: + error = "Cannot deserialize generic object with type: " + raise TypeError(error + str(obj_type)) + + def deserialize_basic(self, attr, data_type): + """Deserialize basic builtin data type from string. + Will attempt to convert to str, int, float and bool. + This function will also accept '1', '0', 'true' and 'false' as + valid bool values. + + :param str attr: response string to be deserialized. + :param str data_type: deserialization data type. + :rtype: str, int, float or bool + :raises: TypeError if string format is not valid. + """ + # If we're here, data is supposed to be a basic type. + # If it's still an XML node, take the text + if isinstance(attr, ET.Element): + attr = attr.text + if not attr: + if data_type == "str": + # None or '', node is empty string. + return "" + else: + # None or '', node with a strong type is None. + # Don't try to model "empty bool" or "empty int" + return None + + if data_type == "bool": + if attr in [True, False, 1, 0]: + return bool(attr) + elif isinstance(attr, basestring): + if attr.lower() in ["true", "1"]: + return True + elif attr.lower() in ["false", "0"]: + return False + raise TypeError("Invalid boolean value: {}".format(attr)) + + if data_type == "str": + return self.deserialize_unicode(attr) + return eval(data_type)(attr) # nosec + + @staticmethod + def deserialize_unicode(data): + """Preserve unicode objects in Python 2, otherwise return data + as a string. + + :param str data: response string to be deserialized. + :rtype: str or unicode + """ + # We might be here because we have an enum modeled as string, + # and we try to deserialize a partial dict with enum inside + if isinstance(data, Enum): + return data + + # Consider this is real string + try: + if isinstance(data, unicode): # type: ignore + return data + except NameError: + return str(data) + else: + return str(data) + + @staticmethod + def deserialize_enum(data, enum_obj): + """Deserialize string into enum object. + + If the string is not a valid enum value it will be returned as-is + and a warning will be logged. + + :param str data: Response string to be deserialized. If this value is + None or invalid it will be returned as-is. + :param Enum enum_obj: Enum object to deserialize to. + :rtype: Enum + """ + if isinstance(data, enum_obj) or data is None: + return data + if isinstance(data, Enum): + data = data.value + if isinstance(data, int): + # Workaround. We might consider remove it in the future. + # https://github.com/Azure/azure-rest-api-specs/issues/141 + try: + return list(enum_obj.__members__.values())[data] + except IndexError: + error = "{!r} is not a valid index for enum {!r}" + raise DeserializationError(error.format(data, enum_obj)) + try: + return enum_obj(str(data)) + except ValueError: + for enum_value in enum_obj: + if enum_value.value.lower() == str(data).lower(): + return enum_value + # We don't fail anymore for unknown value, we deserialize as a string + _LOGGER.warning("Deserializer is not able to find %s as valid enum in %s", data, enum_obj) + return Deserializer.deserialize_unicode(data) + + @staticmethod + def deserialize_bytearray(attr): + """Deserialize string into bytearray. + + :param str attr: response string to be deserialized. + :rtype: bytearray + :raises: TypeError if string format invalid. + """ + if isinstance(attr, ET.Element): + attr = attr.text + return bytearray(b64decode(attr)) # type: ignore + + @staticmethod + def deserialize_base64(attr): + """Deserialize base64 encoded string into string. + + :param str attr: response string to be deserialized. + :rtype: bytearray + :raises: TypeError if string format invalid. + """ + if isinstance(attr, ET.Element): + attr = attr.text + padding = "=" * (3 - (len(attr) + 3) % 4) # type: ignore + attr = attr + padding # type: ignore + encoded = attr.replace("-", "+").replace("_", "/") + return b64decode(encoded) + + @staticmethod + def deserialize_decimal(attr): + """Deserialize string into Decimal object. + + :param str attr: response string to be deserialized. + :rtype: Decimal + :raises: DeserializationError if string format invalid. + """ + if isinstance(attr, ET.Element): + attr = attr.text + try: + return decimal.Decimal(attr) # type: ignore + except decimal.DecimalException as err: + msg = "Invalid decimal {}".format(attr) + raise_with_traceback(DeserializationError, msg, err) + + @staticmethod + def deserialize_long(attr): + """Deserialize string into long (Py2) or int (Py3). + + :param str attr: response string to be deserialized. + :rtype: long or int + :raises: ValueError if string format invalid. + """ + if isinstance(attr, ET.Element): + attr = attr.text + return _long_type(attr) # type: ignore + + @staticmethod + def deserialize_duration(attr): + """Deserialize ISO-8601 formatted string into TimeDelta object. + + :param str attr: response string to be deserialized. + :rtype: TimeDelta + :raises: DeserializationError if string format invalid. + """ + if isinstance(attr, ET.Element): + attr = attr.text + try: + duration = isodate.parse_duration(attr) + except (ValueError, OverflowError, AttributeError) as err: + msg = "Cannot deserialize duration object." + raise_with_traceback(DeserializationError, msg, err) + else: + return duration + + @staticmethod + def deserialize_date(attr): + """Deserialize ISO-8601 formatted string into Date object. + + :param str attr: response string to be deserialized. + :rtype: Date + :raises: DeserializationError if string format invalid. + """ + if isinstance(attr, ET.Element): + attr = attr.text + if re.search(r"[^\W\d_]", attr, re.I + re.U): # type: ignore + raise DeserializationError("Date must have only digits and -. Received: %s" % attr) + # This must NOT use defaultmonth/defaultday. Using None ensure this raises an exception. + return isodate.parse_date(attr, defaultmonth=None, defaultday=None) + + @staticmethod + def deserialize_time(attr): + """Deserialize ISO-8601 formatted string into time object. + + :param str attr: response string to be deserialized. + :rtype: datetime.time + :raises: DeserializationError if string format invalid. + """ + if isinstance(attr, ET.Element): + attr = attr.text + if re.search(r"[^\W\d_]", attr, re.I + re.U): # type: ignore + raise DeserializationError("Date must have only digits and -. Received: %s" % attr) + return isodate.parse_time(attr) + + @staticmethod + def deserialize_rfc(attr): + """Deserialize RFC-1123 formatted string into Datetime object. + + :param str attr: response string to be deserialized. + :rtype: Datetime + :raises: DeserializationError if string format invalid. + """ + if isinstance(attr, ET.Element): + attr = attr.text + try: + parsed_date = email.utils.parsedate_tz(attr) # type: ignore + date_obj = datetime.datetime( + *parsed_date[:6], tzinfo=_FixedOffset(datetime.timedelta(minutes=(parsed_date[9] or 0) / 60)) + ) + if not date_obj.tzinfo: + date_obj = date_obj.astimezone(tz=TZ_UTC) + except ValueError as err: + msg = "Cannot deserialize to rfc datetime object." + raise_with_traceback(DeserializationError, msg, err) + else: + return date_obj + + @staticmethod + def deserialize_iso(attr): + """Deserialize ISO-8601 formatted string into Datetime object. + + :param str attr: response string to be deserialized. + :rtype: Datetime + :raises: DeserializationError if string format invalid. + """ + if isinstance(attr, ET.Element): + attr = attr.text + try: + attr = attr.upper() # type: ignore + match = Deserializer.valid_date.match(attr) + if not match: + raise ValueError("Invalid datetime string: " + attr) + + check_decimal = attr.split(".") + if len(check_decimal) > 1: + decimal_str = "" + for digit in check_decimal[1]: + if digit.isdigit(): + decimal_str += digit + else: + break + if len(decimal_str) > 6: + attr = attr.replace(decimal_str, decimal_str[0:6]) + + date_obj = isodate.parse_datetime(attr) + test_utc = date_obj.utctimetuple() + if test_utc.tm_year > 9999 or test_utc.tm_year < 1: + raise OverflowError("Hit max or min date") + except (ValueError, OverflowError, AttributeError) as err: + msg = "Cannot deserialize datetime object." + raise_with_traceback(DeserializationError, msg, err) + else: + return date_obj + + @staticmethod + def deserialize_unix(attr): + """Serialize Datetime object into IntTime format. + This is represented as seconds. + + :param int attr: Object to be serialized. + :rtype: Datetime + :raises: DeserializationError if format invalid + """ + if isinstance(attr, ET.Element): + attr = int(attr.text) # type: ignore + try: + date_obj = datetime.datetime.fromtimestamp(attr, TZ_UTC) + except ValueError as err: + msg = "Cannot deserialize to unix datetime object." + raise_with_traceback(DeserializationError, msg, err) + else: + return date_obj diff --git a/sdk/eventgrid/azure-eventgrid/azure/eventgrid/_vendor.py b/sdk/eventgrid/azure-eventgrid/azure/eventgrid/_vendor.py new file mode 100644 index 000000000000..e0aa301abac7 --- /dev/null +++ b/sdk/eventgrid/azure-eventgrid/azure/eventgrid/_vendor.py @@ -0,0 +1,38 @@ +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for license information. +# Code generated by Microsoft (R) Python Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is regenerated. +# -------------------------------------------------------------------------- + +from abc import ABC +from typing import List, TYPE_CHECKING, cast + +from ._configuration import EventGridClientConfiguration + +if TYPE_CHECKING: + # pylint: disable=unused-import,ungrouped-imports + from azure.core import PipelineClient + + from ._serialization import Deserializer, Serializer + + +def _format_url_section(template, **kwargs): + components = template.split("/") + while components: + try: + return template.format(**kwargs) + except KeyError as key: + # Need the cast, as for some reasons "split" is typed as list[str | Any] + formatted_components = cast(List[str], template.split("/")) + components = [c for c in formatted_components if "{}".format(key.args[0]) not in c] + template = "/".join(components) + + +class EventGridClientMixinABC(ABC): + """DO NOT use this class. It is for internal typing use only.""" + + _client: "PipelineClient" + _config: EventGridClientConfiguration + _serialize: "Serializer" + _deserialize: "Deserializer" diff --git a/sdk/eventgrid/azure-eventgrid/azure/eventgrid/_version.py b/sdk/eventgrid/azure-eventgrid/azure/eventgrid/_version.py index 3b6361b229a2..c8302cef558b 100644 --- a/sdk/eventgrid/azure-eventgrid/azure/eventgrid/_version.py +++ b/sdk/eventgrid/azure-eventgrid/azure/eventgrid/_version.py @@ -1,12 +1,9 @@ # coding=utf-8 # -------------------------------------------------------------------------- # Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for -# license information. -# -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is -# regenerated. +# Licensed under the MIT License. See License.txt in the project root for license information. +# Code generated by Microsoft (R) Python Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is regenerated. # -------------------------------------------------------------------------- -VERSION = "4.11.1" +VERSION = "4.12.0b1" diff --git a/sdk/eventgrid/azure-eventgrid/azure/eventgrid/aio/__init__.py b/sdk/eventgrid/azure-eventgrid/azure/eventgrid/aio/__init__.py index 0d2dce7aaea2..5a6ddd0f6e2b 100644 --- a/sdk/eventgrid/azure-eventgrid/azure/eventgrid/aio/__init__.py +++ b/sdk/eventgrid/azure-eventgrid/azure/eventgrid/aio/__init__.py @@ -1,9 +1,23 @@ # coding=utf-8 -# ------------------------------------ -# Copyright (c) Microsoft Corporation. -# Licensed under the MIT License. -# ------------------------------------ +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for license information. +# Code generated by Microsoft (R) Python Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is regenerated. +# -------------------------------------------------------------------------- -from ._publisher_client_async import EventGridPublisherClient +from ._client import EventGridClient -__all__ = ["EventGridPublisherClient"] +try: + from ._patch import __all__ as _patch_all + from ._patch import * # pylint: disable=unused-wildcard-import +except ImportError: + _patch_all = [] +from ._patch import patch_sdk as _patch_sdk + +__all__ = [ + "EventGridClient", +] +__all__.extend([p for p in _patch_all if p not in __all__]) + +_patch_sdk() diff --git a/sdk/eventgrid/azure-eventgrid/azure/eventgrid/aio/_client.py b/sdk/eventgrid/azure-eventgrid/azure/eventgrid/aio/_client.py new file mode 100644 index 000000000000..013ea7cf5291 --- /dev/null +++ b/sdk/eventgrid/azure-eventgrid/azure/eventgrid/aio/_client.py @@ -0,0 +1,78 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for license information. +# Code generated by Microsoft (R) Python Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is regenerated. +# -------------------------------------------------------------------------- + +from copy import deepcopy +from typing import Any, Awaitable + +from azure.core import AsyncPipelineClient +from azure.core.credentials import AzureKeyCredential +from azure.core.rest import AsyncHttpResponse, HttpRequest + +from .._serialization import Deserializer, Serializer +from ._configuration import EventGridClientConfiguration +from ._operations import EventGridClientOperationsMixin + + +class EventGridClient(EventGridClientOperationsMixin): # pylint: disable=client-accepts-api-version-keyword + """Azure Messaging EventGrid Client. + + :param endpoint: The host name of the namespace, e.g. + namespaceName1.westus-1.eventgrid.azure.net. Required. + :type endpoint: str + :param credential: Credential needed for the client to connect to Azure. Required. + :type credential: ~azure.core.credentials.AzureKeyCredential + :keyword api_version: The API version to use for this operation. Default value is + "2023-06-01-preview". Note that overriding this default value may result in unsupported + behavior. + :paramtype api_version: str + """ + + def __init__(self, endpoint: str, credential: AzureKeyCredential, **kwargs: Any) -> None: + _endpoint = "{endpoint}" + self._config = EventGridClientConfiguration(endpoint=endpoint, credential=credential, **kwargs) + self._client: AsyncPipelineClient = AsyncPipelineClient(base_url=_endpoint, config=self._config, **kwargs) + + self._serialize = Serializer() + self._deserialize = Deserializer() + self._serialize.client_side_validation = False + + def send_request(self, request: HttpRequest, **kwargs: Any) -> Awaitable[AsyncHttpResponse]: + """Runs the network request through the client's chained policies. + + >>> from azure.core.rest import HttpRequest + >>> request = HttpRequest("GET", "https://www.example.org/") + + >>> response = await client.send_request(request) + + + For more information on this code flow, see https://aka.ms/azsdk/dpcodegen/python/send_request + + :param request: The network request you want to make. Required. + :type request: ~azure.core.rest.HttpRequest + :keyword bool stream: Whether the response payload will be streamed. Defaults to False. + :return: The response of your network call. Does not do error handling on your response. + :rtype: ~azure.core.rest.AsyncHttpResponse + """ + + request_copy = deepcopy(request) + path_format_arguments = { + "endpoint": self._serialize.url("self._config.endpoint", self._config.endpoint, "str", skip_quote=True), + } + + request_copy.url = self._client.format_url(request_copy.url, **path_format_arguments) + return self._client.send_request(request_copy, **kwargs) + + async def close(self) -> None: + await self._client.close() + + async def __aenter__(self) -> "EventGridClient": + await self._client.__aenter__() + return self + + async def __aexit__(self, *exc_details: Any) -> None: + await self._client.__aexit__(*exc_details) diff --git a/sdk/eventgrid/azure-eventgrid/azure/eventgrid/aio/_configuration.py b/sdk/eventgrid/azure-eventgrid/azure/eventgrid/aio/_configuration.py new file mode 100644 index 000000000000..df820835aa70 --- /dev/null +++ b/sdk/eventgrid/azure-eventgrid/azure/eventgrid/aio/_configuration.py @@ -0,0 +1,61 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for license information. +# Code generated by Microsoft (R) Python Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is regenerated. +# -------------------------------------------------------------------------- + +from typing import Any + +from azure.core.configuration import Configuration +from azure.core.credentials import AzureKeyCredential +from azure.core.pipeline import policies + +from .._version import VERSION + + +class EventGridClientConfiguration(Configuration): # pylint: disable=too-many-instance-attributes + """Configuration for EventGridClient. + + Note that all parameters used to create this instance are saved as instance + attributes. + + :param endpoint: The host name of the namespace, e.g. + namespaceName1.westus-1.eventgrid.azure.net. Required. + :type endpoint: str + :param credential: Credential needed for the client to connect to Azure. Required. + :type credential: ~azure.core.credentials.AzureKeyCredential + :keyword api_version: The API version to use for this operation. Default value is + "2023-06-01-preview". Note that overriding this default value may result in unsupported + behavior. + :paramtype api_version: str + """ + + def __init__(self, endpoint: str, credential: AzureKeyCredential, **kwargs: Any) -> None: + super(EventGridClientConfiguration, self).__init__(**kwargs) + api_version: str = kwargs.pop("api_version", "2023-06-01-preview") + + if endpoint is None: + raise ValueError("Parameter 'endpoint' must not be None.") + if credential is None: + raise ValueError("Parameter 'credential' must not be None.") + + self.endpoint = endpoint + self.credential = credential + self.api_version = api_version + kwargs.setdefault("sdk_moniker", "eventgrid/{}".format(VERSION)) + self._configure(**kwargs) + + def _configure(self, **kwargs: Any) -> None: + self.user_agent_policy = kwargs.get("user_agent_policy") or policies.UserAgentPolicy(**kwargs) + self.headers_policy = kwargs.get("headers_policy") or policies.HeadersPolicy(**kwargs) + self.proxy_policy = kwargs.get("proxy_policy") or policies.ProxyPolicy(**kwargs) + self.logging_policy = kwargs.get("logging_policy") or policies.NetworkTraceLoggingPolicy(**kwargs) + self.http_logging_policy = kwargs.get("http_logging_policy") or policies.HttpLoggingPolicy(**kwargs) + self.retry_policy = kwargs.get("retry_policy") or policies.AsyncRetryPolicy(**kwargs) + self.custom_hook_policy = kwargs.get("custom_hook_policy") or policies.CustomHookPolicy(**kwargs) + self.redirect_policy = kwargs.get("redirect_policy") or policies.AsyncRedirectPolicy(**kwargs) + self.authentication_policy = kwargs.get("authentication_policy") + if self.credential and not self.authentication_policy: + self.authentication_policy = policies.AzureKeyCredentialPolicy(self.credential, "SharedAccessKey", **kwargs) diff --git a/sdk/eventgrid/azure-eventgrid/azure/eventgrid/aio/_operations/__init__.py b/sdk/eventgrid/azure-eventgrid/azure/eventgrid/aio/_operations/__init__.py new file mode 100644 index 000000000000..51bf5a7f3dcf --- /dev/null +++ b/sdk/eventgrid/azure-eventgrid/azure/eventgrid/aio/_operations/__init__.py @@ -0,0 +1,19 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for license information. +# Code generated by Microsoft (R) Python Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is regenerated. +# -------------------------------------------------------------------------- + +from ._operations import EventGridClientOperationsMixin + +from ._patch import __all__ as _patch_all +from ._patch import * # pylint: disable=unused-wildcard-import +from ._patch import patch_sdk as _patch_sdk + +__all__ = [ + "EventGridClientOperationsMixin", +] +__all__.extend([p for p in _patch_all if p not in __all__]) +_patch_sdk() diff --git a/sdk/eventgrid/azure-eventgrid/azure/eventgrid/aio/_operations/_operations.py b/sdk/eventgrid/azure-eventgrid/azure/eventgrid/aio/_operations/_operations.py new file mode 100644 index 000000000000..6deb5f1616bc --- /dev/null +++ b/sdk/eventgrid/azure-eventgrid/azure/eventgrid/aio/_operations/_operations.py @@ -0,0 +1,475 @@ +# pylint: disable=too-many-lines +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for license information. +# Code generated by Microsoft (R) Python Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is regenerated. +# -------------------------------------------------------------------------- +import datetime +import json +from typing import Any, Callable, Dict, List, Optional, TypeVar + +from azure.core.exceptions import ( + ClientAuthenticationError, + HttpResponseError, + ResourceExistsError, + ResourceNotFoundError, + ResourceNotModifiedError, + map_error, +) +from azure.core.pipeline import PipelineResponse +from azure.core.pipeline.transport import AsyncHttpResponse +from azure.core.rest import HttpRequest +from azure.core.tracing.decorator_async import distributed_trace_async +from azure.core.utils import case_insensitive_dict + +from ... import models as _models +from ..._model_base import AzureJSONEncoder, _deserialize +from ..._operations._operations import ( + build_event_grid_acknowledge_cloud_events_request, + build_event_grid_publish_cloud_event_request, + build_event_grid_publish_cloud_events_request, + build_event_grid_receive_cloud_events_request, + build_event_grid_reject_cloud_events_request, + build_event_grid_release_cloud_events_request, +) +from .._vendor import EventGridClientMixinABC + +T = TypeVar("T") +ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, Dict[str, Any]], Any]] + + +class EventGridClientOperationsMixin(EventGridClientMixinABC): + @distributed_trace_async + async def publish_cloud_event( # pylint: disable=inconsistent-return-statements + self, topic_name: str, event: _models.CloudEvent, **kwargs: Any + ) -> None: + """Publish Single Cloud Event to namespace topic. In case of success, the server responds with an + HTTP 200 status code with an empty JSON object in response. Otherwise, the server can return + various error codes. For example, 401: which indicates authorization failure, 403: which + indicates quota exceeded or message is too large, 410: which indicates that specific topic is + not found, 400: for bad request, and 500: for internal server error. + + :param topic_name: Topic Name. Required. + :type topic_name: str + :param event: Single Cloud Event being published. Required. + :type event: ~azure.eventgrid.models.CloudEvent + :keyword content_type: content type. Default value is "application/cloudevents+json; + charset=utf-8". + :paramtype content_type: str + :keyword bool stream: Whether to stream the response of this operation. Defaults to False. You + will have to context manage the returned stream. + :return: None + :rtype: None + :raises ~azure.core.exceptions.HttpResponseError: + """ + error_map = { + 401: ClientAuthenticationError, + 404: ResourceNotFoundError, + 409: ResourceExistsError, + 304: ResourceNotModifiedError, + } + error_map.update(kwargs.pop("error_map", {}) or {}) + + _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) + _params = kwargs.pop("params", {}) or {} + + content_type: str = kwargs.pop( + "content_type", _headers.pop("content-type", "application/cloudevents+json; charset=utf-8") + ) + cls: ClsType[None] = kwargs.pop("cls", None) + + _content = json.dumps(event, cls=AzureJSONEncoder) # type: ignore + + request = build_event_grid_publish_cloud_event_request( + topic_name=topic_name, + content_type=content_type, + api_version=self._config.api_version, + content=_content, + headers=_headers, + params=_params, + ) + path_format_arguments = { + "endpoint": self._serialize.url("self._config.endpoint", self._config.endpoint, "str", skip_quote=True), + } + request.url = self._client.format_url(request.url, **path_format_arguments) + + _stream = kwargs.pop("stream", False) + pipeline_response: PipelineResponse = await self._client._pipeline.run( # type: ignore # pylint: disable=protected-access + request, stream=_stream, **kwargs + ) + + response = pipeline_response.http_response + + if response.status_code not in [200]: + map_error(status_code=response.status_code, response=response, error_map=error_map) + raise HttpResponseError(response=response) + + if cls: + return cls(pipeline_response, None, {}) + + @distributed_trace_async + async def publish_cloud_events( # pylint: disable=inconsistent-return-statements + self, topic_name: str, events: List[_models.CloudEvent], **kwargs: Any + ) -> None: + """Publish Batch Cloud Event to namespace topic. In case of success, the server responds with an + HTTP 200 status code with an empty JSON object in response. Otherwise, the server can return + various error codes. For example, 401: which indicates authorization failure, 403: which + indicates quota exceeded or message is too large, 410: which indicates that specific topic is + not found, 400: for bad request, and 500: for internal server error. + + :param topic_name: Topic Name. Required. + :type topic_name: str + :param events: Array of Cloud Events being published. Required. + :type events: list[~azure.eventgrid.models.CloudEvent] + :keyword content_type: content type. Default value is "application/cloudevents-batch+json; + charset=utf-8". + :paramtype content_type: str + :keyword bool stream: Whether to stream the response of this operation. Defaults to False. You + will have to context manage the returned stream. + :return: None + :rtype: None + :raises ~azure.core.exceptions.HttpResponseError: + """ + error_map = { + 401: ClientAuthenticationError, + 404: ResourceNotFoundError, + 409: ResourceExistsError, + 304: ResourceNotModifiedError, + } + error_map.update(kwargs.pop("error_map", {}) or {}) + + _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) + _params = kwargs.pop("params", {}) or {} + + content_type: str = kwargs.pop( + "content_type", _headers.pop("content-type", "application/cloudevents-batch+json; charset=utf-8") + ) + cls: ClsType[None] = kwargs.pop("cls", None) + + _content = json.dumps(events, cls=AzureJSONEncoder) # type: ignore + + request = build_event_grid_publish_cloud_events_request( + topic_name=topic_name, + content_type=content_type, + api_version=self._config.api_version, + content=_content, + headers=_headers, + params=_params, + ) + path_format_arguments = { + "endpoint": self._serialize.url("self._config.endpoint", self._config.endpoint, "str", skip_quote=True), + } + request.url = self._client.format_url(request.url, **path_format_arguments) + + _stream = kwargs.pop("stream", False) + pipeline_response: PipelineResponse = await self._client._pipeline.run( # type: ignore # pylint: disable=protected-access + request, stream=_stream, **kwargs + ) + + response = pipeline_response.http_response + + if response.status_code not in [200]: + map_error(status_code=response.status_code, response=response, error_map=error_map) + raise HttpResponseError(response=response) + + if cls: + return cls(pipeline_response, None, {}) + + @distributed_trace_async + async def receive_cloud_events( + self, + topic_name: str, + event_subscription_name: str, + *, + max_events: Optional[int] = None, + max_wait_time: Optional[datetime.timedelta] = None, + **kwargs: Any + ) -> _models.ReceiveResult: + """Receive Batch of Cloud Events from the Event Subscription. + + :param topic_name: Topic Name. Required. + :type topic_name: str + :param event_subscription_name: Event Subscription Name. Required. + :type event_subscription_name: str + :keyword max_events: Max Events count to be received. Minimum value is 1, while maximum value + is 100 events. If not specified, the default value is 1. Default value is None. + :paramtype max_events: int + :keyword max_wait_time: Max wait time value for receive operation in Seconds. It is the time in + seconds that the server approximately waits for the availability of an event and responds to + the request. If an event is available, the broker responds immediately to the client. Minimum + value is 10 seconds, while maximum value is 120 seconds. If not specified, the default value is + 60 seconds. Default value is None. + :paramtype max_wait_time: ~datetime.timedelta + :keyword bool stream: Whether to stream the response of this operation. Defaults to False. You + will have to context manage the returned stream. + :return: ReceiveResult. The ReceiveResult is compatible with MutableMapping + :rtype: ~azure.eventgrid.models.ReceiveResult + :raises ~azure.core.exceptions.HttpResponseError: + """ + error_map = { + 401: ClientAuthenticationError, + 404: ResourceNotFoundError, + 409: ResourceExistsError, + 304: ResourceNotModifiedError, + } + error_map.update(kwargs.pop("error_map", {}) or {}) + + _headers = kwargs.pop("headers", {}) or {} + _params = kwargs.pop("params", {}) or {} + + cls: ClsType[_models.ReceiveResult] = kwargs.pop("cls", None) + + request = build_event_grid_receive_cloud_events_request( + topic_name=topic_name, + event_subscription_name=event_subscription_name, + max_events=max_events, + max_wait_time=max_wait_time, + api_version=self._config.api_version, + headers=_headers, + params=_params, + ) + path_format_arguments = { + "endpoint": self._serialize.url("self._config.endpoint", self._config.endpoint, "str", skip_quote=True), + } + request.url = self._client.format_url(request.url, **path_format_arguments) + + _stream = kwargs.pop("stream", False) + pipeline_response: PipelineResponse = await self._client._pipeline.run( # type: ignore # pylint: disable=protected-access + request, stream=_stream, **kwargs + ) + + response = pipeline_response.http_response + + if response.status_code not in [200]: + map_error(status_code=response.status_code, response=response, error_map=error_map) + raise HttpResponseError(response=response) + + if _stream: + deserialized = response.iter_bytes() + else: + deserialized = _deserialize(_models.ReceiveResult, response.json()) + + if cls: + return cls(pipeline_response, deserialized, {}) # type: ignore + + return deserialized # type: ignore + + @distributed_trace_async + async def acknowledge_cloud_events( + self, topic_name: str, event_subscription_name: str, lock_tokens: _models.AcknowledgeOptions, **kwargs: Any + ) -> _models.AcknowledgeResult: + """Acknowledge batch of Cloud Events. The server responds with an HTTP 200 status code if at least + one event is successfully acknowledged. The response body will include the set of successfully + acknowledged lockTokens, along with other failed lockTokens with their corresponding error + information. Successfully acknowledged events will no longer be available to any consumer. + + :param topic_name: Topic Name. Required. + :type topic_name: str + :param event_subscription_name: Event Subscription Name. Required. + :type event_subscription_name: str + :param lock_tokens: AcknowledgeOptions. Required. + :type lock_tokens: ~azure.eventgrid.models.AcknowledgeOptions + :keyword content_type: content type. Default value is "application/json; charset=utf-8". + :paramtype content_type: str + :keyword bool stream: Whether to stream the response of this operation. Defaults to False. You + will have to context manage the returned stream. + :return: AcknowledgeResult. The AcknowledgeResult is compatible with MutableMapping + :rtype: ~azure.eventgrid.models.AcknowledgeResult + :raises ~azure.core.exceptions.HttpResponseError: + """ + error_map = { + 401: ClientAuthenticationError, + 404: ResourceNotFoundError, + 409: ResourceExistsError, + 304: ResourceNotModifiedError, + } + error_map.update(kwargs.pop("error_map", {}) or {}) + + _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) + _params = kwargs.pop("params", {}) or {} + + content_type: str = kwargs.pop("content_type", _headers.pop("content-type", "application/json; charset=utf-8")) + cls: ClsType[_models.AcknowledgeResult] = kwargs.pop("cls", None) + + _content = json.dumps(lock_tokens, cls=AzureJSONEncoder) # type: ignore + + request = build_event_grid_acknowledge_cloud_events_request( + topic_name=topic_name, + event_subscription_name=event_subscription_name, + content_type=content_type, + api_version=self._config.api_version, + content=_content, + headers=_headers, + params=_params, + ) + path_format_arguments = { + "endpoint": self._serialize.url("self._config.endpoint", self._config.endpoint, "str", skip_quote=True), + } + request.url = self._client.format_url(request.url, **path_format_arguments) + + _stream = kwargs.pop("stream", False) + pipeline_response: PipelineResponse = await self._client._pipeline.run( # type: ignore # pylint: disable=protected-access + request, stream=_stream, **kwargs + ) + + response = pipeline_response.http_response + + if response.status_code not in [200]: + map_error(status_code=response.status_code, response=response, error_map=error_map) + raise HttpResponseError(response=response) + + if _stream: + deserialized = response.iter_bytes() + else: + deserialized = _deserialize(_models.AcknowledgeResult, response.json()) + + if cls: + return cls(pipeline_response, deserialized, {}) # type: ignore + + return deserialized # type: ignore + + @distributed_trace_async + async def release_cloud_events( + self, topic_name: str, event_subscription_name: str, lock_tokens: _models.ReleaseOptions, **kwargs: Any + ) -> _models.ReleaseResult: + """Release batch of Cloud Events. The server responds with an HTTP 200 status code if at least one + event is successfully released. The response body will include the set of successfully released + lockTokens, along with other failed lockTokens with their corresponding error information. + + :param topic_name: Topic Name. Required. + :type topic_name: str + :param event_subscription_name: Event Subscription Name. Required. + :type event_subscription_name: str + :param lock_tokens: ReleaseOptions. Required. + :type lock_tokens: ~azure.eventgrid.models.ReleaseOptions + :keyword content_type: content type. Default value is "application/json; charset=utf-8". + :paramtype content_type: str + :keyword bool stream: Whether to stream the response of this operation. Defaults to False. You + will have to context manage the returned stream. + :return: ReleaseResult. The ReleaseResult is compatible with MutableMapping + :rtype: ~azure.eventgrid.models.ReleaseResult + :raises ~azure.core.exceptions.HttpResponseError: + """ + error_map = { + 401: ClientAuthenticationError, + 404: ResourceNotFoundError, + 409: ResourceExistsError, + 304: ResourceNotModifiedError, + } + error_map.update(kwargs.pop("error_map", {}) or {}) + + _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) + _params = kwargs.pop("params", {}) or {} + + content_type: str = kwargs.pop("content_type", _headers.pop("content-type", "application/json; charset=utf-8")) + cls: ClsType[_models.ReleaseResult] = kwargs.pop("cls", None) + + _content = json.dumps(lock_tokens, cls=AzureJSONEncoder) # type: ignore + + request = build_event_grid_release_cloud_events_request( + topic_name=topic_name, + event_subscription_name=event_subscription_name, + content_type=content_type, + api_version=self._config.api_version, + content=_content, + headers=_headers, + params=_params, + ) + path_format_arguments = { + "endpoint": self._serialize.url("self._config.endpoint", self._config.endpoint, "str", skip_quote=True), + } + request.url = self._client.format_url(request.url, **path_format_arguments) + + _stream = kwargs.pop("stream", False) + pipeline_response: PipelineResponse = await self._client._pipeline.run( # type: ignore # pylint: disable=protected-access + request, stream=_stream, **kwargs + ) + + response = pipeline_response.http_response + + if response.status_code not in [200]: + map_error(status_code=response.status_code, response=response, error_map=error_map) + raise HttpResponseError(response=response) + + if _stream: + deserialized = response.iter_bytes() + else: + deserialized = _deserialize(_models.ReleaseResult, response.json()) + + if cls: + return cls(pipeline_response, deserialized, {}) # type: ignore + + return deserialized # type: ignore + + @distributed_trace_async + async def reject_cloud_events( + self, topic_name: str, event_subscription_name: str, lock_tokens: _models.RejectOptions, **kwargs: Any + ) -> _models.RejectResult: + """Reject batch of Cloud Events. + + :param topic_name: Topic Name. Required. + :type topic_name: str + :param event_subscription_name: Event Subscription Name. Required. + :type event_subscription_name: str + :param lock_tokens: RejectOptions. Required. + :type lock_tokens: ~azure.eventgrid.models.RejectOptions + :keyword content_type: content type. Default value is "application/json; charset=utf-8". + :paramtype content_type: str + :keyword bool stream: Whether to stream the response of this operation. Defaults to False. You + will have to context manage the returned stream. + :return: RejectResult. The RejectResult is compatible with MutableMapping + :rtype: ~azure.eventgrid.models.RejectResult + :raises ~azure.core.exceptions.HttpResponseError: + """ + error_map = { + 401: ClientAuthenticationError, + 404: ResourceNotFoundError, + 409: ResourceExistsError, + 304: ResourceNotModifiedError, + } + error_map.update(kwargs.pop("error_map", {}) or {}) + + _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) + _params = kwargs.pop("params", {}) or {} + + content_type: str = kwargs.pop("content_type", _headers.pop("content-type", "application/json; charset=utf-8")) + cls: ClsType[_models.RejectResult] = kwargs.pop("cls", None) + + _content = json.dumps(lock_tokens, cls=AzureJSONEncoder) # type: ignore + + request = build_event_grid_reject_cloud_events_request( + topic_name=topic_name, + event_subscription_name=event_subscription_name, + content_type=content_type, + api_version=self._config.api_version, + content=_content, + headers=_headers, + params=_params, + ) + path_format_arguments = { + "endpoint": self._serialize.url("self._config.endpoint", self._config.endpoint, "str", skip_quote=True), + } + request.url = self._client.format_url(request.url, **path_format_arguments) + + _stream = kwargs.pop("stream", False) + pipeline_response: PipelineResponse = await self._client._pipeline.run( # type: ignore # pylint: disable=protected-access + request, stream=_stream, **kwargs + ) + + response = pipeline_response.http_response + + if response.status_code not in [200]: + map_error(status_code=response.status_code, response=response, error_map=error_map) + raise HttpResponseError(response=response) + + if _stream: + deserialized = response.iter_bytes() + else: + deserialized = _deserialize(_models.RejectResult, response.json()) + + if cls: + return cls(pipeline_response, deserialized, {}) # type: ignore + + return deserialized # type: ignore diff --git a/sdk/eventgrid/azure-eventgrid/azure/eventgrid/aio/_operations/_patch.py b/sdk/eventgrid/azure-eventgrid/azure/eventgrid/aio/_operations/_patch.py new file mode 100644 index 000000000000..f7dd32510333 --- /dev/null +++ b/sdk/eventgrid/azure-eventgrid/azure/eventgrid/aio/_operations/_patch.py @@ -0,0 +1,20 @@ +# ------------------------------------ +# Copyright (c) Microsoft Corporation. +# Licensed under the MIT License. +# ------------------------------------ +"""Customize generated code here. + +Follow our quickstart for examples: https://aka.ms/azsdk/python/dpcodegen/python/customize +""" +from typing import List + +__all__: List[str] = [] # Add all objects you want publicly available to users at this package level + + +def patch_sdk(): + """Do not remove from this file. + + `patch_sdk` is a last resort escape hatch that allows you to do customizations + you can't accomplish using the techniques described in + https://aka.ms/azsdk/python/dpcodegen/python/customize + """ diff --git a/sdk/eventgrid/azure-eventgrid/azure/eventgrid/aio/_patch.py b/sdk/eventgrid/azure-eventgrid/azure/eventgrid/aio/_patch.py new file mode 100644 index 000000000000..f7dd32510333 --- /dev/null +++ b/sdk/eventgrid/azure-eventgrid/azure/eventgrid/aio/_patch.py @@ -0,0 +1,20 @@ +# ------------------------------------ +# Copyright (c) Microsoft Corporation. +# Licensed under the MIT License. +# ------------------------------------ +"""Customize generated code here. + +Follow our quickstart for examples: https://aka.ms/azsdk/python/dpcodegen/python/customize +""" +from typing import List + +__all__: List[str] = [] # Add all objects you want publicly available to users at this package level + + +def patch_sdk(): + """Do not remove from this file. + + `patch_sdk` is a last resort escape hatch that allows you to do customizations + you can't accomplish using the techniques described in + https://aka.ms/azsdk/python/dpcodegen/python/customize + """ diff --git a/sdk/eventgrid/azure-eventgrid/azure/eventgrid/aio/_publisher_client_async.py b/sdk/eventgrid/azure-eventgrid/azure/eventgrid/aio/_publisher_client_async.py index 89c950448403..adc9ec2d93c1 100644 --- a/sdk/eventgrid/azure-eventgrid/azure/eventgrid/aio/_publisher_client_async.py +++ b/sdk/eventgrid/azure-eventgrid/azure/eventgrid/aio/_publisher_client_async.py @@ -29,7 +29,7 @@ HttpResponseError, ResourceNotFoundError, ResourceExistsError, - map_error + map_error, ) from .._policies import CloudEventDistributedTracingPolicy from .._models import EventGridEvent @@ -48,14 +48,12 @@ if TYPE_CHECKING: from azure.core.credentials_async import AsyncTokenCredential -SendType = Union[ - CloudEvent, EventGridEvent, Dict, List[CloudEvent], List[EventGridEvent], List[Dict] -] +SendType = Union[CloudEvent, EventGridEvent, Dict, List[CloudEvent], List[EventGridEvent], List[Dict]] ListEventType = Union[List[CloudEvent], List[EventGridEvent], List[Dict]] -class EventGridPublisherClient: # pylint: disable=client-accepts-api-version-keyword +class EventGridPublisherClient: # pylint: disable=client-accepts-api-version-keyword """Asynchronous EventGridPublisherClient publishes events to an EventGrid topic or domain. It can be used to publish either an EventGridEvent, a CloudEvent or a Custom Schema. @@ -86,9 +84,7 @@ class EventGridPublisherClient: # pylint: disable=client-accepts-api-version-key def __init__( self, endpoint: str, - credential: Union[ - "AsyncTokenCredential", AzureKeyCredential, AzureSasCredential - ], + credential: Union["AsyncTokenCredential", AzureKeyCredential, AzureSasCredential], **kwargs: Any ) -> None: self._client = EventGridPublisherClientAsync( @@ -98,14 +94,9 @@ def __init__( @staticmethod def _policies( - credential: Union[ - AzureKeyCredential, AzureSasCredential, "AsyncTokenCredential" - ], - **kwargs: Any + credential: Union[AzureKeyCredential, AzureSasCredential, "AsyncTokenCredential"], **kwargs: Any ) -> List[Any]: - auth_policy = _get_authentication_policy( - credential, AsyncBearerTokenCredentialPolicy - ) + auth_policy = _get_authentication_policy(credential, AsyncBearerTokenCredentialPolicy) sdk_moniker = "eventgridpublisherclient/{}".format(VERSION) policies = [ RequestIdPolicy(**kwargs), @@ -198,15 +189,10 @@ async def send(self, events: SendType, *, channel_name: Optional[str] = None, ** content_type = kwargs.pop("content_type", "application/json; charset=utf-8") if isinstance(events[0], CloudEvent) or _is_cloud_event(events[0]): try: - events = [ - _cloud_event_to_generated(e, **kwargs) - for e in events # pylint: disable=protected-access - ] + events = [_cloud_event_to_generated(e, **kwargs) for e in events] # pylint: disable=protected-access except AttributeError: ## this is either a dictionary or a CNCF cloud event - events = [ - _from_cncf_events(e) for e in events - ] + events = [_from_cncf_events(e) for e in events] content_type = "application/cloudevents-batch+json; charset=utf-8" elif isinstance(events[0], EventGridEvent) or _is_eventgrid_event(events[0]): for event in events: diff --git a/sdk/eventgrid/azure-eventgrid/azure/eventgrid/aio/_vendor.py b/sdk/eventgrid/azure-eventgrid/azure/eventgrid/aio/_vendor.py new file mode 100644 index 000000000000..35bd45e3dc6b --- /dev/null +++ b/sdk/eventgrid/azure-eventgrid/azure/eventgrid/aio/_vendor.py @@ -0,0 +1,26 @@ +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for license information. +# Code generated by Microsoft (R) Python Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is regenerated. +# -------------------------------------------------------------------------- + +from abc import ABC +from typing import TYPE_CHECKING + +from ._configuration import EventGridClientConfiguration + +if TYPE_CHECKING: + # pylint: disable=unused-import,ungrouped-imports + from azure.core import AsyncPipelineClient + + from .._serialization import Deserializer, Serializer + + +class EventGridClientMixinABC(ABC): + """DO NOT use this class. It is for internal typing use only.""" + + _client: "AsyncPipelineClient" + _config: EventGridClientConfiguration + _serialize: "Serializer" + _deserialize: "Deserializer" diff --git a/sdk/eventgrid/azure-eventgrid/azure/eventgrid/models/__init__.py b/sdk/eventgrid/azure-eventgrid/azure/eventgrid/models/__init__.py new file mode 100644 index 000000000000..cc99d92356dd --- /dev/null +++ b/sdk/eventgrid/azure-eventgrid/azure/eventgrid/models/__init__.py @@ -0,0 +1,38 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for license information. +# Code generated by Microsoft (R) Python Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is regenerated. +# -------------------------------------------------------------------------- + +from ._models import AcknowledgeOptions +from ._models import AcknowledgeResult +from ._models import BrokerProperties +from ._models import CloudEvent +from ._models import FailedLockToken +from ._models import ReceiveDetails +from ._models import ReceiveResult +from ._models import RejectOptions +from ._models import RejectResult +from ._models import ReleaseOptions +from ._models import ReleaseResult +from ._patch import __all__ as _patch_all +from ._patch import * # pylint: disable=unused-wildcard-import +from ._patch import patch_sdk as _patch_sdk + +__all__ = [ + "AcknowledgeOptions", + "AcknowledgeResult", + "BrokerProperties", + "CloudEvent", + "FailedLockToken", + "ReceiveDetails", + "ReceiveResult", + "RejectOptions", + "RejectResult", + "ReleaseOptions", + "ReleaseResult", +] +__all__.extend([p for p in _patch_all if p not in __all__]) +_patch_sdk() diff --git a/sdk/eventgrid/azure-eventgrid/azure/eventgrid/models/_models.py b/sdk/eventgrid/azure-eventgrid/azure/eventgrid/models/_models.py new file mode 100644 index 000000000000..2dd22010d7a4 --- /dev/null +++ b/sdk/eventgrid/azure-eventgrid/azure/eventgrid/models/_models.py @@ -0,0 +1,462 @@ +# coding=utf-8 +# pylint: disable=too-many-lines +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for license information. +# Code generated by Microsoft (R) Python Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is regenerated. +# -------------------------------------------------------------------------- + +import datetime +from typing import Any, List, Mapping, Optional, TYPE_CHECKING, overload + +from .. import _model_base +from .._model_base import rest_field + +if TYPE_CHECKING: + # pylint: disable=unused-import,ungrouped-imports + from .. import models as _models + + +class AcknowledgeOptions(_model_base.Model): + """Array of lock token strings for the corresponding received Cloud Events to be acknowledged. + + All required parameters must be populated in order to send to Azure. + + :ivar lock_tokens: String array of lock tokens. Required. + :vartype lock_tokens: list[str] + """ + + lock_tokens: List[str] = rest_field(name="lockTokens") + """String array of lock tokens. Required.""" + + @overload + def __init__( + self, + *, + lock_tokens: List[str], + ): + ... + + @overload + def __init__(self, mapping: Mapping[str, Any]): + """ + :param mapping: raw JSON to initialize the model. + :type mapping: Mapping[str, Any] + """ + + def __init__(self, *args: Any, **kwargs: Any) -> None: # pylint: disable=useless-super-delegation + super().__init__(*args, **kwargs) + + +class AcknowledgeResult(_model_base.Model): + """The result of the Acknowledge operation. + + All required parameters must be populated in order to send to Azure. + + :ivar failed_lock_tokens: Array of LockToken values for failed cloud events. Each LockToken + includes the lock token value along with the related error information (namely, the error code + and description). Required. + :vartype failed_lock_tokens: list[~azure.eventgrid.models.FailedLockToken] + :ivar succeeded_lock_tokens: Array of lock tokens values for the successfully acknowledged + cloud events. Required. + :vartype succeeded_lock_tokens: list[str] + """ + + failed_lock_tokens: List["_models.FailedLockToken"] = rest_field(name="failedLockTokens") + """Array of LockToken values for failed cloud events. Each LockToken includes the lock token value + along with the related error information (namely, the error code and description). Required.""" + succeeded_lock_tokens: List[str] = rest_field(name="succeededLockTokens") + """Array of lock tokens values for the successfully acknowledged cloud events. Required.""" + + @overload + def __init__( + self, + *, + failed_lock_tokens: List["_models.FailedLockToken"], + succeeded_lock_tokens: List[str], + ): + ... + + @overload + def __init__(self, mapping: Mapping[str, Any]): + """ + :param mapping: raw JSON to initialize the model. + :type mapping: Mapping[str, Any] + """ + + def __init__(self, *args: Any, **kwargs: Any) -> None: # pylint: disable=useless-super-delegation + super().__init__(*args, **kwargs) + + +class BrokerProperties(_model_base.Model): + """Properties of the Event Broker operation. + + All required parameters must be populated in order to send to Azure. + + :ivar lock_token: The token used to lock the event. Required. + :vartype lock_token: str + :ivar delivery_count: The attempt count for delivering the event. Required. + :vartype delivery_count: int + """ + + lock_token: str = rest_field(name="lockToken") + """The token used to lock the event. Required.""" + delivery_count: int = rest_field(name="deliveryCount") + """The attempt count for delivering the event. Required.""" + + @overload + def __init__( + self, + *, + lock_token: str, + delivery_count: int, + ): + ... + + @overload + def __init__(self, mapping: Mapping[str, Any]): + """ + :param mapping: raw JSON to initialize the model. + :type mapping: Mapping[str, Any] + """ + + def __init__(self, *args: Any, **kwargs: Any) -> None: # pylint: disable=useless-super-delegation + super().__init__(*args, **kwargs) + + +class CloudEvent(_model_base.Model): + """Properties of an event published to an Azure Messaging EventGrid Namespace topic using the + CloudEvent 1.0 Schema. + + All required parameters must be populated in order to send to Azure. + + :ivar id: An identifier for the event. The combination of id and source must be unique for each + distinct event. Required. + :vartype id: str + :ivar source: Identifies the context in which an event happened. The combination of id and + source must be unique for each distinct event. Required. + :vartype source: str + :ivar data: Event data specific to the event type. + :vartype data: any + :ivar data_base64: Event data specific to the event type, encoded as a base64 string. + :vartype data_base64: bytes + :ivar type: Type of event related to the originating occurrence. Required. + :vartype type: str + :ivar time: The time (in UTC) the event was generated, in RFC3339 format. + :vartype time: ~datetime.datetime + :ivar specversion: The version of the CloudEvents specification which the event uses. Required. + :vartype specversion: str + :ivar dataschema: Identifies the schema that data adheres to. + :vartype dataschema: str + :ivar datacontenttype: Content type of data value. + :vartype datacontenttype: str + :ivar subject: This describes the subject of the event in the context of the event producer + (identified by source). + :vartype subject: str + """ + + id: str = rest_field() + """An identifier for the event. The combination of id and source must be unique for each distinct + event. Required.""" + source: str = rest_field() + """Identifies the context in which an event happened. The combination of id and source must be + unique for each distinct event. Required.""" + data: Optional[Any] = rest_field() + """Event data specific to the event type.""" + data_base64: Optional[bytes] = rest_field() + """Event data specific to the event type, encoded as a base64 string.""" + type: str = rest_field() + """Type of event related to the originating occurrence. Required.""" + time: Optional[datetime.datetime] = rest_field() + """The time (in UTC) the event was generated, in RFC3339 format.""" + specversion: str = rest_field() + """The version of the CloudEvents specification which the event uses. Required.""" + dataschema: Optional[str] = rest_field() + """Identifies the schema that data adheres to.""" + datacontenttype: Optional[str] = rest_field() + """Content type of data value.""" + subject: Optional[str] = rest_field() + """This describes the subject of the event in the context of the event producer (identified by + source).""" + + @overload + def __init__( + self, + *, + id: str, # pylint: disable=redefined-builtin + source: str, + type: str, + specversion: str, + data: Optional[Any] = None, + data_base64: Optional[bytes] = None, + time: Optional[datetime.datetime] = None, + dataschema: Optional[str] = None, + datacontenttype: Optional[str] = None, + subject: Optional[str] = None, + ): + ... + + @overload + def __init__(self, mapping: Mapping[str, Any]): + """ + :param mapping: raw JSON to initialize the model. + :type mapping: Mapping[str, Any] + """ + + def __init__(self, *args: Any, **kwargs: Any) -> None: # pylint: disable=useless-super-delegation + super().__init__(*args, **kwargs) + + +class FailedLockToken(_model_base.Model): + """Failed LockToken information. + + All required parameters must be populated in order to send to Azure. + + :ivar lock_token: LockToken value. Required. + :vartype lock_token: str + :ivar error_code: Error code related to the token. Example of such error codes are BadToken: + which indicates the Token is not formatted correctly, TokenLost: which indicates that token is + not found, and InternalServerError: For any internal server errors. Required. + :vartype error_code: str + :ivar error_description: Description of the token error. Required. + :vartype error_description: str + """ + + lock_token: str = rest_field(name="lockToken") + """LockToken value. Required.""" + error_code: str = rest_field(name="errorCode") + """Error code related to the token. Example of such error codes are BadToken: which indicates the + Token is not formatted correctly, TokenLost: which indicates that token is not found, and + InternalServerError: For any internal server errors. Required.""" + error_description: str = rest_field(name="errorDescription") + """Description of the token error. Required.""" + + @overload + def __init__( + self, + *, + lock_token: str, + error_code: str, + error_description: str, + ): + ... + + @overload + def __init__(self, mapping: Mapping[str, Any]): + """ + :param mapping: raw JSON to initialize the model. + :type mapping: Mapping[str, Any] + """ + + def __init__(self, *args: Any, **kwargs: Any) -> None: # pylint: disable=useless-super-delegation + super().__init__(*args, **kwargs) + + +class ReceiveDetails(_model_base.Model): + """Receive operation details per Cloud Event. + + All required parameters must be populated in order to send to Azure. + + :ivar broker_properties: The Event Broker details. Required. + :vartype broker_properties: ~azure.eventgrid.models.BrokerProperties + :ivar event: Cloud Event details. Required. + :vartype event: ~azure.eventgrid.models.CloudEvent + """ + + broker_properties: "_models.BrokerProperties" = rest_field(name="brokerProperties") + """The Event Broker details. Required.""" + event: "_models.CloudEvent" = rest_field() + """Cloud Event details. Required.""" + + @overload + def __init__( + self, + *, + broker_properties: "_models.BrokerProperties", + event: "_models.CloudEvent", + ): + ... + + @overload + def __init__(self, mapping: Mapping[str, Any]): + """ + :param mapping: raw JSON to initialize the model. + :type mapping: Mapping[str, Any] + """ + + def __init__(self, *args: Any, **kwargs: Any) -> None: # pylint: disable=useless-super-delegation + super().__init__(*args, **kwargs) + + +class ReceiveResult(_model_base.Model): + """Details of the Receive operation response. + + All required parameters must be populated in order to send to Azure. + + :ivar value: Array of receive responses, one per cloud event. Required. + :vartype value: list[~azure.eventgrid.models.ReceiveDetails] + """ + + value: List["_models.ReceiveDetails"] = rest_field() + """Array of receive responses, one per cloud event. Required.""" + + @overload + def __init__( + self, + *, + value: List["_models.ReceiveDetails"], + ): + ... + + @overload + def __init__(self, mapping: Mapping[str, Any]): + """ + :param mapping: raw JSON to initialize the model. + :type mapping: Mapping[str, Any] + """ + + def __init__(self, *args: Any, **kwargs: Any) -> None: # pylint: disable=useless-super-delegation + super().__init__(*args, **kwargs) + + +class RejectOptions(_model_base.Model): + """Array of lock token strings for the corresponding received Cloud Events to be rejected. + + All required parameters must be populated in order to send to Azure. + + :ivar lock_tokens: String array of lock tokens. Required. + :vartype lock_tokens: list[str] + """ + + lock_tokens: List[str] = rest_field(name="lockTokens") + """String array of lock tokens. Required.""" + + @overload + def __init__( + self, + *, + lock_tokens: List[str], + ): + ... + + @overload + def __init__(self, mapping: Mapping[str, Any]): + """ + :param mapping: raw JSON to initialize the model. + :type mapping: Mapping[str, Any] + """ + + def __init__(self, *args: Any, **kwargs: Any) -> None: # pylint: disable=useless-super-delegation + super().__init__(*args, **kwargs) + + +class RejectResult(_model_base.Model): + """The result of the Reject operation. + + All required parameters must be populated in order to send to Azure. + + :ivar failed_lock_tokens: Array of LockToken values for failed cloud events. Each LockToken + includes the lock token value along with the related error information (namely, the error code + and description). Required. + :vartype failed_lock_tokens: list[~azure.eventgrid.models.FailedLockToken] + :ivar succeeded_lock_tokens: Array of lock tokens values for the successfully rejected cloud + events. Required. + :vartype succeeded_lock_tokens: list[str] + """ + + failed_lock_tokens: List["_models.FailedLockToken"] = rest_field(name="failedLockTokens") + """Array of LockToken values for failed cloud events. Each LockToken includes the lock token value + along with the related error information (namely, the error code and description). Required.""" + succeeded_lock_tokens: List[str] = rest_field(name="succeededLockTokens") + """Array of lock tokens values for the successfully rejected cloud events. Required.""" + + @overload + def __init__( + self, + *, + failed_lock_tokens: List["_models.FailedLockToken"], + succeeded_lock_tokens: List[str], + ): + ... + + @overload + def __init__(self, mapping: Mapping[str, Any]): + """ + :param mapping: raw JSON to initialize the model. + :type mapping: Mapping[str, Any] + """ + + def __init__(self, *args: Any, **kwargs: Any) -> None: # pylint: disable=useless-super-delegation + super().__init__(*args, **kwargs) + + +class ReleaseOptions(_model_base.Model): + """Array of lock token strings for the corresponding received Cloud Events to be released. + + All required parameters must be populated in order to send to Azure. + + :ivar lock_tokens: String array of lock tokens. Required. + :vartype lock_tokens: list[str] + """ + + lock_tokens: List[str] = rest_field(name="lockTokens") + """String array of lock tokens. Required.""" + + @overload + def __init__( + self, + *, + lock_tokens: List[str], + ): + ... + + @overload + def __init__(self, mapping: Mapping[str, Any]): + """ + :param mapping: raw JSON to initialize the model. + :type mapping: Mapping[str, Any] + """ + + def __init__(self, *args: Any, **kwargs: Any) -> None: # pylint: disable=useless-super-delegation + super().__init__(*args, **kwargs) + + +class ReleaseResult(_model_base.Model): + """The result of the Release operation. + + All required parameters must be populated in order to send to Azure. + + :ivar failed_lock_tokens: Array of LockToken values for failed cloud events. Each LockToken + includes the lock token value along with the related error information (namely, the error code + and description). Required. + :vartype failed_lock_tokens: list[~azure.eventgrid.models.FailedLockToken] + :ivar succeeded_lock_tokens: Array of lock tokens values for the successfully released cloud + events. Required. + :vartype succeeded_lock_tokens: list[str] + """ + + failed_lock_tokens: List["_models.FailedLockToken"] = rest_field(name="failedLockTokens") + """Array of LockToken values for failed cloud events. Each LockToken includes the lock token value + along with the related error information (namely, the error code and description). Required.""" + succeeded_lock_tokens: List[str] = rest_field(name="succeededLockTokens") + """Array of lock tokens values for the successfully released cloud events. Required.""" + + @overload + def __init__( + self, + *, + failed_lock_tokens: List["_models.FailedLockToken"], + succeeded_lock_tokens: List[str], + ): + ... + + @overload + def __init__(self, mapping: Mapping[str, Any]): + """ + :param mapping: raw JSON to initialize the model. + :type mapping: Mapping[str, Any] + """ + + def __init__(self, *args: Any, **kwargs: Any) -> None: # pylint: disable=useless-super-delegation + super().__init__(*args, **kwargs) diff --git a/sdk/eventgrid/azure-eventgrid/azure/eventgrid/models/_patch.py b/sdk/eventgrid/azure-eventgrid/azure/eventgrid/models/_patch.py new file mode 100644 index 000000000000..f7dd32510333 --- /dev/null +++ b/sdk/eventgrid/azure-eventgrid/azure/eventgrid/models/_patch.py @@ -0,0 +1,20 @@ +# ------------------------------------ +# Copyright (c) Microsoft Corporation. +# Licensed under the MIT License. +# ------------------------------------ +"""Customize generated code here. + +Follow our quickstart for examples: https://aka.ms/azsdk/python/dpcodegen/python/customize +""" +from typing import List + +__all__: List[str] = [] # Add all objects you want publicly available to users at this package level + + +def patch_sdk(): + """Do not remove from this file. + + `patch_sdk` is a last resort escape hatch that allows you to do customizations + you can't accomplish using the techniques described in + https://aka.ms/azsdk/python/dpcodegen/python/customize + """ diff --git a/sdk/eventgrid/azure-eventgrid/azure/eventgrid/py.typed b/sdk/eventgrid/azure-eventgrid/azure/eventgrid/py.typed index e69de29bb2d1..e5aff4f83af8 100644 --- a/sdk/eventgrid/azure-eventgrid/azure/eventgrid/py.typed +++ b/sdk/eventgrid/azure-eventgrid/azure/eventgrid/py.typed @@ -0,0 +1 @@ +# Marker file for PEP 561. \ No newline at end of file diff --git a/sdk/eventgrid/azure-eventgrid/samples/async_samples/sample_authentication_async.py b/sdk/eventgrid/azure-eventgrid/samples/async_samples/sample_authentication_async.py index c155652668af..cb5499f6f9d1 100644 --- a/sdk/eventgrid/azure-eventgrid/samples/async_samples/sample_authentication_async.py +++ b/sdk/eventgrid/azure-eventgrid/samples/async_samples/sample_authentication_async.py @@ -44,14 +44,9 @@ from azure.eventgrid.aio import EventGridPublisherClient from azure.eventgrid import EventGridEvent -event = EventGridEvent( - data={"team": "azure-sdk"}, - subject="Door1", - event_type="Azure.Sdk.Demo", - data_version="2.0" -) +event = EventGridEvent(data={"team": "azure-sdk"}, subject="Door1", event_type="Azure.Sdk.Demo", data_version="2.0") credential = DefaultAzureCredential() endpoint = os.environ["EVENTGRID_TOPIC_ENDPOINT"] client = EventGridPublisherClient(endpoint, credential) -# [END client_auth_with_token_cred_async] \ No newline at end of file +# [END client_auth_with_token_cred_async] diff --git a/sdk/eventgrid/azure-eventgrid/samples/async_samples/sample_publish_cloud_event_using_dict_async.py b/sdk/eventgrid/azure-eventgrid/samples/async_samples/sample_publish_cloud_event_using_dict_async.py index d46585fac3e9..d6737c21420d 100644 --- a/sdk/eventgrid/azure-eventgrid/samples/async_samples/sample_publish_cloud_event_using_dict_async.py +++ b/sdk/eventgrid/azure-eventgrid/samples/async_samples/sample_publish_cloud_event_using_dict_async.py @@ -24,25 +24,27 @@ topic_key = os.environ["EVENTGRID_CLOUD_EVENT_TOPIC_KEY"] endpoint = os.environ["EVENTGRID_CLOUD_EVENT_TOPIC_ENDPOINT"] + async def publish(): credential = AzureKeyCredential(topic_key) client = EventGridPublisherClient(endpoint, credential) # [START publish_cloud_event_dict_async] async with client: - await client.send([ - { - "type": "Contoso.Items.ItemReceived", - "source": "/contoso/items", - "data": { - "itemSku": "Contoso Item SKU #1" - }, - "subject": "Door1", - "specversion": "1.0", - "id": "randomclouduuid11" - } - ]) + await client.send( + [ + { + "type": "Contoso.Items.ItemReceived", + "source": "/contoso/items", + "data": {"itemSku": "Contoso Item SKU #1"}, + "subject": "Door1", + "specversion": "1.0", + "id": "randomclouduuid11", + } + ] + ) # [END publish_cloud_event_dict_async] -if __name__ == '__main__': + +if __name__ == "__main__": asyncio.run(publish()) diff --git a/sdk/eventgrid/azure-eventgrid/samples/async_samples/sample_publish_cncf_cloud_events_async.py b/sdk/eventgrid/azure-eventgrid/samples/async_samples/sample_publish_cncf_cloud_events_async.py index e2036f141e63..aa26709c8f2b 100644 --- a/sdk/eventgrid/azure-eventgrid/samples/async_samples/sample_publish_cncf_cloud_events_async.py +++ b/sdk/eventgrid/azure-eventgrid/samples/async_samples/sample_publish_cncf_cloud_events_async.py @@ -25,19 +25,18 @@ async def publish(): - + credential = AzureKeyCredential(topic_key) client = EventGridPublisherClient(endpoint, credential) - await client.send([ - CloudEvent( - attributes={ - "type": "cloudevent", - "source": "/cncf/cloud/event/1.0", - "subject": "testing-cncf-event" - }, - data=b'This is a cncf cloud event.', - ) - ]) + await client.send( + [ + CloudEvent( + attributes={"type": "cloudevent", "source": "/cncf/cloud/event/1.0", "subject": "testing-cncf-event"}, + data=b"This is a cncf cloud event.", + ) + ] + ) + -if __name__ == '__main__': +if __name__ == "__main__": asyncio.run(publish()) diff --git a/sdk/eventgrid/azure-eventgrid/samples/async_samples/sample_publish_custom_schema_to_a_topic_async.py b/sdk/eventgrid/azure-eventgrid/samples/async_samples/sample_publish_custom_schema_to_a_topic_async.py index 9c6001e4afae..2d7ccbb8cbcf 100644 --- a/sdk/eventgrid/azure-eventgrid/samples/async_samples/sample_publish_custom_schema_to_a_topic_async.py +++ b/sdk/eventgrid/azure-eventgrid/samples/async_samples/sample_publish_custom_schema_to_a_topic_async.py @@ -28,6 +28,7 @@ key = os.environ["EVENTGRID_CUSTOM_EVENT_TOPIC_KEY"] endpoint = os.environ["EVENTGRID_CUSTOM_EVENT_TOPIC_ENDPOINT"] + async def publish_event(): # authenticate client # [START publish_custom_schema_async] @@ -40,13 +41,14 @@ async def publish_event(): "customDataVersion": "2.0", "customId": uuid.uuid4(), "customEventTime": dt.datetime.now(UTC()).isoformat(), - "customData": "sample data" + "customData": "sample data", } async with client: # publish list of events await client.send(custom_schema_event) - + # [END publish_custom_schema_async] -if __name__ == '__main__': + +if __name__ == "__main__": asyncio.run(publish_event()) diff --git a/sdk/eventgrid/azure-eventgrid/samples/async_samples/sample_publish_eg_event_using_dict_async.py b/sdk/eventgrid/azure-eventgrid/samples/async_samples/sample_publish_eg_event_using_dict_async.py index a175bfec0d75..8909b5717b2b 100644 --- a/sdk/eventgrid/azure-eventgrid/samples/async_samples/sample_publish_eg_event_using_dict_async.py +++ b/sdk/eventgrid/azure-eventgrid/samples/async_samples/sample_publish_eg_event_using_dict_async.py @@ -26,35 +26,33 @@ topic_key = os.environ["EVENTGRID_TOPIC_KEY"] endpoint = os.environ["EVENTGRID_TOPIC_ENDPOINT"] + async def publish(): credential = AzureKeyCredential(topic_key) client = EventGridPublisherClient(endpoint, credential) # [START publish_eg_event_dict_async] - event0 = { - "eventType": "Contoso.Items.ItemReceived", - "data": { - "itemSku": "Contoso Item SKU #1" - }, - "subject": "Door1", - "dataVersion": "2.0", - "id": "randomuuid11", - "eventTime": datetime.utcnow() - } - event1 = { - "eventType": "Contoso.Items.ItemReceived", - "data": { - "itemSku": "Contoso Item SKU #2" - }, - "subject": "Door1", - "dataVersion": "2.0", - "id": "randomuuid12", - "eventTime": datetime.utcnow() + event0 = { + "eventType": "Contoso.Items.ItemReceived", + "data": {"itemSku": "Contoso Item SKU #1"}, + "subject": "Door1", + "dataVersion": "2.0", + "id": "randomuuid11", + "eventTime": datetime.utcnow(), + } + event1 = { + "eventType": "Contoso.Items.ItemReceived", + "data": {"itemSku": "Contoso Item SKU #2"}, + "subject": "Door1", + "dataVersion": "2.0", + "id": "randomuuid12", + "eventTime": datetime.utcnow(), } - async with client: + async with client: await client.send([event0, event1]) # [END publish_eg_event_dict_async] -if __name__ == '__main__': + +if __name__ == "__main__": asyncio.run(publish()) diff --git a/sdk/eventgrid/azure-eventgrid/samples/async_samples/sample_publish_eg_events_to_a_domain_async.py b/sdk/eventgrid/azure-eventgrid/samples/async_samples/sample_publish_eg_events_to_a_domain_async.py index 2973273fb180..48272cb30166 100644 --- a/sdk/eventgrid/azure-eventgrid/samples/async_samples/sample_publish_eg_events_to_a_domain_async.py +++ b/sdk/eventgrid/azure-eventgrid/samples/async_samples/sample_publish_eg_events_to_a_domain_async.py @@ -24,30 +24,30 @@ domain_key = os.environ["EVENTGRID_DOMAIN_KEY"] domain_hostname = os.environ["EVENTGRID_DOMAIN_ENDPOINT"] + async def publish(): credential = AzureKeyCredential(domain_key) client = EventGridPublisherClient(domain_hostname, credential) - await client.send([ - EventGridEvent( - topic="MyCustomDomainTopic1", - event_type="Contoso.Items.ItemReceived", - data={ - "itemSku": "Contoso Item SKU #1" - }, - subject="Door1", - data_version="2.0" - ), - EventGridEvent( - topic="MyCustomDomainTopic2", - event_type="Contoso.Items.ItemReceived", - data={ - "itemSku": "Contoso Item SKU #2" - }, - subject="Door1", - data_version="2.0" - ) - ]) + await client.send( + [ + EventGridEvent( + topic="MyCustomDomainTopic1", + event_type="Contoso.Items.ItemReceived", + data={"itemSku": "Contoso Item SKU #1"}, + subject="Door1", + data_version="2.0", + ), + EventGridEvent( + topic="MyCustomDomainTopic2", + event_type="Contoso.Items.ItemReceived", + data={"itemSku": "Contoso Item SKU #2"}, + subject="Door1", + data_version="2.0", + ), + ] + ) + -if __name__ == '__main__': +if __name__ == "__main__": asyncio.run(publish()) diff --git a/sdk/eventgrid/azure-eventgrid/samples/async_samples/sample_publish_eg_events_to_a_topic_async.py b/sdk/eventgrid/azure-eventgrid/samples/async_samples/sample_publish_eg_events_to_a_topic_async.py index f1924f54f66d..4ab76f2a81fa 100644 --- a/sdk/eventgrid/azure-eventgrid/samples/async_samples/sample_publish_eg_events_to_a_topic_async.py +++ b/sdk/eventgrid/azure-eventgrid/samples/async_samples/sample_publish_eg_events_to_a_topic_async.py @@ -24,21 +24,24 @@ topic_key = os.environ["EVENTGRID_TOPIC_KEY"] endpoint = os.environ["EVENTGRID_TOPIC_ENDPOINT"] + async def publish(): credential = AzureKeyCredential(topic_key) client = EventGridPublisherClient(endpoint, credential) - await client.send([ - EventGridEvent( - event_type="Contoso.Items.ItemReceived", - data={ - "itemSku": "Contoso Item SKU #1" - }, - subject="Door1", - data_version="2.0" - ) - ]) + await client.send( + [ + EventGridEvent( + event_type="Contoso.Items.ItemReceived", + data={"itemSku": "Contoso Item SKU #1"}, + subject="Door1", + data_version="2.0", + ) + ] + ) + + # [END publish_eg_event_to_topic_async] -if __name__ == '__main__': +if __name__ == "__main__": asyncio.run(publish()) diff --git a/sdk/eventgrid/azure-eventgrid/samples/async_samples/sample_publish_events_to_a_topic_using_sas_credential_async.py b/sdk/eventgrid/azure-eventgrid/samples/async_samples/sample_publish_events_to_a_topic_using_sas_credential_async.py index 26e4c745ddaf..c70bfd11aef4 100644 --- a/sdk/eventgrid/azure-eventgrid/samples/async_samples/sample_publish_events_to_a_topic_using_sas_credential_async.py +++ b/sdk/eventgrid/azure-eventgrid/samples/async_samples/sample_publish_events_to_a_topic_using_sas_credential_async.py @@ -23,21 +23,23 @@ sas = os.environ["EVENTGRID_SAS"] endpoint = os.environ["EVENTGRID_TOPIC_ENDPOINT"] + async def publish(): credential = AzureSasCredential(sas) client = EventGridPublisherClient(endpoint, credential) async with client: - await client.send([ - EventGridEvent( - event_type="Contoso.Items.ItemReceived", - data={ - "itemSku": "Contoso Item SKU #1" - }, - subject="Door1", - data_version="2.0" - ) - ]) + await client.send( + [ + EventGridEvent( + event_type="Contoso.Items.ItemReceived", + data={"itemSku": "Contoso Item SKU #1"}, + subject="Door1", + data_version="2.0", + ) + ] + ) + -if __name__ == '__main__': +if __name__ == "__main__": asyncio.run(publish()) diff --git a/sdk/eventgrid/azure-eventgrid/samples/async_samples/sample_publish_events_using_cloud_events_1.0_schema_async.py b/sdk/eventgrid/azure-eventgrid/samples/async_samples/sample_publish_events_using_cloud_events_1.0_schema_async.py index 8e238dfee5a1..1d3ea0337e33 100644 --- a/sdk/eventgrid/azure-eventgrid/samples/async_samples/sample_publish_events_using_cloud_events_1.0_schema_async.py +++ b/sdk/eventgrid/azure-eventgrid/samples/async_samples/sample_publish_events_using_cloud_events_1.0_schema_async.py @@ -24,21 +24,24 @@ topic_key = os.environ["EVENTGRID_CLOUD_EVENT_TOPIC_KEY"] endpoint = os.environ["EVENTGRID_CLOUD_EVENT_TOPIC_ENDPOINT"] + async def publish(): credential = AzureKeyCredential(topic_key) client = EventGridPublisherClient(endpoint, credential) - await client.send([ - CloudEvent( - type="Contoso.Items.ItemReceived", - source="/contoso/items", - data={ - "itemSku": "Contoso Item SKU #1" - }, - subject="Door1" - ) - ]) + await client.send( + [ + CloudEvent( + type="Contoso.Items.ItemReceived", + source="/contoso/items", + data={"itemSku": "Contoso Item SKU #1"}, + subject="Door1", + ) + ] + ) + + # [END publish_cloud_event_to_topic_async] -if __name__ == '__main__': +if __name__ == "__main__": asyncio.run(publish()) diff --git a/sdk/eventgrid/azure-eventgrid/samples/async_samples/sample_publish_to_channel_async.py b/sdk/eventgrid/azure-eventgrid/samples/async_samples/sample_publish_to_channel_async.py index 15a6526f0088..f12070c35a1e 100644 --- a/sdk/eventgrid/azure-eventgrid/samples/async_samples/sample_publish_to_channel_async.py +++ b/sdk/eventgrid/azure-eventgrid/samples/async_samples/sample_publish_to_channel_async.py @@ -22,26 +22,28 @@ from azure.core.credentials import AzureKeyCredential from azure.core.messaging import CloudEvent -topic_key = os.environ['EVENTGRID_PARTNER_NAMESPACE_TOPIC_KEY'] -endpoint = os.environ['EVENTGRID_PARTNER_NAMESPACE_TOPIC_ENDPOINT'] +topic_key = os.environ["EVENTGRID_PARTNER_NAMESPACE_TOPIC_KEY"] +endpoint = os.environ["EVENTGRID_PARTNER_NAMESPACE_TOPIC_ENDPOINT"] + +channel_name = os.environ["EVENTGRID_PARTNER_CHANNEL_NAME"] -channel_name = os.environ['EVENTGRID_PARTNER_CHANNEL_NAME'] async def publish(): credential = AzureKeyCredential(topic_key) client = EventGridPublisherClient(endpoint, credential) async with client: - await client.send([ - CloudEvent( - type="Contoso.Items.ItemReceived", - source="/contoso/items", - data={ - "itemSku": "Contoso Item SKU #1" - }, - subject="Door1" - ) - ], - channel_name=channel_name) + await client.send( + [ + CloudEvent( + type="Contoso.Items.ItemReceived", + source="/contoso/items", + data={"itemSku": "Contoso Item SKU #1"}, + subject="Door1", + ) + ], + channel_name=channel_name, + ) + -if __name__ == '__main__': +if __name__ == "__main__": asyncio.run(publish()) diff --git a/sdk/eventgrid/azure-eventgrid/samples/consume_samples/consume_cloud_events_from_eventhub.py b/sdk/eventgrid/azure-eventgrid/samples/consume_samples/consume_cloud_events_from_eventhub.py index 9aee835f0bfc..b569b239664e 100644 --- a/sdk/eventgrid/azure-eventgrid/samples/consume_samples/consume_cloud_events_from_eventhub.py +++ b/sdk/eventgrid/azure-eventgrid/samples/consume_samples/consume_cloud_events_from_eventhub.py @@ -24,19 +24,20 @@ CONNECTION_STR = os.environ["EVENT_HUB_CONN_STR"] EVENTHUB_NAME = os.environ["EVENT_HUB_NAME"] + + def on_event(partition_context, event): dict_event = CloudEvent.from_json(event) print("data: {}\n".format(dict_event.data)) + consumer_client = EventHubConsumerClient.from_connection_string( conn_str=CONNECTION_STR, - consumer_group='$Default', + consumer_group="$Default", eventhub_name=EVENTHUB_NAME, ) with consumer_client: event_list = consumer_client.receive( - on_event=on_event, - starting_position="-1", # "-1" is from the beginning of the partition. - prefetch=5 + on_event=on_event, starting_position="-1", prefetch=5 # "-1" is from the beginning of the partition. ) diff --git a/sdk/eventgrid/azure-eventgrid/samples/consume_samples/consume_cloud_events_from_storage_queue.py b/sdk/eventgrid/azure-eventgrid/samples/consume_samples/consume_cloud_events_from_storage_queue.py index 4db27504690e..f184b889e1cc 100644 --- a/sdk/eventgrid/azure-eventgrid/samples/consume_samples/consume_cloud_events_from_storage_queue.py +++ b/sdk/eventgrid/azure-eventgrid/samples/consume_samples/consume_cloud_events_from_storage_queue.py @@ -20,17 +20,16 @@ import json # all types of CloudEvents below produce same DeserializedEvent -connection_str = os.environ['AZURE_STORAGE_CONNECTION_STRING'] -queue_name = os.environ['STORAGE_QUEUE_NAME'] +connection_str = os.environ["AZURE_STORAGE_CONNECTION_STRING"] +queue_name = os.environ["STORAGE_QUEUE_NAME"] with QueueServiceClient.from_connection_string(connection_str) as qsc: - payload = qsc.get_queue_client( - queue=queue_name, - message_decode_policy=BinaryBase64DecodePolicy() - ).peek_messages(max_messages=32) + payload = qsc.get_queue_client(queue=queue_name, message_decode_policy=BinaryBase64DecodePolicy()).peek_messages( + max_messages=32 + ) ## deserialize payload into a list of typed Events events = [CloudEvent.from_json(msg) for msg in payload] for event in events: - print(type(event)) ## CloudEvent + print(type(event)) ## CloudEvent diff --git a/sdk/eventgrid/azure-eventgrid/samples/consume_samples/consume_eventgrid_events_from_service_bus_queue.py b/sdk/eventgrid/azure-eventgrid/samples/consume_samples/consume_eventgrid_events_from_service_bus_queue.py index 9889ea87c112..25c6a3b49d05 100644 --- a/sdk/eventgrid/azure-eventgrid/samples/consume_samples/consume_eventgrid_events_from_service_bus_queue.py +++ b/sdk/eventgrid/azure-eventgrid/samples/consume_samples/consume_eventgrid_events_from_service_bus_queue.py @@ -23,14 +23,14 @@ import json # all types of EventGridEvents below produce same DeserializedEvent -connection_str = os.environ['SERVICE_BUS_CONNECTION_STR'] -queue_name = os.environ['SERVICE_BUS_QUEUE_NAME'] +connection_str = os.environ["SERVICE_BUS_CONNECTION_STR"] +queue_name = os.environ["SERVICE_BUS_QUEUE_NAME"] with ServiceBusClient.from_connection_string(connection_str) as sb_client: - payload = sb_client.get_queue_receiver(queue_name).receive_messages() + payload = sb_client.get_queue_receiver(queue_name).receive_messages() ## deserialize payload into a list of typed Events events = [EventGridEvent.from_json(msg) for msg in payload] for event in events: - print(type(event)) ## EventGridEvent + print(type(event)) ## EventGridEvent diff --git a/sdk/eventgrid/azure-eventgrid/samples/consume_samples/functionsapp/EventGridTrigger1/__init__.py b/sdk/eventgrid/azure-eventgrid/samples/consume_samples/functionsapp/EventGridTrigger1/__init__.py index b7a38ab6b5b3..56c82506fc64 100644 --- a/sdk/eventgrid/azure-eventgrid/samples/consume_samples/functionsapp/EventGridTrigger1/__init__.py +++ b/sdk/eventgrid/azure-eventgrid/samples/consume_samples/functionsapp/EventGridTrigger1/__init__.py @@ -5,16 +5,19 @@ import azure.functions as func from azure.eventgrid import EventGridEvent + def main(event: func.EventGridEvent): logging.info(sys.version) logging.info(event) - result = json.dumps({ - 'id': event.id, - 'data': event.get_json(), - 'topic': event.topic, - 'subject': event.subject, - 'event_type': event.event_type - }) + result = json.dumps( + { + "id": event.id, + "data": event.get_json(), + "topic": event.topic, + "subject": event.subject, + "event_type": event.event_type, + } + ) logging.info(result) deserialized_event = EventGridEvent.from_dict(json.loads(result)) ## can only be EventGridEvent diff --git a/sdk/eventgrid/azure-eventgrid/samples/publish_samples/publish_cloud_events_to_custom_topic_sample.py b/sdk/eventgrid/azure-eventgrid/samples/publish_samples/publish_cloud_events_to_custom_topic_sample.py index 541b07a61205..6a6ef301fb8e 100644 --- a/sdk/eventgrid/azure-eventgrid/samples/publish_samples/publish_cloud_events_to_custom_topic_sample.py +++ b/sdk/eventgrid/azure-eventgrid/samples/publish_samples/publish_cloud_events_to_custom_topic_sample.py @@ -30,21 +30,20 @@ credential = AzureKeyCredential(key) client = EventGridPublisherClient(endpoint, credential) -services = ["EventGrid", "ServiceBus", "EventHubs", "Storage"] # possible values for data field +services = ["EventGrid", "ServiceBus", "EventHubs", "Storage"] # possible values for data field + def publish_event(): # publish events for _ in range(3): - event_list = [] # list of events to publish + event_list = [] # list of events to publish # create events and append to list for j in range(randint(1, 1)): - sample_members = sample(services, k=randint(1, 4)) # select random subset of team members + sample_members = sample(services, k=randint(1, 4)) # select random subset of team members data_dict = {"team": sample_members} event = CloudEvent( - type="Azure.Sdk.Sample", - source="https://egsample.dev/sampleevent", - data={"team": sample_members} - ) + type="Azure.Sdk.Sample", source="https://egsample.dev/sampleevent", data={"team": sample_members} + ) event_list.append(event) # publish list of events @@ -52,5 +51,6 @@ def publish_event(): print("Batch of size {} published".format(len(event_list))) time.sleep(randint(1, 5)) + if __name__ == "__main__": publish_event() diff --git a/sdk/eventgrid/azure-eventgrid/samples/publish_samples/publish_cloud_events_to_domain_topic_sample.py b/sdk/eventgrid/azure-eventgrid/samples/publish_samples/publish_cloud_events_to_domain_topic_sample.py index 4deab71ad6b5..ba9182a40beb 100644 --- a/sdk/eventgrid/azure-eventgrid/samples/publish_samples/publish_cloud_events_to_domain_topic_sample.py +++ b/sdk/eventgrid/azure-eventgrid/samples/publish_samples/publish_cloud_events_to_domain_topic_sample.py @@ -32,21 +32,18 @@ credential = AzureKeyCredential(domain_key) client = EventGridPublisherClient(domain_endpoint, credential) + def publish_event(): # publish events for _ in range(3): - event_list = [] # list of events to publish - services = ["EventGrid", "ServiceBus", "EventHubs", "Storage"] # possible values for data field + event_list = [] # list of events to publish + services = ["EventGrid", "ServiceBus", "EventHubs", "Storage"] # possible values for data field # create events and append to list for j in range(randint(1, 3)): - sample_members = sample(services, k=randint(1, 4)) # select random subset of team members - event = CloudEvent( - type="Azure.Sdk.Demo", - source='domainname', - data={"team": sample_members} - ) + sample_members = sample(services, k=randint(1, 4)) # select random subset of team members + event = CloudEvent(type="Azure.Sdk.Demo", source="domainname", data={"team": sample_members}) event_list.append(event) # publish list of events @@ -54,5 +51,6 @@ def publish_event(): print("Batch of size {} published".format(len(event_list))) time.sleep(randint(1, 5)) -if __name__ == '__main__': + +if __name__ == "__main__": publish_event() diff --git a/sdk/eventgrid/azure-eventgrid/samples/publish_samples/publish_custom_schema_events_to_topic_sample.py b/sdk/eventgrid/azure-eventgrid/samples/publish_samples/publish_custom_schema_events_to_topic_sample.py index c24b7c3fddb4..0ab11a428c5a 100644 --- a/sdk/eventgrid/azure-eventgrid/samples/publish_samples/publish_custom_schema_events_to_topic_sample.py +++ b/sdk/eventgrid/azure-eventgrid/samples/publish_samples/publish_custom_schema_events_to_topic_sample.py @@ -27,6 +27,7 @@ key = os.environ["EVENTGRID_CUSTOM_EVENT_TOPIC_KEY"] endpoint = os.environ["EVENTGRID_CUSTOM_EVENT_TOPIC_ENDPOINT"] + def publish_event(): # authenticate client credential = AzureKeyCredential(key) @@ -38,13 +39,13 @@ def publish_event(): "customDataVersion": "2.0", "customId": uuid.uuid4(), "customEventTime": dt.datetime.now(UTC()).isoformat(), - "customData": "sample data" + "customData": "sample data", } # publish events - for _ in range(3): + for _ in range(3): - event_list = [] # list of events to publish + event_list = [] # list of events to publish # create events and append to list for j in range(randint(1, 3)): event_list.append(custom_schema_event) @@ -55,5 +56,5 @@ def publish_event(): time.sleep(randint(1, 5)) -if __name__ == '__main__': +if __name__ == "__main__": publish_event() diff --git a/sdk/eventgrid/azure-eventgrid/samples/publish_samples/publish_event_grid_events_to_custom_topic_sample.py b/sdk/eventgrid/azure-eventgrid/samples/publish_samples/publish_event_grid_events_to_custom_topic_sample.py index 209d3ca48582..16247a4763bf 100644 --- a/sdk/eventgrid/azure-eventgrid/samples/publish_samples/publish_event_grid_events_to_custom_topic_sample.py +++ b/sdk/eventgrid/azure-eventgrid/samples/publish_samples/publish_event_grid_events_to_custom_topic_sample.py @@ -28,22 +28,20 @@ # authenticate client credential = AzureKeyCredential(key) client = EventGridPublisherClient(endpoint, credential) -services = ["EventGrid", "ServiceBus", "EventHubs", "Storage"] # possible values for data field +services = ["EventGrid", "ServiceBus", "EventHubs", "Storage"] # possible values for data field + def publish_event(): # publish events for _ in range(3): - event_list = [] # list of events to publish + event_list = [] # list of events to publish # create events and append to list for j in range(randint(1, 3)): - sample_members = sample(services, k=randint(1, 4)) # select random subset of team members + sample_members = sample(services, k=randint(1, 4)) # select random subset of team members event = EventGridEvent( - subject="Door1", - data={"team": sample_members}, - event_type="Azure.Sdk.Demo", - data_version="2.0" - ) + subject="Door1", data={"team": sample_members}, event_type="Azure.Sdk.Demo", data_version="2.0" + ) event_list.append(event) # publish list of events @@ -51,5 +49,6 @@ def publish_event(): print("Batch of size {} published".format(len(event_list))) time.sleep(randint(1, 5)) -if __name__ == '__main__': + +if __name__ == "__main__": publish_event() diff --git a/sdk/eventgrid/azure-eventgrid/samples/publish_samples/publish_with_shared_access_signature_sample.py b/sdk/eventgrid/azure-eventgrid/samples/publish_samples/publish_with_shared_access_signature_sample.py index 9e0c02c15914..fc57050087f3 100644 --- a/sdk/eventgrid/azure-eventgrid/samples/publish_samples/publish_with_shared_access_signature_sample.py +++ b/sdk/eventgrid/azure-eventgrid/samples/publish_samples/publish_with_shared_access_signature_sample.py @@ -34,21 +34,20 @@ credential = AzureSasCredential(signature) client = EventGridPublisherClient(endpoint, credential) -services = ["EventGrid", "ServiceBus", "EventHubs", "Storage"] # possible values for data field +services = ["EventGrid", "ServiceBus", "EventHubs", "Storage"] # possible values for data field + def publish_event(): # publish events for _ in range(3): - event_list = [] # list of events to publish + event_list = [] # list of events to publish # create events and append to list for j in range(randint(1, 3)): - sample_members = sample(services, k=randint(1, 4)) # select random subset of team members + sample_members = sample(services, k=randint(1, 4)) # select random subset of team members event = CloudEvent( - type="Azure.Sdk.Demo", - source="https://egdemo.dev/demowithsignature", - data={"team": sample_members} - ) + type="Azure.Sdk.Demo", source="https://egdemo.dev/demowithsignature", data={"team": sample_members} + ) event_list.append(event) # publish list of events @@ -56,5 +55,6 @@ def publish_event(): print("Batch of size {} published".format(len(event_list))) time.sleep(randint(1, 5)) -if __name__ == '__main__': + +if __name__ == "__main__": publish_event() diff --git a/sdk/eventgrid/azure-eventgrid/samples/sync_samples/sample_authentication.py b/sdk/eventgrid/azure-eventgrid/samples/sync_samples/sample_authentication.py index d700513188b7..1258671589fa 100644 --- a/sdk/eventgrid/azure-eventgrid/samples/sync_samples/sample_authentication.py +++ b/sdk/eventgrid/azure-eventgrid/samples/sync_samples/sample_authentication.py @@ -46,4 +46,4 @@ credential = DefaultAzureCredential() endpoint = os.environ["EVENTGRID_TOPIC_ENDPOINT"] client = EventGridPublisherClient(endpoint, credential) -# [END client_auth_with_token_cred] \ No newline at end of file +# [END client_auth_with_token_cred] diff --git a/sdk/eventgrid/azure-eventgrid/samples/sync_samples/sample_generate_sas.py b/sdk/eventgrid/azure-eventgrid/samples/sync_samples/sample_generate_sas.py index 4c8eb1141a91..deb0d72d6649 100644 --- a/sdk/eventgrid/azure-eventgrid/samples/sync_samples/sample_generate_sas.py +++ b/sdk/eventgrid/azure-eventgrid/samples/sync_samples/sample_generate_sas.py @@ -22,11 +22,11 @@ topic_key = os.environ["EVENTGRID_TOPIC_KEY"] endpoint = os.environ["EVENTGRID_TOPIC_ENDPOINT"] -#represents the expiration date for sas +# represents the expiration date for sas expiration_date_utc = datetime.utcnow() + timedelta(hours=10) signature = generate_sas(endpoint, topic_key, expiration_date_utc) # [END generate_sas] -print(signature) \ No newline at end of file +print(signature) diff --git a/sdk/eventgrid/azure-eventgrid/samples/sync_samples/sample_publish_cloud_event_using_dict.py b/sdk/eventgrid/azure-eventgrid/samples/sync_samples/sample_publish_cloud_event_using_dict.py index 25be49359a50..0630d409ebd6 100644 --- a/sdk/eventgrid/azure-eventgrid/samples/sync_samples/sample_publish_cloud_event_using_dict.py +++ b/sdk/eventgrid/azure-eventgrid/samples/sync_samples/sample_publish_cloud_event_using_dict.py @@ -26,16 +26,16 @@ client = EventGridPublisherClient(endpoint, credential) # [START publish_cloud_event_dict] -client.send([ - { - "type": "Contoso.Items.ItemReceived", - "source": "/contoso/items", - "data": { - "itemSku": "Contoso Item SKU #1" - }, - "subject": "Door1", - "specversion": "1.0", - "id": "randomclouduuid11" - } -]) +client.send( + [ + { + "type": "Contoso.Items.ItemReceived", + "source": "/contoso/items", + "data": {"itemSku": "Contoso Item SKU #1"}, + "subject": "Door1", + "specversion": "1.0", + "id": "randomclouduuid11", + } + ] +) # [END publish_cloud_event_dict] diff --git a/sdk/eventgrid/azure-eventgrid/samples/sync_samples/sample_publish_cncf_cloud_events.py b/sdk/eventgrid/azure-eventgrid/samples/sync_samples/sample_publish_cncf_cloud_events.py index a54a7442a15a..b686d9220ed1 100644 --- a/sdk/eventgrid/azure-eventgrid/samples/sync_samples/sample_publish_cncf_cloud_events.py +++ b/sdk/eventgrid/azure-eventgrid/samples/sync_samples/sample_publish_cncf_cloud_events.py @@ -25,13 +25,11 @@ credential = AzureKeyCredential(topic_key) client = EventGridPublisherClient(endpoint, credential) -client.send([ - CloudEvent( - attributes={ - "type": "cloudevent", - "source": "/cncf/cloud/event/1.0", - "subject": "testing-cncf-event" - }, - data=b'This is a cncf cloud event.', - ) -]) +client.send( + [ + CloudEvent( + attributes={"type": "cloudevent", "source": "/cncf/cloud/event/1.0", "subject": "testing-cncf-event"}, + data=b"This is a cncf cloud event.", + ) + ] +) diff --git a/sdk/eventgrid/azure-eventgrid/samples/sync_samples/sample_publish_custom_schema_to_a_topic.py b/sdk/eventgrid/azure-eventgrid/samples/sync_samples/sample_publish_custom_schema_to_a_topic.py index abfaba2c0083..4650aae15d61 100644 --- a/sdk/eventgrid/azure-eventgrid/samples/sync_samples/sample_publish_custom_schema_to_a_topic.py +++ b/sdk/eventgrid/azure-eventgrid/samples/sync_samples/sample_publish_custom_schema_to_a_topic.py @@ -27,6 +27,7 @@ key = os.environ["EVENTGRID_CUSTOM_EVENT_TOPIC_KEY"] endpoint = os.environ["EVENTGRID_CUSTOM_EVENT_TOPIC_ENDPOINT"] + def publish_event(): # authenticate client credential = AzureKeyCredential(key) @@ -39,11 +40,12 @@ def publish_event(): "customDataVersion": "2.0", "customId": uuid.uuid4(), "customEventTime": dt.datetime.now(UTC()).isoformat(), - "customData": "sample data" + "customData": "sample data", } client.send(custom_schema_event) # [END publish_custom_schema] -if __name__ == '__main__': + +if __name__ == "__main__": publish_event() diff --git a/sdk/eventgrid/azure-eventgrid/samples/sync_samples/sample_publish_eg_event_using_dict.py b/sdk/eventgrid/azure-eventgrid/samples/sync_samples/sample_publish_eg_event_using_dict.py index 97b84b922592..b7154a0e07b8 100644 --- a/sdk/eventgrid/azure-eventgrid/samples/sync_samples/sample_publish_eg_event_using_dict.py +++ b/sdk/eventgrid/azure-eventgrid/samples/sync_samples/sample_publish_eg_event_using_dict.py @@ -25,33 +25,31 @@ topic_key = os.environ["EVENTGRID_TOPIC_KEY"] endpoint = os.environ["EVENTGRID_TOPIC_ENDPOINT"] + def publish(): # [START publish_eg_event_dict] credential = AzureKeyCredential(topic_key) client = EventGridPublisherClient(endpoint, credential) - event0 = { - "eventType": "Contoso.Items.ItemReceived", - "data": { - "itemSku": "Contoso Item SKU #1" - }, - "subject": "Door1", - "dataVersion": "2.0", - "id": "randomuuid11", - "eventTime": datetime.now(UTC()) - } - event1 = { - "eventType": "Contoso.Items.ItemReceived", - "data": { - "itemSku": "Contoso Item SKU #2" - }, - "subject": "Door1", - "dataVersion": "2.0", - "id": "randomuuid12", - "eventTime": datetime.now(UTC()) - } + event0 = { + "eventType": "Contoso.Items.ItemReceived", + "data": {"itemSku": "Contoso Item SKU #1"}, + "subject": "Door1", + "dataVersion": "2.0", + "id": "randomuuid11", + "eventTime": datetime.now(UTC()), + } + event1 = { + "eventType": "Contoso.Items.ItemReceived", + "data": {"itemSku": "Contoso Item SKU #2"}, + "subject": "Door1", + "dataVersion": "2.0", + "id": "randomuuid12", + "eventTime": datetime.now(UTC()), + } client.send([event0, event1]) # [END publish_eg_event_dict] -if __name__ == '__main__': + +if __name__ == "__main__": publish() diff --git a/sdk/eventgrid/azure-eventgrid/samples/sync_samples/sample_publish_eg_events_to_a_domain.py b/sdk/eventgrid/azure-eventgrid/samples/sync_samples/sample_publish_eg_events_to_a_domain.py index e66d26c7a36a..960f91aeae70 100644 --- a/sdk/eventgrid/azure-eventgrid/samples/sync_samples/sample_publish_eg_events_to_a_domain.py +++ b/sdk/eventgrid/azure-eventgrid/samples/sync_samples/sample_publish_eg_events_to_a_domain.py @@ -24,23 +24,21 @@ credential = AzureKeyCredential(domain_key) client = EventGridPublisherClient(domain_hostname, credential) -client.send([ - EventGridEvent( - topic="MyCustomDomainTopic1", - event_type="Contoso.Items.ItemReceived", - data={ - "itemSku": "Contoso Item SKU #1" - }, - subject="Door1", - data_version="2.0" - ), - EventGridEvent( - topic="MyCustomDomainTopic2", - event_type="Contoso.Items.ItemReceived", - data={ - "itemSku": "Contoso Item SKU #2" - }, - subject="Door1", - data_version="2.0" - ) -]) +client.send( + [ + EventGridEvent( + topic="MyCustomDomainTopic1", + event_type="Contoso.Items.ItemReceived", + data={"itemSku": "Contoso Item SKU #1"}, + subject="Door1", + data_version="2.0", + ), + EventGridEvent( + topic="MyCustomDomainTopic2", + event_type="Contoso.Items.ItemReceived", + data={"itemSku": "Contoso Item SKU #2"}, + subject="Door1", + data_version="2.0", + ), + ] +) diff --git a/sdk/eventgrid/azure-eventgrid/samples/sync_samples/sample_publish_eg_events_to_a_topic.py b/sdk/eventgrid/azure-eventgrid/samples/sync_samples/sample_publish_eg_events_to_a_topic.py index 9138a8ab1899..a0e3ab58b6dc 100644 --- a/sdk/eventgrid/azure-eventgrid/samples/sync_samples/sample_publish_eg_events_to_a_topic.py +++ b/sdk/eventgrid/azure-eventgrid/samples/sync_samples/sample_publish_eg_events_to_a_topic.py @@ -25,14 +25,14 @@ credential = AzureKeyCredential(topic_key) client = EventGridPublisherClient(endpoint, credential) -client.send([ - EventGridEvent( - event_type="Contoso.Items.ItemReceived", - data={ - "itemSku": "Contoso Item SKU #1" - }, - subject="Door1", - data_version="2.0" - ) -]) +client.send( + [ + EventGridEvent( + event_type="Contoso.Items.ItemReceived", + data={"itemSku": "Contoso Item SKU #1"}, + subject="Door1", + data_version="2.0", + ) + ] +) # [END publish_eg_event_to_topic] diff --git a/sdk/eventgrid/azure-eventgrid/samples/sync_samples/sample_publish_events_to_a_topic_using_sas_credential.py b/sdk/eventgrid/azure-eventgrid/samples/sync_samples/sample_publish_events_to_a_topic_using_sas_credential.py index 78ad100c40f9..6051d883fbd4 100644 --- a/sdk/eventgrid/azure-eventgrid/samples/sync_samples/sample_publish_events_to_a_topic_using_sas_credential.py +++ b/sdk/eventgrid/azure-eventgrid/samples/sync_samples/sample_publish_events_to_a_topic_using_sas_credential.py @@ -25,13 +25,13 @@ credential = AzureSasCredential(sas) client = EventGridPublisherClient(endpoint, credential) -client.send([ - EventGridEvent( - event_type="Contoso.Items.ItemReceived", - data={ - "itemSku": "Contoso Item SKU #1" - }, - subject="Door1", - data_version="2.0" - ) -]) +client.send( + [ + EventGridEvent( + event_type="Contoso.Items.ItemReceived", + data={"itemSku": "Contoso Item SKU #1"}, + subject="Door1", + data_version="2.0", + ) + ] +) diff --git a/sdk/eventgrid/azure-eventgrid/samples/sync_samples/sample_publish_events_using_cloud_events_1.0_schema.py b/sdk/eventgrid/azure-eventgrid/samples/sync_samples/sample_publish_events_using_cloud_events_1.0_schema.py index 9544f3fcd209..a0297f9950a3 100644 --- a/sdk/eventgrid/azure-eventgrid/samples/sync_samples/sample_publish_events_using_cloud_events_1.0_schema.py +++ b/sdk/eventgrid/azure-eventgrid/samples/sync_samples/sample_publish_events_using_cloud_events_1.0_schema.py @@ -26,14 +26,14 @@ credential = AzureKeyCredential(topic_key) client = EventGridPublisherClient(endpoint, credential) -client.send([ - CloudEvent( - type="Contoso.Items.ItemReceived", - source="/contoso/items", - data={ - "itemSku": "Contoso Item SKU #1" - }, - subject="Door1" - ) -]) +client.send( + [ + CloudEvent( + type="Contoso.Items.ItemReceived", + source="/contoso/items", + data={"itemSku": "Contoso Item SKU #1"}, + subject="Door1", + ) + ] +) # [END publish_cloud_event_to_topic] diff --git a/sdk/eventgrid/azure-eventgrid/setup.py b/sdk/eventgrid/azure-eventgrid/setup.py index e903d958c0d5..edd7b5532d83 100644 --- a/sdk/eventgrid/azure-eventgrid/setup.py +++ b/sdk/eventgrid/azure-eventgrid/setup.py @@ -1,71 +1,70 @@ -#!/usr/bin/env python - -#------------------------------------------------------------------------- +# coding=utf-8 +# -------------------------------------------------------------------------- # Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for -# license information. -#-------------------------------------------------------------------------- +# Licensed under the MIT License. See License.txt in the project root for license information. +# Code generated by Microsoft (R) Python Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is regenerated. +# -------------------------------------------------------------------------- +# coding: utf-8 +import os import re -import os.path -from io import open -from setuptools import find_packages, setup +from setuptools import setup, find_packages + -# Change the PACKAGE_NAME only to change folder and different name PACKAGE_NAME = "azure-eventgrid" -PACKAGE_PPRINT_NAME = "Event Grid" +PACKAGE_PPRINT_NAME = "Azure Event Grid" # a-b-c => a/b/c -package_folder_path = PACKAGE_NAME.replace('-', '/') -# a-b-c => a.b.c -namespace_name = PACKAGE_NAME.replace('-', '.') +package_folder_path = PACKAGE_NAME.replace("-", "/") # Version extraction inspired from 'requests' -with open(os.path.join(package_folder_path, 'version.py') - if os.path.exists(os.path.join(package_folder_path, 'version.py')) - else os.path.join(package_folder_path, '_version.py'), 'r') as fd: - version = re.search(r'^VERSION\s*=\s*[\'"]([^\'"]*)[\'"]', - fd.read(), re.MULTILINE).group(1) +with open(os.path.join(package_folder_path, "_version.py"), "r") as fd: + version = re.search(r'^VERSION\s*=\s*[\'"]([^\'"]*)[\'"]', fd.read(), re.MULTILINE).group(1) if not version: - raise RuntimeError('Cannot find version information') + raise RuntimeError("Cannot find version information") -with open('README.md', encoding='utf-8') as f: - readme = f.read() -with open('CHANGELOG.md', encoding='utf-8') as f: - changelog = f.read() setup( name=PACKAGE_NAME, version=version, - include_package_data=True, - description='Microsoft Azure {} Client Library for Python'.format(PACKAGE_PPRINT_NAME), - long_description=readme + '\n\n' + changelog, - long_description_content_type='text/markdown', - license='MIT License', - author='Microsoft Corporation', - author_email='azpysdkhelp@microsoft.com', - url='https://github.com/Azure/azure-sdk-for-python', + description="Microsoft {} Client Library for Python".format(PACKAGE_PPRINT_NAME), + long_description=open("README.md", "r").read(), + long_description_content_type="text/markdown", + license="MIT License", + author="Microsoft Corporation", + author_email="azpysdkhelp@microsoft.com", + url="https://github.com/Azure/azure-sdk-for-python/tree/main/sdk", + keywords="azure, azure sdk", classifiers=[ - "Development Status :: 5 - Production/Stable", - 'Programming Language :: Python', - 'Programming Language :: Python :: 3 :: Only', - 'Programming Language :: Python :: 3.7', - 'Programming Language :: Python :: 3.8', - 'Programming Language :: Python :: 3.9', - 'Programming Language :: Python :: 3.10', - 'License :: OSI Approved :: MIT License', + "Development Status :: 4 - Beta", + "Programming Language :: Python", + "Programming Language :: Python :: 3 :: Only", + "Programming Language :: Python :: 3", + "Programming Language :: Python :: 3.7", + "Programming Language :: Python :: 3.8", + "Programming Language :: Python :: 3.9", + "Programming Language :: Python :: 3.10", + "Programming Language :: Python :: 3.11", + "License :: OSI Approved :: MIT License", ], - python_requires=">=3.7", zip_safe=False, - packages=find_packages(exclude=[ - 'tests', - 'samples', - # Exclude packages that will be covered by PEP420 or nspkg - 'azure', - ]), + packages=find_packages( + exclude=[ + "tests", + # Exclude packages that will be covered by PEP420 or nspkg + "azure", + ] + ), + include_package_data=True, + package_data={ + "pytyped": ["py.typed"], + }, install_requires=[ - 'isodate>=0.6.0', - 'azure-core<2.0.0,>=1.24.0', + "isodate<1.0.0,>=0.6.1", + "azure-core<2.0.0,>=1.24.0", + "typing-extensions>=4.3.0; python_version<'3.8.0'", ], + python_requires=">=3.7", ) diff --git a/sdk/eventgrid/azure-eventgrid/swagger/_constants.py b/sdk/eventgrid/azure-eventgrid/swagger/_constants.py index 4561014452ac..ba9894284450 100644 --- a/sdk/eventgrid/azure-eventgrid/swagger/_constants.py +++ b/sdk/eventgrid/azure-eventgrid/swagger/_constants.py @@ -2,7 +2,6 @@ ####These files are used to extract the enums - update this list as necessary ############################################################################# files = [ - "https://raw.githubusercontent.com/Azure/azure-rest-api-specs/main/specification/eventgrid/data-plane/Microsoft.Communication/stable/2018-01-01/AzureCommunicationServices.json", "https://raw.githubusercontent.com/Azure/azure-rest-api-specs/main/specification/eventgrid/data-plane/Microsoft.ApiManagement/stable/2018-01-01/APIManagement.json", "https://raw.githubusercontent.com/Azure/azure-rest-api-specs/main/specification/eventgrid/data-plane/Microsoft.AppConfiguration/stable/2018-01-01/AppConfiguration.json", @@ -24,39 +23,39 @@ "https://raw.githubusercontent.com/Azure/azure-rest-api-specs/main/specification/eventgrid/data-plane/Microsoft.Storage/stable/2018-01-01/Storage.json", "https://raw.githubusercontent.com/Azure/azure-rest-api-specs/main/specification/eventgrid/data-plane/Microsoft.Web/stable/2018-01-01/Web.json", "https://raw.githubusercontent.com/Azure/azure-rest-api-specs/main/specification/eventgrid/data-plane/Microsoft.HealthcareApis/stable/2018-01-01/HealthcareApis.json", - ] +] ####################################################### ### Used for backward compatibility. Don't change this ####################################################### backward_compat = { - 'AcsChatMemberAddedToThreadWithUserEventName': "Microsoft.Communication.ChatMemberAddedToThreadWithUser", - 'ResourceWriteFailureEventName': "Microsoft.Resources.ResourceWriteFailure", - 'IoTHubDeviceDeletedEventName': "Microsoft.Devices.DeviceDeleted", - 'IoTHubDeviceDisconnectedEventName': "Microsoft.Devices.DeviceDisconnected", - 'ResourceDeleteFailureEventName': "Microsoft.Resources.ResourceDeleteFailure", - 'ResourceDeleteCancelEventName': "Microsoft.Resources.ResourceDeleteCancel", - 'AcsChatThreadParticipantAddedEventName': "Microsoft.Communication.ChatThreadParticipantAdded", - 'ResourceDeleteSuccessEventName': "Microsoft.Resources.ResourceDeleteSuccess", - 'EventGridSubscriptionValidationEventName': "Microsoft.EventGrid.SubscriptionValidationEvent", - 'ResourceWriteSuccessEventName': "Microsoft.Resources.ResourceWriteSuccess", - 'ResourceActionSuccessEventName': "Microsoft.Resources.ResourceActionSuccess", - 'ResourceWriteCancelEventName': "Microsoft.Resources.ResourceWriteCancel", - 'ResourceActionFailureEventName': "Microsoft.Resources.ResourceActionFailure", - 'AcsChatMemberRemovedFromThreadWithUserEventName': "Microsoft.Communication.ChatMemberRemovedFromThreadWithUser", - 'IoTHubDeviceConnectedEventName': "Microsoft.Devices.DeviceConnected", - 'EventGridSubscriptionDeletedEventName': "Microsoft.EventGrid.SubscriptionDeletedEvent", - 'AcsChatThreadParticipantRemovedEventName': "Microsoft.Communication.ChatThreadParticipantRemoved", - 'ResourceActionCancelEventName': "Microsoft.Resources.ResourceActionCancel", - 'IoTHubDeviceCreatedEventName': "Microsoft.Devices.DeviceCreated", + "AcsChatMemberAddedToThreadWithUserEventName": "Microsoft.Communication.ChatMemberAddedToThreadWithUser", + "ResourceWriteFailureEventName": "Microsoft.Resources.ResourceWriteFailure", + "IoTHubDeviceDeletedEventName": "Microsoft.Devices.DeviceDeleted", + "IoTHubDeviceDisconnectedEventName": "Microsoft.Devices.DeviceDisconnected", + "ResourceDeleteFailureEventName": "Microsoft.Resources.ResourceDeleteFailure", + "ResourceDeleteCancelEventName": "Microsoft.Resources.ResourceDeleteCancel", + "AcsChatThreadParticipantAddedEventName": "Microsoft.Communication.ChatThreadParticipantAdded", + "ResourceDeleteSuccessEventName": "Microsoft.Resources.ResourceDeleteSuccess", + "EventGridSubscriptionValidationEventName": "Microsoft.EventGrid.SubscriptionValidationEvent", + "ResourceWriteSuccessEventName": "Microsoft.Resources.ResourceWriteSuccess", + "ResourceActionSuccessEventName": "Microsoft.Resources.ResourceActionSuccess", + "ResourceWriteCancelEventName": "Microsoft.Resources.ResourceWriteCancel", + "ResourceActionFailureEventName": "Microsoft.Resources.ResourceActionFailure", + "AcsChatMemberRemovedFromThreadWithUserEventName": "Microsoft.Communication.ChatMemberRemovedFromThreadWithUser", + "IoTHubDeviceConnectedEventName": "Microsoft.Devices.DeviceConnected", + "EventGridSubscriptionDeletedEventName": "Microsoft.EventGrid.SubscriptionDeletedEvent", + "AcsChatThreadParticipantRemovedEventName": "Microsoft.Communication.ChatThreadParticipantRemoved", + "ResourceActionCancelEventName": "Microsoft.Resources.ResourceActionCancel", + "IoTHubDeviceCreatedEventName": "Microsoft.Devices.DeviceCreated", } additional_events = { - 'ContainerRegistryArtifactEventName': 'Microsoft.AppConfiguration.KeyValueModified', - 'KeyVaultAccessPolicyChangedEventName': 'Microsoft.KeyVault.VaultAccessPolicyChanged', - 'ContainerRegistryEventName': 'Microsoft.ContainerRegistry.ChartPushed', - 'ServiceBusDeadletterMessagesAvailableWithNoListenerEventName': 'Microsoft.ServiceBus.DeadletterMessagesAvailableWithNoListeners' + "ContainerRegistryArtifactEventName": "Microsoft.AppConfiguration.KeyValueModified", + "KeyVaultAccessPolicyChangedEventName": "Microsoft.KeyVault.VaultAccessPolicyChanged", + "ContainerRegistryEventName": "Microsoft.ContainerRegistry.ChartPushed", + "ServiceBusDeadletterMessagesAvailableWithNoListenerEventName": "Microsoft.ServiceBus.DeadletterMessagesAvailableWithNoListeners", } -EXCEPTIONS = ['ContainerRegistryArtifactEventData', 'ContainerRegistryEventData'] +EXCEPTIONS = ["ContainerRegistryArtifactEventData", "ContainerRegistryEventData"] diff --git a/sdk/eventgrid/azure-eventgrid/swagger/postprocess_eventnames.py b/sdk/eventgrid/azure-eventgrid/swagger/postprocess_eventnames.py index d198152e00c1..bb3bf3794f28 100644 --- a/sdk/eventgrid/azure-eventgrid/swagger/postprocess_eventnames.py +++ b/sdk/eventgrid/azure-eventgrid/swagger/postprocess_eventnames.py @@ -6,14 +6,15 @@ from azure.eventgrid._generated import models from _constants import files, backward_compat, additional_events, EXCEPTIONS + def extract(definitions): if not definitions: return tups = [] for event in definitions: - if event.endswith('Data') and event not in EXCEPTIONS: + if event.endswith("Data") and event not in EXCEPTIONS: try: - key, txt = "Name".join(event.rsplit('Data', 1)), definitions[event]['description'] + key, txt = "Name".join(event.rsplit("Data", 1)), definitions[event]["description"] val = re.findall("Microsoft.[a-zA-Z]+.[a-zA-Z]+", txt) tups.append((key, val[0])) except: @@ -21,6 +22,7 @@ def extract(definitions): sys.exit(1) return tups + def generate_enum_content(tuples): print("# These names at the top are 'corrected' aliases of duplicate values that appear below, which are") print("# deprecated but maintained for backwards compatibility.") @@ -32,10 +34,11 @@ def generate_enum_content(tuples): for k, v in additional_events.items(): print(k + " = '" + v + "'\n") + definitions = {} for fp in files: data = json.loads(urlopen(fp).read()) - definitions.update(data.get('definitions')) + definitions.update(data.get("definitions")) tup_list = extract(definitions) tup_list.sort(key=lambda tup: tup[0]) generate_enum_content(tup_list) diff --git a/sdk/eventgrid/azure-eventgrid/tests/_mocks.py b/sdk/eventgrid/azure-eventgrid/tests/_mocks.py index d06c77a30a89..eb4ebc6f19ce 100644 --- a/sdk/eventgrid/azure-eventgrid/tests/_mocks.py +++ b/sdk/eventgrid/azure-eventgrid/tests/_mocks.py @@ -3,35 +3,35 @@ # storage cloud event cloud_storage_dict = { - "id":"a0517898-9fa4-4e70-b4a3-afda1dd68672", - "source":"/subscriptions/{subscription-id}/resourceGroups/{resource-group}/providers/Microsoft.Storage/storageAccounts/{storage-account}", - "data":{ - "api":"PutBlockList", - "client_request_id":"6d79dbfb-0e37-4fc4-981f-442c9ca65760", - "request_id":"831e1650-001e-001b-66ab-eeb76e000000", - "e_tag":"0x8D4BCC2E4835CD0", - "content_type":"application/octet-stream", - "content_length":524288, - "blob_type":"BlockBlob", - "url":"https://oc2d2817345i60006.blob.core.windows.net/oc2d2817345i200097container/oc2d2817345i20002296blob", - "sequencer":"00000000000004420000000000028963", - "storage_diagnostics":{"batchId":"b68529f3-68cd-4744-baa4-3c0498ec19f0"} + "id": "a0517898-9fa4-4e70-b4a3-afda1dd68672", + "source": "/subscriptions/{subscription-id}/resourceGroups/{resource-group}/providers/Microsoft.Storage/storageAccounts/{storage-account}", + "data": { + "api": "PutBlockList", + "client_request_id": "6d79dbfb-0e37-4fc4-981f-442c9ca65760", + "request_id": "831e1650-001e-001b-66ab-eeb76e000000", + "e_tag": "0x8D4BCC2E4835CD0", + "content_type": "application/octet-stream", + "content_length": 524288, + "blob_type": "BlockBlob", + "url": "https://oc2d2817345i60006.blob.core.windows.net/oc2d2817345i200097container/oc2d2817345i20002296blob", + "sequencer": "00000000000004420000000000028963", + "storage_diagnostics": {"batchId": "b68529f3-68cd-4744-baa4-3c0498ec19f0"}, }, - "type":"Microsoft.Storage.BlobCreated", - "time":"2020-08-07T01:11:49.765846Z", - "specversion":"1.0" + "type": "Microsoft.Storage.BlobCreated", + "time": "2020-08-07T01:11:49.765846Z", + "specversion": "1.0", } cloud_storage_string = json.dumps(cloud_storage_dict) cloud_storage_bytes = cloud_storage_string.encode("utf-8") # custom cloud event cloud_custom_dict = { - "id":"de0fd76c-4ef4-4dfb-ab3a-8f24a307e033", - "source":"https://egtest.dev/cloudcustomevent", - "data":{"team": "event grid squad"}, - "type":"Azure.Sdk.Sample", - "time":"2020-08-07T02:06:08.11969Z", - "specversion":"1.0" + "id": "de0fd76c-4ef4-4dfb-ab3a-8f24a307e033", + "source": "https://egtest.dev/cloudcustomevent", + "data": {"team": "event grid squad"}, + "type": "Azure.Sdk.Sample", + "time": "2020-08-07T02:06:08.11969Z", + "specversion": "1.0", } cloud_custom_string = json.dumps(cloud_custom_dict) cloud_custom_bytes = cloud_custom_string.encode("utf-8") @@ -39,25 +39,25 @@ # storage eg event # spell-checker:ignore swpill eventgridegsub egtopicsamplesub eg_storage_dict = { - "id":"bbab6625-dc56-4b22-abeb-afcc72e5290c", - "subject":"/blobServices/default/containers/oc2d2817345i200097container/blobs/oc2d2817345i20002296blob", - "data":{ - "api":"PutBlockList", - "clientRequestId":"6d79dbfb-0e37-4fc4-981f-442c9ca65760", - "requestId":"831e1650-001e-001b-66ab-eeb76e000000", - "eTag":"0x8D4BCC2E4835CD0", - "contentType":"application/octet-stream", - "contentLength":524288, - "blobType":"BlockBlob", - "url":"https://oc2d2817345i60006.blob.core.windows.net/oc2d2817345i200097container/oc2d2817345i20002296blob", - "sequencer":"00000000000004420000000000028963", - "storageDiagnostics":{"batchId":"b68529f3-68cd-4744-baa4-3c0498ec19f0"} + "id": "bbab6625-dc56-4b22-abeb-afcc72e5290c", + "subject": "/blobServices/default/containers/oc2d2817345i200097container/blobs/oc2d2817345i20002296blob", + "data": { + "api": "PutBlockList", + "clientRequestId": "6d79dbfb-0e37-4fc4-981f-442c9ca65760", + "requestId": "831e1650-001e-001b-66ab-eeb76e000000", + "eTag": "0x8D4BCC2E4835CD0", + "contentType": "application/octet-stream", + "contentLength": 524288, + "blobType": "BlockBlob", + "url": "https://oc2d2817345i60006.blob.core.windows.net/oc2d2817345i200097container/oc2d2817345i20002296blob", + "sequencer": "00000000000004420000000000028963", + "storageDiagnostics": {"batchId": "b68529f3-68cd-4744-baa4-3c0498ec19f0"}, }, - "eventType":"Microsoft.Storage.BlobCreated", - "dataVersion":"2.0", - "metadataVersion":"1", - "eventTime":"2020-08-07T02:28:23.867525Z", - "topic":"/subscriptions/faa080af-c1d8-40ad-9cce-e1a450ca5b57/resourceGroups/t-swpill-test/providers/Microsoft.EventGrid/topics/eventgridegsub" + "eventType": "Microsoft.Storage.BlobCreated", + "dataVersion": "2.0", + "metadataVersion": "1", + "eventTime": "2020-08-07T02:28:23.867525Z", + "topic": "/subscriptions/faa080af-c1d8-40ad-9cce-e1a450ca5b57/resourceGroups/t-swpill-test/providers/Microsoft.EventGrid/topics/eventgridegsub", } eg_storage_string = json.dumps(eg_storage_dict) @@ -65,14 +65,14 @@ # custom eg event eg_custom_dict = { - "id":"3a30afef-b604-4b67-973e-7dfff7e178a7", - "subject":"Test EG Custom Event", - "data":{"team":"event grid squad"}, - "eventType":"Azure.Sdk.Sample", - "dataVersion":"2.0", - "metadataVersion":"1", - "eventTime":"2020-08-07T02:19:05.16916Z", - "topic":"/subscriptions/f8aa80ae-d1c8-60ad-9bce-e1a850ba5b67/resourceGroups/sample-resource-group-test/providers/Microsoft.EventGrid/topics/egtopicsamplesub" + "id": "3a30afef-b604-4b67-973e-7dfff7e178a7", + "subject": "Test EG Custom Event", + "data": {"team": "event grid squad"}, + "eventType": "Azure.Sdk.Sample", + "dataVersion": "2.0", + "metadataVersion": "1", + "eventTime": "2020-08-07T02:19:05.16916Z", + "topic": "/subscriptions/f8aa80ae-d1c8-60ad-9bce-e1a850ba5b67/resourceGroups/sample-resource-group-test/providers/Microsoft.EventGrid/topics/egtopicsamplesub", } eg_custom_string = json.dumps(eg_custom_dict) eg_custom_bytes = eg_custom_string.encode("utf-8") diff --git a/sdk/eventgrid/azure-eventgrid/tests/conftest.py b/sdk/eventgrid/azure-eventgrid/tests/conftest.py index c26a3608ff0b..52626e8a7fbb 100644 --- a/sdk/eventgrid/azure-eventgrid/tests/conftest.py +++ b/sdk/eventgrid/azure-eventgrid/tests/conftest.py @@ -27,7 +27,11 @@ import sys import pytest from devtools_testutils import test_proxy -from devtools_testutils.sanitizers import add_remove_header_sanitizer, add_general_regex_sanitizer, set_custom_default_matcher +from devtools_testutils.sanitizers import ( + add_remove_header_sanitizer, + add_general_regex_sanitizer, + set_custom_default_matcher, +) # Ignore async tests for Python < 3.5 collect_ignore_glob = [] @@ -35,6 +39,7 @@ collect_ignore_glob.append("*_async.py") collect_ignore_glob.append("test_cncf*") + @pytest.fixture(scope="session", autouse=True) def add_aeg_sanitizer(test_proxy): # this can be reverted to set_bodiless_matcher() after tests are re-recorded and don't contain these headers @@ -43,6 +48,5 @@ def add_aeg_sanitizer(test_proxy): ) add_remove_header_sanitizer(headers="aeg-sas-key, aeg-sas-token") add_general_regex_sanitizer( - value="fakeresource", - regex="(?<=\\/\\/)[a-z-]+(?=\\.westus2-1\\.eventgrid\\.azure\\.net/api/events)" + value="fakeresource", regex="(?<=\\/\\/)[a-z-]+(?=\\.westus2-1\\.eventgrid\\.azure\\.net/api/events)" ) diff --git a/sdk/eventgrid/azure-eventgrid/tests/eventgrid_preparer.py b/sdk/eventgrid/azure-eventgrid/tests/eventgrid_preparer.py index 90403144b6d1..6c0c22c1fb91 100644 --- a/sdk/eventgrid/azure-eventgrid/tests/eventgrid_preparer.py +++ b/sdk/eventgrid/azure-eventgrid/tests/eventgrid_preparer.py @@ -6,20 +6,28 @@ from devtools_testutils.resource_testcase import RESOURCE_GROUP_PARAM -EVENTGRID_TOPIC_PARAM = 'eventgrid_topic' -EVENTGRID_TOPIC_LOCATION = 'westus' +EVENTGRID_TOPIC_PARAM = "eventgrid_topic" +EVENTGRID_TOPIC_LOCATION = "westus" CLOUD_EVENT_SCHEMA = InputSchema.cloud_event_schema_v1_0 CUSTOM_EVENT_SCHEMA = InputSchema.custom_event_schema -ID_JSON_FIELD = JsonField(source_field='customId') -TOPIC_JSON_FIELD = JsonField(source_field='customTopic') -EVENT_TIME_JSON_FIELD = JsonField(source_field='customEventTime') -EVENT_TYPE_JSON_FIELD_WITH_DEFAULT = JsonFieldWithDefault(source_field='customEventType', default_value='') -SUBJECT_JSON_FIELD_WITH_DEFAULT = JsonFieldWithDefault(source_field='customSubject', default_value='') -DATA_VERSION_JSON_FIELD_WITH_DEFAULT = JsonFieldWithDefault(source_field='customDataVersion', default_value='') -CUSTOM_JSON_INPUT_SCHEMA_MAPPING = JsonInputSchemaMapping(id=ID_JSON_FIELD, topic=TOPIC_JSON_FIELD, event_time=EVENT_TIME_JSON_FIELD, event_type=EVENT_TYPE_JSON_FIELD_WITH_DEFAULT, subject=SUBJECT_JSON_FIELD_WITH_DEFAULT, data_version=DATA_VERSION_JSON_FIELD_WITH_DEFAULT) +ID_JSON_FIELD = JsonField(source_field="customId") +TOPIC_JSON_FIELD = JsonField(source_field="customTopic") +EVENT_TIME_JSON_FIELD = JsonField(source_field="customEventTime") +EVENT_TYPE_JSON_FIELD_WITH_DEFAULT = JsonFieldWithDefault(source_field="customEventType", default_value="") +SUBJECT_JSON_FIELD_WITH_DEFAULT = JsonFieldWithDefault(source_field="customSubject", default_value="") +DATA_VERSION_JSON_FIELD_WITH_DEFAULT = JsonFieldWithDefault(source_field="customDataVersion", default_value="") +CUSTOM_JSON_INPUT_SCHEMA_MAPPING = JsonInputSchemaMapping( + id=ID_JSON_FIELD, + topic=TOPIC_JSON_FIELD, + event_time=EVENT_TIME_JSON_FIELD, + event_type=EVENT_TYPE_JSON_FIELD_WITH_DEFAULT, + subject=SUBJECT_JSON_FIELD_WITH_DEFAULT, + data_version=DATA_VERSION_JSON_FIELD_WITH_DEFAULT, +) EventGridPreparer = functools.partial( - PowerShellPreparer, "eventgrid", + PowerShellPreparer, + "eventgrid", eventgrid_topic_endpoint="https://fakeresource.westus2-1.eventgrid.azure.net/api/events", eventgrid_topic_key="fakekeyfakekeyfakekeyfakekeyfakekeyfakekeyA=", eventgrid_domain_endpoint="https://fakeresource.westus2-1.eventgrid.azure.net/api/events", diff --git a/sdk/eventgrid/azure-eventgrid/tests/perfstress_tests/send.py b/sdk/eventgrid/azure-eventgrid/tests/perfstress_tests/send.py index d7cb6011f696..6d78bf1d0c81 100644 --- a/sdk/eventgrid/azure-eventgrid/tests/perfstress_tests/send.py +++ b/sdk/eventgrid/azure-eventgrid/tests/perfstress_tests/send.py @@ -1,8 +1,8 @@ -#------------------------------------------------------------------------- +# ------------------------------------------------------------------------- # Copyright (c) Microsoft Corporation. All rights reserved. # Licensed under the MIT License. See License.txt in the project root for # license information. -#-------------------------------------------------------------------------- +# -------------------------------------------------------------------------- import asyncio from azure_devtools.perfstress_tests import PerfStressTest @@ -12,6 +12,7 @@ from azure.core.credentials import AzureKeyCredential + class EventGridPerfTest(PerfStressTest): def __init__(self, arguments): super().__init__(arguments) @@ -21,29 +22,23 @@ def __init__(self, arguments): endpoint = self.get_from_env("EG_TOPIC_HOSTNAME") # Create clients - self.publisher_client = SyncPublisherClient( - endpoint=endpoint, - credential=AzureKeyCredential(topic_key) - ) - self.async_publisher_client = AsyncPublisherClient( - endpoint=endpoint, - credential=AzureKeyCredential(topic_key) - ) + self.publisher_client = SyncPublisherClient(endpoint=endpoint, credential=AzureKeyCredential(topic_key)) + self.async_publisher_client = AsyncPublisherClient(endpoint=endpoint, credential=AzureKeyCredential(topic_key)) self.event_list = [] for _ in range(self.args.num_events): - self.event_list.append(EventGridEvent( - event_type="Contoso.Items.ItemReceived", - data={ - "services": ["EventGrid", "ServiceBus", "EventHubs", "Storage"] - }, - subject="Door1", - data_version="2.0" - )) + self.event_list.append( + EventGridEvent( + event_type="Contoso.Items.ItemReceived", + data={"services": ["EventGrid", "ServiceBus", "EventHubs", "Storage"]}, + subject="Door1", + data_version="2.0", + ) + ) async def close(self): """This is run after cleanup. - + Use this to close any open handles or clients. """ await self.async_publisher_client.close() @@ -51,7 +46,7 @@ async def close(self): def run_sync(self): """The synchronous perf test. - + Try to keep this minimal and focused. Using only a single client API. Avoid putting any ancillary logic (e.g. generating UUIDs), and put this in the setup/init instead so that we're only measuring the client API call. @@ -60,7 +55,7 @@ def run_sync(self): async def run_async(self): """The asynchronous perf test. - + Try to keep this minimal and focused. Using only a single client API. Avoid putting any ancillary logic (e.g. generating UUIDs), and put this in the setup/init instead so that we're only measuring the client API call. @@ -70,4 +65,6 @@ async def run_async(self): @staticmethod def add_arguments(parser): super(EventGridPerfTest, EventGridPerfTest).add_arguments(parser) - parser.add_argument('-n', '--num-events', nargs='?', type=int, help='Number of events to be sent. Defaults to 100', default=100) + parser.add_argument( + "-n", "--num-events", nargs="?", type=int, help="Number of events to be sent. Defaults to 100", default=100 + ) diff --git a/sdk/eventgrid/azure-eventgrid/tests/test_cloud_event_tracing.py b/sdk/eventgrid/azure-eventgrid/tests/test_cloud_event_tracing.py index 0aad89a0d898..39ad0605f76e 100644 --- a/sdk/eventgrid/azure-eventgrid/tests/test_cloud_event_tracing.py +++ b/sdk/eventgrid/azure-eventgrid/tests/test_cloud_event_tracing.py @@ -1,67 +1,63 @@ -#------------------------------------------------------------------------- +# ------------------------------------------------------------------------- # Copyright (c) Microsoft Corporation. All rights reserved. # Licensed under the MIT License. See License.txt in the project root for # license information. -#-------------------------------------------------------------------------- +# -------------------------------------------------------------------------- import pytest import json -from azure.core.pipeline import ( - PipelineRequest, - PipelineContext -) +from azure.core.pipeline import PipelineRequest, PipelineContext from azure.core.pipeline.transport import HttpRequest from azure.core.messaging import CloudEvent from azure.eventgrid._policies import CloudEventDistributedTracingPolicy -from _mocks import ( - cloud_storage_dict -) +from _mocks import cloud_storage_dict + # spell-checker:disable _content_type = "application/cloudevents-batch+json; charset=utf-8" _traceparent_value = "00-4bf92f3577b34da6a3ce929d0e0e4736-00f067aa0ba902b7-01" _tracestate_value = "rojo=00-4bf92f3577b34da6a3ce929d0e0e4736-00f067aa0ba902b7-01,congo=lZWRzIHRoNhcm5hbCBwbGVhc3VyZS4" # spell-chcker:enable -class EventGridSerializationTests(object): +class EventGridSerializationTests(object): def test_cloud_event_policy_copies(self): policy = CloudEventDistributedTracingPolicy() body = json.dumps([cloud_storage_dict]) - universal_request = HttpRequest('POST', 'http://127.0.0.1/', data=body) - universal_request.headers['content-type'] = _content_type - universal_request.headers['traceparent'] = _traceparent_value - universal_request.headers['tracestate'] = _tracestate_value + universal_request = HttpRequest("POST", "http://127.0.0.1/", data=body) + universal_request.headers["content-type"] = _content_type + universal_request.headers["traceparent"] = _traceparent_value + universal_request.headers["tracestate"] = _tracestate_value request = PipelineRequest(universal_request, PipelineContext(None)) resp = policy.on_request(request) body = json.loads(request.http_request.body) - + for item in body: - assert 'traceparent' in item - assert 'tracestate' in item + assert "traceparent" in item + assert "tracestate" in item def test_cloud_event_policy_no_copy_if_trace_exists(self): policy = CloudEventDistributedTracingPolicy() - cloud_storage_dict.update({'traceparent': 'exists', 'tracestate': 'state_exists'}) + cloud_storage_dict.update({"traceparent": "exists", "tracestate": "state_exists"}) body = json.dumps([cloud_storage_dict]) - universal_request = HttpRequest('POST', 'http://127.0.0.1/', data=body) - universal_request.headers['content-type'] = _content_type - universal_request.headers['traceparent'] = _traceparent_value - universal_request.headers['tracestate'] = _tracestate_value + universal_request = HttpRequest("POST", "http://127.0.0.1/", data=body) + universal_request.headers["content-type"] = _content_type + universal_request.headers["traceparent"] = _traceparent_value + universal_request.headers["tracestate"] = _tracestate_value request = PipelineRequest(universal_request, PipelineContext(None)) resp = policy.on_request(request) body = json.loads(request.http_request.body) - + for item in body: - assert 'traceparent' in item - assert 'tracestate' in item - assert item['traceparent'] == 'exists' - assert item['tracestate'] == 'state_exists' + assert "traceparent" in item + assert "tracestate" in item + assert item["traceparent"] == "exists" + assert item["tracestate"] == "state_exists" diff --git a/sdk/eventgrid/azure-eventgrid/tests/test_cncf_events.py b/sdk/eventgrid/azure-eventgrid/tests/test_cncf_events.py index fb9d753a9f91..0e1e56be6992 100644 --- a/sdk/eventgrid/azure-eventgrid/tests/test_cncf_events.py +++ b/sdk/eventgrid/azure-eventgrid/tests/test_cncf_events.py @@ -1,4 +1,3 @@ - import json from devtools_testutils import AzureRecordedTestCase, recorded_by_proxy @@ -10,6 +9,7 @@ EventGridPreparer, ) + class TestEventGridPublisherClientCncf(AzureRecordedTestCase): def create_eg_publisher_client(self, endpoint): credential = self.get_credential(EventGridPublisherClient) @@ -26,13 +26,14 @@ def test_send_cloud_event_data_dict(self, variables, eventgrid_cloud_event_topic } data = {"message": "Hello World!"} cloud_event = CloudEvent(attributes, data) + def callback(request): req = json.loads(request.http_request.body) assert req[0].get("data") is not None assert isinstance(req[0], dict) assert req[0].get("type") == "com.example.sampletype1" assert req[0].get("source") == "https://example.com/event-producer" - + client.send(cloud_event, raw_request_hook=callback) @EventGridPreparer() @@ -43,13 +44,14 @@ def test_send_cloud_event_data_base64_using_data(self, variables, eventgrid_clou "type": "com.example.sampletype1", "source": "https://example.com/event-producer", } - data = b'hello world' + data = b"hello world" cloud_event = CloudEvent(attributes, data) + def callback(request): req = json.loads(request.http_request.body) assert req[0].get("data_base64") is not None assert req[0].get("data") is None - + client.send(cloud_event, raw_request_hook=callback) @EventGridPreparer() @@ -73,10 +75,12 @@ def test_send_cloud_event_data_str(self, variables, eventgrid_cloud_event_topic_ "source": "https://example.com/event-producer", } data = "hello world" + def callback(request): req = json.loads(request.http_request.body) assert req[0].get("data_base64") is None assert req[0].get("data") is not None + cloud_event = CloudEvent(attributes, data) client.send(cloud_event, raw_request_hook=callback) @@ -99,8 +103,8 @@ def test_send_cloud_event_data_with_extensions(self, variables, eventgrid_cloud_ attributes = { "type": "com.example.sampletype1", "source": "https://example.com/event-producer", - "ext1": "extension" + "ext1": "extension", } data = "hello world" cloud_event = CloudEvent(attributes, data) - client.send([cloud_event]) \ No newline at end of file + client.send([cloud_event]) diff --git a/sdk/eventgrid/azure-eventgrid/tests/test_cncf_events_async.py b/sdk/eventgrid/azure-eventgrid/tests/test_cncf_events_async.py index d2226590c998..9b3d6f285809 100644 --- a/sdk/eventgrid/azure-eventgrid/tests/test_cncf_events_async.py +++ b/sdk/eventgrid/azure-eventgrid/tests/test_cncf_events_async.py @@ -1,4 +1,3 @@ - import json import pytest from devtools_testutils import AzureRecordedTestCase, CachedResourceGroupPreparer @@ -12,6 +11,7 @@ EventGridPreparer, ) + class TestEventGridPublisherClientCncf(AzureRecordedTestCase): def create_eg_publisher_client(self, endpoint): credential = self.get_credential(EventGridPublisherClient, is_async=True) @@ -29,13 +29,14 @@ async def test_send_cloud_event_data_dict(self, variables, eventgrid_cloud_event } data = {"message": "Hello World!"} cloud_event = CloudEvent(attributes, data) + def callback(request): req = json.loads(request.http_request.body) assert req[0].get("data") is not None assert isinstance(req[0], dict) assert req[0].get("type") == "com.example.sampletype1" assert req[0].get("source") == "https://example.com/event-producer" - + await client.send(cloud_event, raw_request_hook=callback) @EventGridPreparer() @@ -47,13 +48,14 @@ async def test_send_cloud_event_data_base64_using_data(self, variables, eventgri "type": "com.example.sampletype1", "source": "https://example.com/event-producer", } - data = b'hello world' + data = b"hello world" cloud_event = CloudEvent(attributes, data) + def callback(request): req = json.loads(request.http_request.body) assert req[0].get("data_base64") is not None assert req[0].get("data") is None - + await client.send(cloud_event, raw_request_hook=callback) @EventGridPreparer() @@ -80,11 +82,12 @@ async def test_send_cloud_event_data_str(self, variables, eventgrid_cloud_event_ } data = "hello world" cloud_event = CloudEvent(attributes, data) + def callback(request): req = json.loads(request.http_request.body) assert req[0].get("data_base64") is None assert req[0].get("data") is not None - + await client.send(cloud_event, raw_request_hook=callback) @EventGridPreparer() @@ -108,7 +111,7 @@ async def test_send_cloud_event_data_with_extensions(self, variables, eventgrid_ attributes = { "type": "com.example.sampletype1", "source": "https://example.com/event-producer", - "ext1": "extension" + "ext1": "extension", } data = "hello world" cloud_event = CloudEvent(attributes, data) diff --git a/sdk/eventgrid/azure-eventgrid/tests/test_eg_event_get_bytes.py b/sdk/eventgrid/azure-eventgrid/tests/test_eg_event_get_bytes.py index 1ae80ecde2c9..9a701caf2f1c 100644 --- a/sdk/eventgrid/azure-eventgrid/tests/test_eg_event_get_bytes.py +++ b/sdk/eventgrid/azure-eventgrid/tests/test_eg_event_get_bytes.py @@ -9,6 +9,7 @@ from azure.eventgrid._messaging_shared import _get_json_content from azure.eventgrid import EventGridEvent + class MockQueueMessage(object): def __init__(self, content=None): self.id = uuid.uuid4() @@ -19,31 +20,33 @@ def __init__(self, content=None): self.pop_receipt = None self.next_visible_on = None + class MockServiceBusReceivedMessage(object): def __init__(self, body=None, **kwargs): - self.body=body - self.application_properties=None - self.session_id=None - self.message_id='3f6c5441-5be5-4f33-80c3-3ffeb6a090ce' - self.content_type='application/cloudevents+json; charset=utf-8' - self.correlation_id=None - self.to=None - self.reply_to=None - self.reply_to_session_id=None - self.subject=None - self.time_to_live=datetime.timedelta(days=14) - self.partition_key=None - self.scheduled_enqueue_time_utc=None - self.auto_renew_error=None, - self.dead_letter_error_description=None - self.dead_letter_reason=None - self.dead_letter_source=None - self.delivery_count=13 - self.enqueued_sequence_number=0 - self.enqueued_time_utc=datetime.datetime(2021, 7, 22, 22, 27, 41, 236000) - self.expires_at_utc=datetime.datetime(2021, 8, 5, 22, 27, 41, 236000) - self.sequence_number=11219 - self.lock_token='233146e3-d5a6-45eb-826f-691d82fb8b13' + self.body = body + self.application_properties = None + self.session_id = None + self.message_id = "3f6c5441-5be5-4f33-80c3-3ffeb6a090ce" + self.content_type = "application/cloudevents+json; charset=utf-8" + self.correlation_id = None + self.to = None + self.reply_to = None + self.reply_to_session_id = None + self.subject = None + self.time_to_live = datetime.timedelta(days=14) + self.partition_key = None + self.scheduled_enqueue_time_utc = None + self.auto_renew_error = (None,) + self.dead_letter_error_description = None + self.dead_letter_reason = None + self.dead_letter_source = None + self.delivery_count = 13 + self.enqueued_sequence_number = 0 + self.enqueued_time_utc = datetime.datetime(2021, 7, 22, 22, 27, 41, 236000) + self.expires_at_utc = datetime.datetime(2021, 8, 5, 22, 27, 41, 236000) + self.sequence_number = 11219 + self.lock_token = "233146e3-d5a6-45eb-826f-691d82fb8b13" + class MockEventhubData(object): def __init__(self, body=None): @@ -53,7 +56,7 @@ def __init__(self, body=None): raise ValueError("EventData cannot be None.") # Internal usage only for transforming AmqpAnnotatedMessage to outgoing EventData - self.body=body + self.body = body self._raw_amqp_message = "some amqp data" self.message_id = None self.content_type = None @@ -66,7 +69,7 @@ def __init__(self, data=None): def __iter__(self): return self - + def __next__(self): if not self.data: return """{"id":"f208feff-099b-4bda-a341-4afd0fa02fef","subject":"https://egsample.dev/sampleevent","data":"ServiceBus","event_type":"Azure.Sdk.Sample","event_time":"2021-07-22T22:27:38.960209Z","data_version":"1.0"}""" @@ -81,14 +84,15 @@ def __init__(self, data=None): def __iter__(self): return self - + def __next__(self): if not self.data: return b'[{"id":"f208feff-099b-4bda-a341-4afd0fa02fef","subject":"https://egsample.dev/sampleevent","data":"Eventhub","event_type":"Azure.Sdk.Sample","event_time":"2021-07-22T22:27:38.960209Z","data_version":"1.0"}]' return self.data - + next = __next__ + def test_get_bytes_storage_queue(): cloud_storage_dict = """{ "id":"a0517898-9fa4-4e70-b4a3-afda1dd68672", @@ -112,73 +116,73 @@ def test_get_bytes_storage_queue(): obj = MockQueueMessage(content=cloud_storage_dict) dict = _get_json_content(obj) - assert dict.get('data') == { - "api":"PutBlockList", - "client_request_id":"6d79dbfb-0e37-4fc4-981f-442c9ca65760", - "request_id":"831e1650-001e-001b-66ab-eeb76e000000", - "e_tag":"0x8D4BCC2E4835CD0", - "content_type":"application/octet-stream", - "content_length":524288, - "blob_type":"BlockBlob", - "url":"https://oc2d2817345i60006.blob.core.windows.net/oc2d2817345i200097container/oc2d2817345i20002296blob", - "sequencer":"00000000000004420000000000028963", - "storage_diagnostics":{"batchId":"b68529f3-68cd-4744-baa4-3c0498ec19f0"} - } - assert dict.get('data_version') == "1.0" + assert dict.get("data") == { + "api": "PutBlockList", + "client_request_id": "6d79dbfb-0e37-4fc4-981f-442c9ca65760", + "request_id": "831e1650-001e-001b-66ab-eeb76e000000", + "e_tag": "0x8D4BCC2E4835CD0", + "content_type": "application/octet-stream", + "content_length": 524288, + "blob_type": "BlockBlob", + "url": "https://oc2d2817345i60006.blob.core.windows.net/oc2d2817345i200097container/oc2d2817345i20002296blob", + "sequencer": "00000000000004420000000000028963", + "storage_diagnostics": {"batchId": "b68529f3-68cd-4744-baa4-3c0498ec19f0"}, + } + assert dict.get("data_version") == "1.0" + def test_get_bytes_storage_queue_wrong_content(): - string = u'This is a random string which must fail' + string = "This is a random string which must fail" obj = MockQueueMessage(content=string) with pytest.raises(ValueError, match="Failed to load JSON content from the object."): _get_json_content(obj) + def test_get_bytes_servicebus(): obj = MockServiceBusReceivedMessage( body=MockBody(), - message_id='3f6c5441-5be5-4f33-80c3-3ffeb6a090ce', - content_type='application/cloudevents+json; charset=utf-8', + message_id="3f6c5441-5be5-4f33-80c3-3ffeb6a090ce", + content_type="application/cloudevents+json; charset=utf-8", time_to_live=datetime.timedelta(days=14), delivery_count=13, enqueued_sequence_number=0, enqueued_time_utc=datetime.datetime(2021, 7, 22, 22, 27, 41, 236000), expires_at_utc=datetime.datetime(2021, 8, 5, 22, 27, 41, 236000), sequence_number=11219, - lock_token='233146e3-d5a6-45eb-826f-691d82fb8b13' + lock_token="233146e3-d5a6-45eb-826f-691d82fb8b13", ) dict = _get_json_content(obj) - assert dict.get('data') == "ServiceBus" - assert dict.get('data_version') == '1.0' + assert dict.get("data") == "ServiceBus" + assert dict.get("data_version") == "1.0" + def test_get_bytes_servicebus_wrong_content(): obj = MockServiceBusReceivedMessage( - body=MockBody(data='random'), - message_id='3f6c5441-5be5-4f33-80c3-3ffeb6a090ce', - content_type='application/json; charset=utf-8', + body=MockBody(data="random"), + message_id="3f6c5441-5be5-4f33-80c3-3ffeb6a090ce", + content_type="application/json; charset=utf-8", time_to_live=datetime.timedelta(days=14), delivery_count=13, enqueued_sequence_number=0, enqueued_time_utc=datetime.datetime(2021, 7, 22, 22, 27, 41, 236000), expires_at_utc=datetime.datetime(2021, 8, 5, 22, 27, 41, 236000), sequence_number=11219, - lock_token='233146e3-d5a6-45eb-826f-691d82fb8b13' + lock_token="233146e3-d5a6-45eb-826f-691d82fb8b13", ) with pytest.raises(ValueError, match="Failed to load JSON content from the object."): dict = _get_json_content(obj) def test_get_bytes_eventhubs(): - obj = MockEventhubData( - body=MockEhBody() - ) + obj = MockEventhubData(body=MockEhBody()) dict = _get_json_content(obj) - assert dict.get('data') == 'Eventhub' - assert dict.get('data_version') == '1.0' + assert dict.get("data") == "Eventhub" + assert dict.get("data_version") == "1.0" + def test_get_bytes_eventhubs_wrong_content(): - obj = MockEventhubData( - body=MockEhBody(data='random string') - ) + obj = MockEventhubData(body=MockEhBody(data="random string")) with pytest.raises(ValueError, match="Failed to load JSON content from the object."): dict = _get_json_content(obj) @@ -186,43 +190,44 @@ def test_get_bytes_eventhubs_wrong_content(): def test_get_bytes_random_obj(): json_str = '{"id": "de0fd76c-4ef4-4dfb-ab3a-8f24a307e033", "subject": "https://egtest.dev/cloudcustomevent", "data": {"team": "event grid squad"}, "event_type": "Azure.Sdk.Sample", "event_time": "2020-08-07T02:06:08.11969Z", "data_version": "1.0"}' - random_obj = { - "id":"de0fd76c-4ef4-4dfb-ab3a-8f24a307e033", - "subject":"https://egtest.dev/cloudcustomevent", - "data":{"team": "event grid squad"}, - "event_type":"Azure.Sdk.Sample", - "event_time":"2020-08-07T02:06:08.11969Z", - "data_version":"1.0", + random_obj = { + "id": "de0fd76c-4ef4-4dfb-ab3a-8f24a307e033", + "subject": "https://egtest.dev/cloudcustomevent", + "data": {"team": "event grid squad"}, + "event_type": "Azure.Sdk.Sample", + "event_time": "2020-08-07T02:06:08.11969Z", + "data_version": "1.0", } assert _get_json_content(json_str) == random_obj + def test_from_json_sb(): obj = MockServiceBusReceivedMessage( body=MockBody(), - message_id='3f6c5441-5be5-4f33-80c3-3ffeb6a090ce', - content_type='application/cloudevents+json; charset=utf-8', + message_id="3f6c5441-5be5-4f33-80c3-3ffeb6a090ce", + content_type="application/cloudevents+json; charset=utf-8", time_to_live=datetime.timedelta(days=14), delivery_count=13, enqueued_sequence_number=0, enqueued_time_utc=datetime.datetime(2021, 7, 22, 22, 27, 41, 236000), expires_at_utc=datetime.datetime(2021, 8, 5, 22, 27, 41, 236000), sequence_number=11219, - lock_token='233146e3-d5a6-45eb-826f-691d82fb8b13' + lock_token="233146e3-d5a6-45eb-826f-691d82fb8b13", ) event = EventGridEvent.from_json(obj) assert event.id == "f208feff-099b-4bda-a341-4afd0fa02fef" assert event.data == "ServiceBus" + def test_from_json_eh(): - obj = MockEventhubData( - body=MockEhBody() - ) + obj = MockEventhubData(body=MockEhBody()) event = EventGridEvent.from_json(obj) assert event.id == "f208feff-099b-4bda-a341-4afd0fa02fef" assert event.data == "Eventhub" + def test_from_json_storage(): eg_storage_dict = """{ "id":"a0517898-9fa4-4e70-b4a3-afda1dd68672", @@ -246,17 +251,17 @@ def test_from_json_storage(): obj = MockQueueMessage(content=eg_storage_dict) event = EventGridEvent.from_json(obj) assert event.data == { - "api":"PutBlockList", - "client_request_id":"6d79dbfb-0e37-4fc4-981f-442c9ca65760", - "request_id":"831e1650-001e-001b-66ab-eeb76e000000", - "e_tag":"0x8D4BCC2E4835CD0", - "content_type":"application/octet-stream", - "content_length":524288, - "blob_type":"BlockBlob", - "url":"https://oc2d2817345i60006.blob.core.windows.net/oc2d2817345i200097container/oc2d2817345i20002296blob", - "sequencer":"00000000000004420000000000028963", - "storage_diagnostics":{"batchId":"b68529f3-68cd-4744-baa4-3c0498ec19f0"} - } + "api": "PutBlockList", + "client_request_id": "6d79dbfb-0e37-4fc4-981f-442c9ca65760", + "request_id": "831e1650-001e-001b-66ab-eeb76e000000", + "e_tag": "0x8D4BCC2E4835CD0", + "content_type": "application/octet-stream", + "content_length": 524288, + "blob_type": "BlockBlob", + "url": "https://oc2d2817345i60006.blob.core.windows.net/oc2d2817345i200097container/oc2d2817345i20002296blob", + "sequencer": "00000000000004420000000000028963", + "storage_diagnostics": {"batchId": "b68529f3-68cd-4744-baa4-3c0498ec19f0"}, + } def test_from_json(): diff --git a/sdk/eventgrid/azure-eventgrid/tests/test_eg_publisher_client.py b/sdk/eventgrid/azure-eventgrid/tests/test_eg_publisher_client.py index 5bb81aaa9a0d..7314f68f199b 100644 --- a/sdk/eventgrid/azure-eventgrid/tests/test_eg_publisher_client.py +++ b/sdk/eventgrid/azure-eventgrid/tests/test_eg_publisher_client.py @@ -1,8 +1,8 @@ -#------------------------------------------------------------------------- +# ------------------------------------------------------------------------- # Copyright (c) Microsoft Corporation. All rights reserved. # Licensed under the MIT License. See License.txt in the project root for # license information. -#-------------------------------------------------------------------------- +# -------------------------------------------------------------------------- import logging import sys @@ -19,7 +19,7 @@ except ImportError: from urlparse import urlparse -from devtools_testutils import AzureRecordedTestCase, recorded_by_proxy +from devtools_testutils import AzureRecordedTestCase, recorded_by_proxy from azure.core.credentials import AzureKeyCredential, AzureSasCredential from azure.core.messaging import CloudEvent @@ -31,6 +31,7 @@ EventGridPreparer, ) + class TestEventGridPublisherClient(AzureRecordedTestCase): def create_eg_publisher_client(self, endpoint): credential = self.get_credential(EventGridPublisherClient) @@ -42,25 +43,21 @@ def create_eg_publisher_client(self, endpoint): def test_send_event_grid_event_data_dict(self, variables, eventgrid_topic_endpoint): client = self.create_eg_publisher_client(eventgrid_topic_endpoint) eg_event = EventGridEvent( - subject="sample", - data={"sample": "eventgridevent"}, - event_type="Sample.EventGrid.Event", - data_version="2.0" - ) + subject="sample", data={"sample": "eventgridevent"}, event_type="Sample.EventGrid.Event", data_version="2.0" + ) client.send(eg_event) @EventGridPreparer() @recorded_by_proxy - def test_send_event_grid_event_fails_without_full_url(self, variables, eventgrid_topic_key, eventgrid_topic_endpoint): + def test_send_event_grid_event_fails_without_full_url( + self, variables, eventgrid_topic_key, eventgrid_topic_endpoint + ): akc_credential = AzureKeyCredential(eventgrid_topic_key) parsed_url = urlparse(eventgrid_topic_endpoint) client = EventGridPublisherClient(parsed_url.netloc, akc_credential) eg_event = EventGridEvent( - subject="sample", - data={"sample": "eventgridevent"}, - event_type="Sample.EventGrid.Event", - data_version="2.0" - ) + subject="sample", data={"sample": "eventgridevent"}, event_type="Sample.EventGrid.Event", data_version="2.0" + ) with pytest.raises(ValueError): client.send(eg_event) @@ -69,17 +66,11 @@ def test_send_event_grid_event_fails_without_full_url(self, variables, eventgrid def test_send_event_grid_event_data_as_list(self, variables, eventgrid_topic_endpoint): client = self.create_eg_publisher_client(eventgrid_topic_endpoint) eg_event1 = EventGridEvent( - subject="sample", - data=u"eventgridevent", - event_type="Sample.EventGrid.Event", - data_version="2.0" - ) + subject="sample", data="eventgridevent", event_type="Sample.EventGrid.Event", data_version="2.0" + ) eg_event2 = EventGridEvent( - subject="sample2", - data=u"eventgridevent2", - event_type="Sample.EventGrid.Event", - data_version="2.0" - ) + subject="sample2", data="eventgridevent2", event_type="Sample.EventGrid.Event", data_version="2.0" + ) client.send([eg_event1, eg_event2]) @EventGridPreparer() @@ -87,11 +78,8 @@ def test_send_event_grid_event_data_as_list(self, variables, eventgrid_topic_end def test_send_event_grid_event_data_str(self, variables, eventgrid_topic_endpoint): client = self.create_eg_publisher_client(eventgrid_topic_endpoint) eg_event = EventGridEvent( - subject="sample", - data=u"eventgridevent", - event_type="Sample.EventGrid.Event", - data_version="2.0" - ) + subject="sample", data="eventgridevent", event_type="Sample.EventGrid.Event", data_version="2.0" + ) client.send(eg_event) @EventGridPreparer() @@ -99,11 +87,8 @@ def test_send_event_grid_event_data_str(self, variables, eventgrid_topic_endpoin def test_send_event_grid_event_data_bytes(self, variables, eventgrid_topic_endpoint): client = self.create_eg_publisher_client(eventgrid_topic_endpoint) eg_event = EventGridEvent( - subject="sample", - data=b"eventgridevent", - event_type="Sample.EventGrid.Event", - data_version="2.0" - ) + subject="sample", data=b"eventgridevent", event_type="Sample.EventGrid.Event", data_version="2.0" + ) with pytest.raises(TypeError, match="Data in EventGridEvent cannot be bytes*"): client.send(eg_event) @@ -112,12 +97,12 @@ def test_send_event_grid_event_data_bytes(self, variables, eventgrid_topic_endpo def test_send_event_grid_event_dict_data_bytes(self, variables, eventgrid_topic_endpoint): client = self.create_eg_publisher_client(eventgrid_topic_endpoint) eg_event = { - "subject":"sample", - "data":b"eventgridevent", - "eventType":"Sample.EventGrid.Event", - "dataVersion":"2.0", - "id": uuid.uuid4(), - "eventTime": datetime.now() + "subject": "sample", + "data": b"eventgridevent", + "eventType": "Sample.EventGrid.Event", + "dataVersion": "2.0", + "id": uuid.uuid4(), + "eventTime": datetime.now(), } with pytest.raises(TypeError, match="Data in EventGridEvent cannot be bytes*"): client.send(eg_event) @@ -127,16 +112,15 @@ def test_send_event_grid_event_dict_data_bytes(self, variables, eventgrid_topic_ def test_send_event_grid_event_dict_data_dict(self, variables, eventgrid_topic_endpoint): client = self.create_eg_publisher_client(eventgrid_topic_endpoint) eg_event = { - "subject":"sample", - "data":{"key1": "Sample.EventGrid.Event"}, - "eventType":"Sample.EventGrid.Event", - "dataVersion":"2.0", - "id": uuid.uuid4(), - "eventTime": datetime.now() + "subject": "sample", + "data": {"key1": "Sample.EventGrid.Event"}, + "eventType": "Sample.EventGrid.Event", + "dataVersion": "2.0", + "id": uuid.uuid4(), + "eventTime": datetime.now(), } client.send(eg_event) - ### CLOUD EVENT TESTS @EventGridPreparer() @@ -144,10 +128,8 @@ def test_send_event_grid_event_dict_data_dict(self, variables, eventgrid_topic_e def test_send_cloud_event_data_dict(self, variables, eventgrid_cloud_event_topic_endpoint): client = self.create_eg_publisher_client(eventgrid_cloud_event_topic_endpoint) cloud_event = CloudEvent( - source = "http://samplesource.dev", - data = {"sample": "cloudevent"}, - type="Sample.Cloud.Event" - ) + source="http://samplesource.dev", data={"sample": "cloudevent"}, type="Sample.Cloud.Event" + ) client.send(cloud_event) @pytest.mark.skip("https://github.com/Azure/azure-sdk-for-python/issues/16993") @@ -155,12 +137,8 @@ def test_send_cloud_event_data_dict(self, variables, eventgrid_cloud_event_topic @recorded_by_proxy def test_send_cloud_event_data_NULL(self, variables, eventgrid_cloud_event_topic_endpoint): client = self.create_eg_publisher_client(eventgrid_cloud_event_topic_endpoint) - cloud_event = CloudEvent( - source = "http://samplesource.dev", - data = NULL, - type="Sample.Cloud.Event" - ) - + cloud_event = CloudEvent(source="http://samplesource.dev", data=NULL, type="Sample.Cloud.Event") + def callback(request): req = json.loads(request.http_request.body) assert req[0].get("data") is None @@ -171,11 +149,7 @@ def callback(request): @recorded_by_proxy def test_send_cloud_event_data_base64_using_data(self, variables, eventgrid_cloud_event_topic_endpoint): client = self.create_eg_publisher_client(eventgrid_cloud_event_topic_endpoint) - cloud_event = CloudEvent( - source = "http://samplesource.dev", - data = b'cloudevent', - type="Sample.Cloud.Event" - ) + cloud_event = CloudEvent(source="http://samplesource.dev", data=b"cloudevent", type="Sample.Cloud.Event") def callback(request): req = json.loads(request.http_request.body) @@ -187,54 +161,38 @@ def callback(request): def test_send_cloud_event_fails_on_providing_data_and_b64(self): with pytest.raises(ValueError, match="Unexpected keyword arguments data_base64.*"): cloud_event = CloudEvent( - source = "http://samplesource.dev", - data_base64 = b'cloudevent', - data = "random data", - type="Sample.Cloud.Event" - ) + source="http://samplesource.dev", + data_base64=b"cloudevent", + data="random data", + type="Sample.Cloud.Event", + ) @EventGridPreparer() @recorded_by_proxy def test_send_cloud_event_data_none(self, variables, eventgrid_cloud_event_topic_endpoint): client = self.create_eg_publisher_client(eventgrid_cloud_event_topic_endpoint) - cloud_event = CloudEvent( - source = "http://samplesource.dev", - data = None, - type="Sample.Cloud.Event" - ) + cloud_event = CloudEvent(source="http://samplesource.dev", data=None, type="Sample.Cloud.Event") client.send(cloud_event) @EventGridPreparer() @recorded_by_proxy def test_send_cloud_event_data_str(self, variables, eventgrid_cloud_event_topic_endpoint): client = self.create_eg_publisher_client(eventgrid_cloud_event_topic_endpoint) - cloud_event = CloudEvent( - source = "http://samplesource.dev", - data = "cloudevent", - type="Sample.Cloud.Event" - ) + cloud_event = CloudEvent(source="http://samplesource.dev", data="cloudevent", type="Sample.Cloud.Event") client.send(cloud_event) @EventGridPreparer() @recorded_by_proxy def test_send_cloud_event_data_bytes(self, variables, eventgrid_cloud_event_topic_endpoint): client = self.create_eg_publisher_client(eventgrid_cloud_event_topic_endpoint) - cloud_event = CloudEvent( - source = "http://samplesource.dev", - data = b"cloudevent", - type="Sample.Cloud.Event" - ) + cloud_event = CloudEvent(source="http://samplesource.dev", data=b"cloudevent", type="Sample.Cloud.Event") client.send(cloud_event) @EventGridPreparer() @recorded_by_proxy def test_send_cloud_event_data_as_list(self, variables, eventgrid_cloud_event_topic_endpoint): client = self.create_eg_publisher_client(eventgrid_cloud_event_topic_endpoint) - cloud_event = CloudEvent( - source = "http://samplesource.dev", - data = "cloudevent", - type="Sample.Cloud.Event" - ) + cloud_event = CloudEvent(source="http://samplesource.dev", data="cloudevent", type="Sample.Cloud.Event") client.send([cloud_event]) @EventGridPreparer() @@ -242,30 +200,27 @@ def test_send_cloud_event_data_as_list(self, variables, eventgrid_cloud_event_to def test_send_cloud_event_data_with_extensions(self, variables, eventgrid_cloud_event_topic_endpoint): client = self.create_eg_publisher_client(eventgrid_cloud_event_topic_endpoint) cloud_event = CloudEvent( - source = "http://samplesource.dev", - data = "cloudevent", - type="Sample.Cloud.Event", - extensions={ - 'reasoncode':204, - 'extension':'hello' - } - ) + source="http://samplesource.dev", + data="cloudevent", + type="Sample.Cloud.Event", + extensions={"reasoncode": 204, "extension": "hello"}, + ) client.send([cloud_event]) internal = _cloud_event_to_generated(cloud_event).serialize() - assert 'reasoncode' in internal - assert 'extension' in internal - assert internal['reasoncode'] == 204 + assert "reasoncode" in internal + assert "extension" in internal + assert internal["reasoncode"] == 204 @EventGridPreparer() @recorded_by_proxy def test_send_cloud_event_dict(self, variables, eventgrid_cloud_event_topic_endpoint): client = self.create_eg_publisher_client(eventgrid_cloud_event_topic_endpoint) cloud_event1 = { - "id": "1234", - "source": "http://samplesource.dev", - "specversion": "1.0", - "data": "cloudevent", - "type": "Sample.Cloud.Event" + "id": "1234", + "source": "http://samplesource.dev", + "specversion": "1.0", + "data": "cloudevent", + "type": "Sample.Cloud.Event", } client.send(cloud_event1) @@ -277,11 +232,8 @@ def test_send_signature_credential(self, variables, eventgrid_topic_key, eventgr credential = AzureSasCredential(signature) client = EventGridPublisherClient(eventgrid_topic_endpoint, credential) eg_event = EventGridEvent( - subject="sample", - data={"sample": "eventgridevent"}, - event_type="Sample.EventGrid.Event", - data_version="2.0" - ) + subject="sample", data={"sample": "eventgridevent"}, event_type="Sample.EventGrid.Event", data_version="2.0" + ) client.send(eg_event) @EventGridPreparer() @@ -289,19 +241,19 @@ def test_send_signature_credential(self, variables, eventgrid_topic_key, eventgr def test_send_NONE_credential(self, variables, eventgrid_topic_endpoint): with pytest.raises(ValueError, match="Parameter 'self._credential' must not be None."): client = EventGridPublisherClient(eventgrid_topic_endpoint, None) - + @EventGridPreparer() @recorded_by_proxy def test_send_custom_schema_event(self, variables, eventgrid_custom_event_topic_endpoint): client = self.create_eg_publisher_client(eventgrid_custom_event_topic_endpoint) custom_event = { - "customSubject": "sample", - "customEventType": "sample.event", - "customDataVersion": "2.0", - "customId": "1234", - "customEventTime": dt.datetime.now(UTC()).isoformat(), - "customData": "sample data" - } + "customSubject": "sample", + "customEventType": "sample.event", + "customDataVersion": "2.0", + "customId": "1234", + "customEventTime": dt.datetime.now(UTC()).isoformat(), + "customData": "sample data", + } client.send(custom_event) @EventGridPreparer() @@ -309,26 +261,29 @@ def test_send_custom_schema_event(self, variables, eventgrid_custom_event_topic_ def test_send_custom_schema_event_as_list(self, variables, eventgrid_custom_event_topic_endpoint): client = self.create_eg_publisher_client(eventgrid_custom_event_topic_endpoint) custom_event1 = { - "customSubject": "sample", - "customEventType": "sample.event", - "customDataVersion": "2.0", - "customId": "1234", - "customEventTime": dt.datetime.now(UTC()).isoformat(), - "customData": "sample data" - } + "customSubject": "sample", + "customEventType": "sample.event", + "customDataVersion": "2.0", + "customId": "1234", + "customEventTime": dt.datetime.now(UTC()).isoformat(), + "customData": "sample data", + } custom_event2 = { - "customSubject": "sample2", - "customEventType": "sample.event", - "customDataVersion": "2.0", - "customId": "12345", - "customEventTime": dt.datetime.now(UTC()).isoformat(), - "customData": "sample data 2" - } + "customSubject": "sample2", + "customEventType": "sample.event", + "customDataVersion": "2.0", + "customId": "12345", + "customEventTime": dt.datetime.now(UTC()).isoformat(), + "customData": "sample data 2", + } client.send([custom_event1, custom_event2]) def test_send_throws_with_bad_credential(self): bad_credential = "I am a bad credential" - with pytest.raises(ValueError, match="The provided credential should be an instance of a TokenCredential, AzureSasCredential or AzureKeyCredential"): + with pytest.raises( + ValueError, + match="The provided credential should be an instance of a TokenCredential, AzureSasCredential or AzureKeyCredential", + ): client = EventGridPublisherClient("eventgrid_endpoint", bad_credential) @pytest.mark.live_test_only @@ -338,28 +293,21 @@ def test_send_token_credential(self, variables, eventgrid_topic_endpoint): credential = self.get_credential(EventGridPublisherClient) client = EventGridPublisherClient(eventgrid_topic_endpoint, credential) eg_event = EventGridEvent( - subject="sample", - data={"sample": "eventgridevent"}, - event_type="Sample.EventGrid.Event", - data_version="2.0" - ) + subject="sample", data={"sample": "eventgridevent"}, event_type="Sample.EventGrid.Event", data_version="2.0" + ) client.send(eg_event) @pytest.mark.live_test_only @EventGridPreparer() @recorded_by_proxy def test_send_partner_namespace(self, variables): - eventgrid_partner_namespace_endpoint = os.environ['EVENTGRID_PARTNER_NAMESPACE_TOPIC_ENDPOINT'] - eventgrid_partner_namespace_key = os.environ['EVENTGRID_PARTNER_NAMESPACE_TOPIC_KEY'] - channel_name = os.environ['EVENTGRID_PARTNER_CHANNEL_NAME'] + eventgrid_partner_namespace_endpoint = os.environ["EVENTGRID_PARTNER_NAMESPACE_TOPIC_ENDPOINT"] + eventgrid_partner_namespace_key = os.environ["EVENTGRID_PARTNER_NAMESPACE_TOPIC_KEY"] + channel_name = os.environ["EVENTGRID_PARTNER_CHANNEL_NAME"] credential = AzureKeyCredential(eventgrid_partner_namespace_key) client = EventGridPublisherClient(eventgrid_partner_namespace_endpoint, eventgrid_partner_namespace_key) - cloud_event = CloudEvent( - source = "http://samplesource.dev", - data = "cloudevent", - type="Sample.Cloud.Event" - ) - + cloud_event = CloudEvent(source="http://samplesource.dev", data="cloudevent", type="Sample.Cloud.Event") + def callback(request): req = json.loads(request.http_request.headers) assert req.get("aeg-channel-name") == channel_name diff --git a/sdk/eventgrid/azure-eventgrid/tests/test_eg_publisher_client_async.py b/sdk/eventgrid/azure-eventgrid/tests/test_eg_publisher_client_async.py index 220d63807c87..21af8f1cf6cd 100644 --- a/sdk/eventgrid/azure-eventgrid/tests/test_eg_publisher_client_async.py +++ b/sdk/eventgrid/azure-eventgrid/tests/test_eg_publisher_client_async.py @@ -1,8 +1,8 @@ -#------------------------------------------------------------------------- +# ------------------------------------------------------------------------- # Copyright (c) Microsoft Corporation. All rights reserved. # Licensed under the MIT License. See License.txt in the project root for # license information. -#-------------------------------------------------------------------------- +# -------------------------------------------------------------------------- import logging import asyncio @@ -25,9 +25,7 @@ from azure.eventgrid.aio import EventGridPublisherClient from azure.eventgrid._helpers import _cloud_event_to_generated -from eventgrid_preparer import ( - EventGridPreparer -) +from eventgrid_preparer import EventGridPreparer class TestEventGridPublisherClient(AzureRecordedTestCase): @@ -42,46 +40,35 @@ def create_eg_publisher_client(self, endpoint): async def test_send_event_grid_event_data_dict(self, variables, eventgrid_topic_endpoint): client = self.create_eg_publisher_client(eventgrid_topic_endpoint) eg_event = EventGridEvent( - subject="sample", - data={"sample": "eventgridevent"}, - event_type="Sample.EventGrid.Event", - data_version="2.0" - ) + subject="sample", data={"sample": "eventgridevent"}, event_type="Sample.EventGrid.Event", data_version="2.0" + ) await client.send(eg_event) - @EventGridPreparer() @recorded_by_proxy_async @pytest.mark.asyncio async def test_send_event_grid_event_data_as_list(self, variables, eventgrid_topic_endpoint): client = self.create_eg_publisher_client(eventgrid_topic_endpoint) eg_event1 = EventGridEvent( - subject="sample", - data="eventgridevent", - event_type="Sample.EventGrid.Event", - data_version="2.0" - ) + subject="sample", data="eventgridevent", event_type="Sample.EventGrid.Event", data_version="2.0" + ) eg_event2 = EventGridEvent( - subject="sample2", - data="eventgridevent2", - event_type="Sample.EventGrid.Event", - data_version="2.0" - ) + subject="sample2", data="eventgridevent2", event_type="Sample.EventGrid.Event", data_version="2.0" + ) await client.send([eg_event1, eg_event2]) @EventGridPreparer() @recorded_by_proxy_async @pytest.mark.asyncio - async def test_send_event_grid_event_fails_without_full_url(self, variables, eventgrid_topic_key, eventgrid_topic_endpoint): + async def test_send_event_grid_event_fails_without_full_url( + self, variables, eventgrid_topic_key, eventgrid_topic_endpoint + ): akc_credential = AzureKeyCredential(eventgrid_topic_key) parsed_url = urlparse(eventgrid_topic_endpoint) client = EventGridPublisherClient(parsed_url.netloc, akc_credential) eg_event = EventGridEvent( - subject="sample", - data={"sample": "eventgridevent"}, - event_type="Sample.EventGrid.Event", - data_version="2.0" - ) + subject="sample", data={"sample": "eventgridevent"}, event_type="Sample.EventGrid.Event", data_version="2.0" + ) with pytest.raises(ValueError): await client.send(eg_event) @@ -91,11 +78,8 @@ async def test_send_event_grid_event_fails_without_full_url(self, variables, eve async def test_send_event_grid_event_data_str(self, variables, eventgrid_topic_endpoint): client = self.create_eg_publisher_client(eventgrid_topic_endpoint) eg_event = EventGridEvent( - subject="sample", - data="eventgridevent", - event_type="Sample.EventGrid.Event", - data_version="2.0" - ) + subject="sample", data="eventgridevent", event_type="Sample.EventGrid.Event", data_version="2.0" + ) await client.send(eg_event) @EventGridPreparer() @@ -104,11 +88,8 @@ async def test_send_event_grid_event_data_str(self, variables, eventgrid_topic_e async def test_send_event_grid_event_data_bytes(self, variables, eventgrid_topic_endpoint): client = self.create_eg_publisher_client(eventgrid_topic_endpoint) eg_event = EventGridEvent( - subject="sample", - data=b"eventgridevent", - event_type="Sample.EventGrid.Event", - data_version="2.0" - ) + subject="sample", data=b"eventgridevent", event_type="Sample.EventGrid.Event", data_version="2.0" + ) with pytest.raises(TypeError, match="Data in EventGridEvent cannot be bytes*"): await client.send(eg_event) @@ -118,12 +99,12 @@ async def test_send_event_grid_event_data_bytes(self, variables, eventgrid_topic async def test_send_event_grid_event_dict_data_bytes(self, variables, eventgrid_topic_endpoint): client = self.create_eg_publisher_client(eventgrid_topic_endpoint) eg_event = { - "subject":"sample", - "data":b"eventgridevent", - "eventType":"Sample.EventGrid.Event", - "dataVersion":"2.0", - "id": "123-ddf-133-324255ffd", - "eventTime": dt.datetime.utcnow() + "subject": "sample", + "data": b"eventgridevent", + "eventType": "Sample.EventGrid.Event", + "dataVersion": "2.0", + "id": "123-ddf-133-324255ffd", + "eventTime": dt.datetime.utcnow(), } with pytest.raises(TypeError, match="Data in EventGridEvent cannot be bytes*"): await client.send(eg_event) @@ -134,23 +115,16 @@ async def test_send_event_grid_event_dict_data_bytes(self, variables, eventgrid_ async def test_send_cloud_event_data_dict(self, variables, eventgrid_cloud_event_topic_endpoint): client = self.create_eg_publisher_client(eventgrid_cloud_event_topic_endpoint) cloud_event = CloudEvent( - source = "http://samplesource.dev", - data = {"sample": "cloudevent"}, - type="Sample.Cloud.Event" - ) + source="http://samplesource.dev", data={"sample": "cloudevent"}, type="Sample.Cloud.Event" + ) await client.send(cloud_event) - @EventGridPreparer() @recorded_by_proxy_async @pytest.mark.asyncio async def test_send_cloud_event_data_str(self, variables, eventgrid_cloud_event_topic_endpoint): client = self.create_eg_publisher_client(eventgrid_cloud_event_topic_endpoint) - cloud_event = CloudEvent( - source = "http://samplesource.dev", - data = "cloudevent", - type="Sample.Cloud.Event" - ) + cloud_event = CloudEvent(source="http://samplesource.dev", data="cloudevent", type="Sample.Cloud.Event") await client.send(cloud_event) @EventGridPreparer() @@ -158,11 +132,7 @@ async def test_send_cloud_event_data_str(self, variables, eventgrid_cloud_event_ @pytest.mark.asyncio async def test_send_cloud_event_data_bytes(self, variables, eventgrid_cloud_event_topic_endpoint): client = self.create_eg_publisher_client(eventgrid_cloud_event_topic_endpoint) - cloud_event = CloudEvent( - source = "http://samplesource.dev", - data = b"cloudevent", - type="Sample.Cloud.Event" - ) + cloud_event = CloudEvent(source="http://samplesource.dev", data=b"cloudevent", type="Sample.Cloud.Event") await client.send(cloud_event) @EventGridPreparer() @@ -170,34 +140,25 @@ async def test_send_cloud_event_data_bytes(self, variables, eventgrid_cloud_even @pytest.mark.asyncio async def test_send_cloud_event_data_as_list(self, variables, eventgrid_cloud_event_topic_endpoint): client = self.create_eg_publisher_client(eventgrid_cloud_event_topic_endpoint) - cloud_event = CloudEvent( - source = "http://samplesource.dev", - data = "cloudevent", - type="Sample.Cloud.Event" - ) + cloud_event = CloudEvent(source="http://samplesource.dev", data="cloudevent", type="Sample.Cloud.Event") await client.send([cloud_event]) - @EventGridPreparer() @recorded_by_proxy_async @pytest.mark.asyncio async def test_send_cloud_event_data_with_extensions(self, variables, eventgrid_cloud_event_topic_endpoint): client = self.create_eg_publisher_client(eventgrid_cloud_event_topic_endpoint) cloud_event = CloudEvent( - source = "http://samplesource.dev", - data = "cloudevent", - type="Sample.Cloud.Event", - extensions={ - 'reasoncode':204, - 'extension':'hello' - } - ) + source="http://samplesource.dev", + data="cloudevent", + type="Sample.Cloud.Event", + extensions={"reasoncode": 204, "extension": "hello"}, + ) await client.send([cloud_event]) internal = _cloud_event_to_generated(cloud_event).serialize() - assert 'reasoncode' in internal - assert 'extension' in internal - assert internal['reasoncode'] == 204 - + assert "reasoncode" in internal + assert "extension" in internal + assert internal["reasoncode"] == 204 @EventGridPreparer() @recorded_by_proxy_async @@ -205,11 +166,11 @@ async def test_send_cloud_event_data_with_extensions(self, variables, eventgrid_ async def test_send_cloud_event_dict(self, variables, eventgrid_cloud_event_topic_endpoint): client = self.create_eg_publisher_client(eventgrid_cloud_event_topic_endpoint) cloud_event1 = { - "id": "1234", - "source": "http://samplesource.dev", - "specversion": "1.0", - "data": "cloudevent", - "type": "Sample.Cloud.Event" + "id": "1234", + "source": "http://samplesource.dev", + "specversion": "1.0", + "data": "cloudevent", + "type": "Sample.Cloud.Event", } await client.send(cloud_event1) @@ -218,11 +179,7 @@ async def test_send_cloud_event_dict(self, variables, eventgrid_cloud_event_topi @pytest.mark.asyncio async def test_send_cloud_event_data_none(self, variables, eventgrid_cloud_event_topic_endpoint): client = self.create_eg_publisher_client(eventgrid_cloud_event_topic_endpoint) - cloud_event = CloudEvent( - source = "http://samplesource.dev", - data = None, - type="Sample.Cloud.Event" - ) + cloud_event = CloudEvent(source="http://samplesource.dev", data=None, type="Sample.Cloud.Event") await client.send(cloud_event) @pytest.mark.skip("https://github.com/Azure/azure-sdk-for-python/issues/16993") @@ -231,11 +188,8 @@ async def test_send_cloud_event_data_none(self, variables, eventgrid_cloud_event @pytest.mark.asyncio async def test_send_cloud_event_data_NULL(self, variables, eventgrid_cloud_event_topic_endpoint): client = self.create_eg_publisher_client(eventgrid_cloud_event_topic_endpoint) - cloud_event = CloudEvent( - source = "http://samplesource.dev", - data = NULL, - type="Sample.Cloud.Event" - ) + cloud_event = CloudEvent(source="http://samplesource.dev", data=NULL, type="Sample.Cloud.Event") + def callback(request): req = json.loads(request.http_request.body) assert req[0].get("data") is None @@ -251,51 +205,46 @@ async def test_send_signature_credential(self, variables, eventgrid_topic_key, e credential = AzureSasCredential(signature) client = EventGridPublisherClient(eventgrid_topic_endpoint, credential) eg_event = EventGridEvent( - subject="sample", - data={"sample": "eventgridevent"}, - event_type="Sample.EventGrid.Event", - data_version="2.0" - ) + subject="sample", data={"sample": "eventgridevent"}, event_type="Sample.EventGrid.Event", data_version="2.0" + ) await client.send(eg_event) - @EventGridPreparer() @recorded_by_proxy_async @pytest.mark.asyncio async def test_send_custom_schema_event(self, variables, eventgrid_custom_event_topic_endpoint): client = self.create_eg_publisher_client(eventgrid_custom_event_topic_endpoint) custom_event = { - "customSubject": "sample", - "customEventType": "sample.event", - "customDataVersion": "2.0", - "customId": "1234", - "customEventTime": dt.datetime.now(UTC()).isoformat(), - "customData": "sample data" - } + "customSubject": "sample", + "customEventType": "sample.event", + "customDataVersion": "2.0", + "customId": "1234", + "customEventTime": dt.datetime.now(UTC()).isoformat(), + "customData": "sample data", + } await client.send(custom_event) - @EventGridPreparer() @recorded_by_proxy_async @pytest.mark.asyncio async def test_send_custom_schema_event_as_list(self, variables, eventgrid_custom_event_topic_endpoint): client = self.create_eg_publisher_client(eventgrid_custom_event_topic_endpoint) custom_event1 = { - "customSubject": "sample", - "customEventType": "sample.event", - "customDataVersion": "2.0", - "customId": "1234", - "customEventTime": dt.datetime.now(UTC()).isoformat(), - "customData": "sample data" - } + "customSubject": "sample", + "customEventType": "sample.event", + "customDataVersion": "2.0", + "customId": "1234", + "customEventTime": dt.datetime.now(UTC()).isoformat(), + "customData": "sample data", + } custom_event2 = { - "customSubject": "sample2", - "customEventType": "sample.event", - "customDataVersion": "2.0", - "customId": "12345", - "customEventTime": dt.datetime.now(UTC()).isoformat(), - "customData": "sample data 2" - } + "customSubject": "sample2", + "customEventType": "sample.event", + "customDataVersion": "2.0", + "customId": "12345", + "customEventTime": dt.datetime.now(UTC()).isoformat(), + "customData": "sample data 2", + } await client.send([custom_event1, custom_event2]) @EventGridPreparer() @@ -303,12 +252,8 @@ async def test_send_custom_schema_event_as_list(self, variables, eventgrid_custo @pytest.mark.asyncio async def test_send_and_close_async_session(self, variables, eventgrid_cloud_event_topic_endpoint): client = self.create_eg_publisher_client(eventgrid_cloud_event_topic_endpoint) - async with client: # this throws if client can't close - cloud_event = CloudEvent( - source = "http://samplesource.dev", - data = "cloudevent", - type="Sample.Cloud.Event" - ) + async with client: # this throws if client can't close + cloud_event = CloudEvent(source="http://samplesource.dev", data="cloudevent", type="Sample.Cloud.Event") await client.send(cloud_event) @pytest.mark.skip() @@ -326,27 +271,21 @@ async def test_send_token_credential(self, variables, eventgrid_topic_endpoint): credential = self.get_credential(EventGridPublisherClient) client = EventGridPublisherClient(eventgrid_topic_endpoint, credential) eg_event = EventGridEvent( - subject="sample", - data={"sample": "eventgridevent"}, - event_type="Sample.EventGrid.Event", - data_version="2.0" - ) + subject="sample", data={"sample": "eventgridevent"}, event_type="Sample.EventGrid.Event", data_version="2.0" + ) await client.send(eg_event) @pytest.mark.live_test_only @EventGridPreparer() @recorded_by_proxy_async async def test_send_partner_namespace(self, variables): - eventgrid_partner_namespace_endpoint = os.environ['EVENTGRID_PARTNER_NAMESPACE_TOPIC_ENDPOINT'] - eventgrid_partner_namespace_key = os.environ['EVENTGRID_PARTNER_NAMESPACE_TOPIC_KEY'] - channel_name = os.environ['EVENTGRID_PARTNER_CHANNEL_NAME'] + eventgrid_partner_namespace_endpoint = os.environ["EVENTGRID_PARTNER_NAMESPACE_TOPIC_ENDPOINT"] + eventgrid_partner_namespace_key = os.environ["EVENTGRID_PARTNER_NAMESPACE_TOPIC_KEY"] + channel_name = os.environ["EVENTGRID_PARTNER_CHANNEL_NAME"] credential = AzureKeyCredential(eventgrid_partner_namespace_key) client = EventGridPublisherClient(eventgrid_partner_namespace_endpoint, eventgrid_partner_namespace_key) - cloud_event = CloudEvent( - source = "http://samplesource.dev", - data = "cloudevent", - type="Sample.Cloud.Event" - ) + cloud_event = CloudEvent(source="http://samplesource.dev", data="cloudevent", type="Sample.Cloud.Event") + def callback(request): req = json.loads(request.http_request.headers) assert req.get("aeg-channel-name") == channel_name diff --git a/sdk/eventgrid/azure-eventgrid/tests/test_exceptions.py b/sdk/eventgrid/azure-eventgrid/tests/test_exceptions.py index 25026fa5594f..575a22a79471 100644 --- a/sdk/eventgrid/azure-eventgrid/tests/test_exceptions.py +++ b/sdk/eventgrid/azure-eventgrid/tests/test_exceptions.py @@ -1,8 +1,8 @@ -#------------------------------------------------------------------------- +# ------------------------------------------------------------------------- # Copyright (c) Microsoft Corporation. All rights reserved. # Licensed under the MIT License. See License.txt in the project root for # license information. -#-------------------------------------------------------------------------- +# -------------------------------------------------------------------------- import logging import sys @@ -11,11 +11,7 @@ import pytest import uuid from datetime import datetime, timedelta -from azure.core.exceptions import ( - HttpResponseError, - ClientAuthenticationError, - ServiceRequestError -) +from azure.core.exceptions import HttpResponseError, ClientAuthenticationError, ServiceRequestError from msrest.serialization import UTC import datetime as dt @@ -37,6 +33,7 @@ EventGridPreparer, ) + class TestEventGridPublisherClientExceptions(AzureMgmtRecordedTestCase): def create_eg_publisher_client(self, endpoint): credential = self.get_credential(EventGridPublisherClient) @@ -49,11 +46,8 @@ def test_raise_on_auth_error(self, variables, eventgrid_topic_endpoint): akc_credential = AzureKeyCredential("bad credential") client = EventGridPublisherClient(eventgrid_topic_endpoint, akc_credential) eg_event = EventGridEvent( - subject="sample", - data={"sample": "eventgridevent"}, - event_type="Sample.EventGrid.Event", - data_version="2.0" - ) + subject="sample", data={"sample": "eventgridevent"}, event_type="Sample.EventGrid.Event", data_version="2.0" + ) with pytest.raises(ClientAuthenticationError, match="The request authorization key is not authorized for*"): client.send(eg_event) @@ -61,13 +55,12 @@ def test_raise_on_auth_error(self, variables, eventgrid_topic_endpoint): @recorded_by_proxy def test_raise_on_bad_resource(self, variables, eventgrid_topic_key): akc_credential = AzureKeyCredential(eventgrid_topic_key) - client = EventGridPublisherClient("https://bad-resource.westus-1.eventgrid.azure.net/api/events", akc_credential) + client = EventGridPublisherClient( + "https://bad-resource.westus-1.eventgrid.azure.net/api/events", akc_credential + ) eg_event = EventGridEvent( - subject="sample", - data={"sample": "eventgridevent"}, - event_type="Sample.EventGrid.Event", - data_version="2.0" - ) + subject="sample", data={"sample": "eventgridevent"}, event_type="Sample.EventGrid.Event", data_version="2.0" + ) with pytest.raises(HttpResponseError): client.send(eg_event) @@ -76,15 +69,10 @@ def test_raise_on_bad_resource(self, variables, eventgrid_topic_key): def test_raise_on_large_payload(self, variables, eventgrid_topic_endpoint): client = self.create_eg_publisher_client(eventgrid_topic_endpoint) - path = os.path.abspath(os.path.join(os.path.abspath(__file__), "..", "./large_data.json")) + path = os.path.abspath(os.path.join(os.path.abspath(__file__), "..", "./large_data.json")) with open(path) as json_file: data = json.load(json_file) - eg_event = EventGridEvent( - subject="sample", - data=data, - event_type="Sample.EventGrid.Event", - data_version="2.0" - ) + eg_event = EventGridEvent(subject="sample", data=data, event_type="Sample.EventGrid.Event", data_version="2.0") with pytest.raises(HttpResponseError) as err: client.send(eg_event) assert "The maximum size (1536000) has been exceeded." in err.value.message diff --git a/sdk/eventgrid/azure-eventgrid/tests/test_exceptions_async.py b/sdk/eventgrid/azure-eventgrid/tests/test_exceptions_async.py index d7285e54f36c..366eb384158e 100644 --- a/sdk/eventgrid/azure-eventgrid/tests/test_exceptions_async.py +++ b/sdk/eventgrid/azure-eventgrid/tests/test_exceptions_async.py @@ -1,8 +1,8 @@ -#------------------------------------------------------------------------- +# ------------------------------------------------------------------------- # Copyright (c) Microsoft Corporation. All rights reserved. # Licensed under the MIT License. See License.txt in the project root for # license information. -#-------------------------------------------------------------------------- +# -------------------------------------------------------------------------- import logging import sys @@ -35,6 +35,7 @@ EventGridPreparer, ) + class TestEventGridPublisherClientExceptionsAsync(AzureRecordedTestCase): def create_eg_publisher_client(self, endpoint): credential = self.get_credential(EventGridPublisherClient, is_async=True) @@ -48,11 +49,8 @@ async def test_raise_on_auth_error(self, variables, eventgrid_topic_endpoint): akc_credential = AzureKeyCredential("bad credential") client = EventGridPublisherClient(eventgrid_topic_endpoint, akc_credential) eg_event = EventGridEvent( - subject="sample", - data={"sample": "eventgridevent"}, - event_type="Sample.EventGrid.Event", - data_version="2.0" - ) + subject="sample", data={"sample": "eventgridevent"}, event_type="Sample.EventGrid.Event", data_version="2.0" + ) with pytest.raises(ClientAuthenticationError, match="The request authorization key is not authorized for*"): await client.send(eg_event) @@ -61,13 +59,12 @@ async def test_raise_on_auth_error(self, variables, eventgrid_topic_endpoint): @pytest.mark.asyncio async def test_raise_on_bad_resource(self, variables, eventgrid_topic_key): akc_credential = AzureKeyCredential(eventgrid_topic_key) - client = EventGridPublisherClient("https://bad-resource.westus-1.eventgrid.azure.net/api/events", akc_credential) + client = EventGridPublisherClient( + "https://bad-resource.westus-1.eventgrid.azure.net/api/events", akc_credential + ) eg_event = EventGridEvent( - subject="sample", - data={"sample": "eventgridevent"}, - event_type="Sample.EventGrid.Event", - data_version="2.0" - ) + subject="sample", data={"sample": "eventgridevent"}, event_type="Sample.EventGrid.Event", data_version="2.0" + ) with pytest.raises(HttpResponseError): await client.send(eg_event) @@ -77,15 +74,10 @@ async def test_raise_on_bad_resource(self, variables, eventgrid_topic_key): async def test_raise_on_large_payload(self, variables, eventgrid_topic_endpoint): client = self.create_eg_publisher_client(eventgrid_topic_endpoint) - path = os.path.abspath(os.path.join(os.path.abspath(__file__), "..", "./large_data.json")) + path = os.path.abspath(os.path.join(os.path.abspath(__file__), "..", "./large_data.json")) with open(path) as json_file: data = json.load(json_file) - eg_event = EventGridEvent( - subject="sample", - data=data, - event_type="Sample.EventGrid.Event", - data_version="2.0" - ) + eg_event = EventGridEvent(subject="sample", data=data, event_type="Sample.EventGrid.Event", data_version="2.0") with pytest.raises(HttpResponseError) as err: await client.send(eg_event) assert "The maximum size (1536000) has been exceeded." in err.value.message diff --git a/sdk/eventgrid/azure-eventgrid/tests/test_serialization.py b/sdk/eventgrid/azure-eventgrid/tests/test_serialization.py index b5b74349afc5..9a191f847290 100644 --- a/sdk/eventgrid/azure-eventgrid/tests/test_serialization.py +++ b/sdk/eventgrid/azure-eventgrid/tests/test_serialization.py @@ -1,8 +1,8 @@ -#------------------------------------------------------------------------- +# ------------------------------------------------------------------------- # Copyright (c) Microsoft Corporation. All rights reserved. # Licensed under the MIT License. See License.txt in the project root for # license information. -#-------------------------------------------------------------------------- +# -------------------------------------------------------------------------- import logging import sys @@ -22,94 +22,83 @@ cloud_storage_dict, cloud_storage_string, cloud_storage_bytes, - ) +) -class EventGridSerializationTests(AzureMgmtTestCase): +class EventGridSerializationTests(AzureMgmtTestCase): def _assert_cloud_event_serialized(self, expected, actual): - assert expected['source'] == actual['source'] - assert expected['type'] == actual['type'] - assert actual['specversion'] == '1.0' - assert 'id' in actual - assert 'time' in actual + assert expected["source"] == actual["source"] + assert expected["type"] == actual["type"] + assert actual["specversion"] == "1.0" + assert "id" in actual + assert "time" in actual # Cloud Event tests def test_cloud_event_serialization_extension_bytes(self, **kwargs): data = b"cloudevent" cloud_event = CloudEvent( - source="http://samplesource.dev", - data=data, - type="Sample.Cloud.Event", - extensions={'e1':1, 'e2':2} - ) - - cloud_event.subject = "subject" # to test explicit setting of prop - encoded = base64.b64encode(data).decode('utf-8') + source="http://samplesource.dev", data=data, type="Sample.Cloud.Event", extensions={"e1": 1, "e2": 2} + ) + + cloud_event.subject = "subject" # to test explicit setting of prop + encoded = base64.b64encode(data).decode("utf-8") internal = _cloud_event_to_generated(cloud_event) assert internal.additional_properties is not None - assert 'e1' in internal.additional_properties + assert "e1" in internal.additional_properties - json = internal.serialize() + json = internal.serialize() expected = { - 'source':'http://samplesource.dev', - 'data_base64': encoded, - 'type':'Sample.Cloud.Event', - 'reason_code':204, - 'e1':1, - 'e2':2 + "source": "http://samplesource.dev", + "data_base64": encoded, + "type": "Sample.Cloud.Event", + "reason_code": 204, + "e1": 1, + "e2": 2, } self._assert_cloud_event_serialized(expected, json) - assert expected['data_base64'] == json['data_base64'] - + assert expected["data_base64"] == json["data_base64"] def test_cloud_event_serialization_extension_string(self, **kwargs): data = "cloudevent" cloud_event = CloudEvent( - source="http://samplesource.dev", - data=data, - type="Sample.Cloud.Event", - extensions={'e1':1, 'e2':2} - ) - - cloud_event.subject = "subject" # to test explicit setting of prop + source="http://samplesource.dev", data=data, type="Sample.Cloud.Event", extensions={"e1": 1, "e2": 2} + ) + + cloud_event.subject = "subject" # to test explicit setting of prop internal = _cloud_event_to_generated(cloud_event) assert internal.additional_properties is not None - assert 'e1' in internal.additional_properties + assert "e1" in internal.additional_properties - json = internal.serialize() + json = internal.serialize() expected = { - 'source':'http://samplesource.dev', - 'data': data, - 'type':'Sample.Cloud.Event', - 'reason_code':204, - 'e1':1, - 'e2':2 + "source": "http://samplesource.dev", + "data": data, + "type": "Sample.Cloud.Event", + "reason_code": 204, + "e1": 1, + "e2": 2, } self._assert_cloud_event_serialized(expected, json) if sys.version_info > (3, 5): - assert expected['data'] == json['data'] + assert expected["data"] == json["data"] else: - encoded = base64.b64encode(data).decode('utf-8') - expected['data_base64'] = encoded - assert expected['data_base64'] == json['data_base64'] - assert 'data' not in json + encoded = base64.b64encode(data).decode("utf-8") + expected["data_base64"] = encoded + assert expected["data_base64"] == json["data_base64"] + assert "data" not in json def test_event_grid_event_raises_on_no_data(self): with pytest.raises(TypeError): - eg_event = EventGridEvent( - subject="sample", - event_type="Sample.EventGrid.Event", - data_version="2.0" - ) + eg_event = EventGridEvent(subject="sample", event_type="Sample.EventGrid.Event", data_version="2.0") def test_import_from_system_events(self): - var = SystemEventNames.AcsChatMemberAddedToThreadWithUserEventName + var = SystemEventNames.AcsChatMemberAddedToThreadWithUserEventName assert var == "Microsoft.Communication.ChatMemberAddedToThreadWithUser" assert SystemEventNames.KeyVaultKeyNearExpiryEventName == "Microsoft.KeyVault.KeyNearExpiry" var = SystemEventNames.ServiceBusActiveMessagesAvailableWithNoListenersEventName @@ -121,17 +110,17 @@ def test_import_from_system_events(self): def test_eg_event_repr(self): event = EventGridEvent( - subject="sample2", - data="eventgridevent2", - event_type="Sample.EventGrid.Event", - data_version="2.0" - ) - + subject="sample2", data="eventgridevent2", event_type="Sample.EventGrid.Event", data_version="2.0" + ) + assert "EventGridEvent(subject=sample2" in event.__repr__() def test_servicebus_system_events_alias(self): val = "Microsoft.ServiceBus.DeadletterMessagesAvailableWithNoListeners" - assert SystemEventNames.ServiceBusDeadletterMessagesAvailableWithNoListenerEventName == SystemEventNames.ServiceBusDeadletterMessagesAvailableWithNoListenersEventName + assert ( + SystemEventNames.ServiceBusDeadletterMessagesAvailableWithNoListenerEventName + == SystemEventNames.ServiceBusDeadletterMessagesAvailableWithNoListenersEventName + ) assert SystemEventNames.ServiceBusDeadletterMessagesAvailableWithNoListenerEventName == val assert SystemEventNames.ServiceBusDeadletterMessagesAvailableWithNoListenersEventName == val assert SystemEventNames(val) == SystemEventNames.ServiceBusDeadletterMessagesAvailableWithNoListenerEventName diff --git a/sdk/eventgrid/azure-eventgrid/tsp-location.yaml b/sdk/eventgrid/azure-eventgrid/tsp-location.yaml new file mode 100644 index 000000000000..89d50ac8e4fe --- /dev/null +++ b/sdk/eventgrid/azure-eventgrid/tsp-location.yaml @@ -0,0 +1,4 @@ +cleanup: false +commit: f6f647a40fbbc43dd6e65c4436448129ab7e1fdd +directory: specification/eventgrid/Azure.Messaging.EventGrid +repo: Azure/azure-rest-api-specs