diff --git a/sdk/healthdataaiservices/azure-health-deidentification/CHANGELOG.md b/sdk/healthdataaiservices/azure-health-deidentification/CHANGELOG.md index be2efaec75db..ced1026fefe2 100644 --- a/sdk/healthdataaiservices/azure-health-deidentification/CHANGELOG.md +++ b/sdk/healthdataaiservices/azure-health-deidentification/CHANGELOG.md @@ -1,6 +1,8 @@ # Release History -## 1.0.0b2 (Unreleased) +## 1.0.0b2 (2024-11-15) + +- Stable release of Azure Health Deidentification client library ### Features Added diff --git a/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/__init__.py b/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/__init__.py index 01d9492ed4f6..75df74cc79f6 100644 --- a/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/__init__.py +++ b/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/__init__.py @@ -5,15 +5,21 @@ # Code generated by Microsoft (R) Python Code Generator. # Changes may cause incorrect behavior and will be lost if the code is regenerated. # -------------------------------------------------------------------------- +# pylint: disable=wrong-import-position -from ._client import DeidentificationClient +from typing import TYPE_CHECKING + +if TYPE_CHECKING: + from ._patch import * # pylint: disable=unused-wildcard-import + +from ._client import DeidentificationClient # type: ignore from ._version import VERSION __version__ = VERSION try: from ._patch import __all__ as _patch_all - from ._patch import * # pylint: disable=unused-wildcard-import + from ._patch import * except ImportError: _patch_all = [] from ._patch import patch_sdk as _patch_sdk @@ -21,6 +27,6 @@ __all__ = [ "DeidentificationClient", ] -__all__.extend([p for p in _patch_all if p not in __all__]) +__all__.extend([p for p in _patch_all if p not in __all__]) # pyright: ignore _patch_sdk() diff --git a/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/_client.py b/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/_client.py index 787817c317d9..dcee388d9467 100644 --- a/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/_client.py +++ b/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/_client.py @@ -19,29 +19,25 @@ from ._serialization import Deserializer, Serializer if TYPE_CHECKING: - # pylint: disable=unused-import,ungrouped-imports from azure.core.credentials import TokenCredential -class DeidentificationClient( - DeidentificationClientOperationsMixin -): # pylint: disable=client-accepts-api-version-keyword +class DeidentificationClient(DeidentificationClientOperationsMixin): """DeidentificationClient. :param endpoint: Url of your De-identification Service. Required. :type endpoint: str :param credential: Credential used to authenticate requests to the service. Required. :type credential: ~azure.core.credentials.TokenCredential - :keyword api_version: The API version to use for this operation. Default value is - "2024-07-12-preview". Note that overriding this default value may result in unsupported - behavior. + :keyword api_version: The API version to use for this operation. Default value is "2024-11-15". + Note that overriding this default value may result in unsupported behavior. :paramtype api_version: str :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. """ def __init__(self, endpoint: str, credential: "TokenCredential", **kwargs: Any) -> None: - _endpoint = "https://{endpoint}" + _endpoint = "{endpoint}" self._config = DeidentificationClientConfiguration(endpoint=endpoint, credential=credential, **kwargs) _policies = kwargs.pop("policies", None) if _policies is None: @@ -86,7 +82,7 @@ def send_request(self, request: HttpRequest, *, stream: bool = False, **kwargs: request_copy = deepcopy(request) path_format_arguments = { - "endpoint": self._serialize.url("self._config.endpoint", self._config.endpoint, "str"), + "endpoint": self._serialize.url("self._config.endpoint", self._config.endpoint, "str", skip_quote=True), } request_copy.url = self._client.format_url(request_copy.url, **path_format_arguments) diff --git a/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/_configuration.py b/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/_configuration.py index 35237a99ba8f..1ac729392e01 100644 --- a/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/_configuration.py +++ b/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/_configuration.py @@ -13,11 +13,10 @@ from ._version import VERSION if TYPE_CHECKING: - # pylint: disable=unused-import,ungrouped-imports from azure.core.credentials import TokenCredential -class DeidentificationClientConfiguration: # pylint: disable=too-many-instance-attributes,name-too-long +class DeidentificationClientConfiguration: # pylint: disable=too-many-instance-attributes """Configuration for DeidentificationClient. Note that all parameters used to create this instance are saved as instance @@ -27,14 +26,13 @@ class DeidentificationClientConfiguration: # pylint: disable=too-many-instance- :type endpoint: str :param credential: Credential used to authenticate requests to the service. Required. :type credential: ~azure.core.credentials.TokenCredential - :keyword api_version: The API version to use for this operation. Default value is - "2024-07-12-preview". Note that overriding this default value may result in unsupported - behavior. + :keyword api_version: The API version to use for this operation. Default value is "2024-11-15". + Note that overriding this default value may result in unsupported behavior. :paramtype api_version: str """ def __init__(self, endpoint: str, credential: "TokenCredential", **kwargs: Any) -> None: - api_version: str = kwargs.pop("api_version", "2024-07-12-preview") + api_version: str = kwargs.pop("api_version", "2024-11-15") if endpoint is None: raise ValueError("Parameter 'endpoint' must not be None.") diff --git a/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/_model_base.py b/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/_model_base.py index 43fd8c7e9b1b..e6a2730f9276 100644 --- a/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/_model_base.py +++ b/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/_model_base.py @@ -1,10 +1,11 @@ +# pylint: disable=too-many-lines # coding=utf-8 # -------------------------------------------------------------------------- # Copyright (c) Microsoft Corporation. All rights reserved. # Licensed under the MIT License. See License.txt in the project root for # license information. # -------------------------------------------------------------------------- -# pylint: disable=protected-access, arguments-differ, signature-differs, broad-except +# pylint: disable=protected-access, broad-except import copy import calendar @@ -19,6 +20,7 @@ import email.utils from datetime import datetime, date, time, timedelta, timezone from json import JSONEncoder +import xml.etree.ElementTree as ET from typing_extensions import Self import isodate from azure.core.exceptions import DeserializationError @@ -123,7 +125,7 @@ def _serialize_datetime(o, format: typing.Optional[str] = None): def _is_readonly(p): try: - return p._visibility == ["read"] # pylint: disable=protected-access + return p._visibility == ["read"] except AttributeError: return False @@ -286,6 +288,12 @@ def _deserialize_decimal(attr): return decimal.Decimal(str(attr)) +def _deserialize_int_as_str(attr): + if isinstance(attr, int): + return attr + return int(attr) + + _DESERIALIZE_MAPPING = { datetime: _deserialize_datetime, date: _deserialize_date, @@ -307,9 +315,11 @@ def _deserialize_decimal(attr): def get_deserializer(annotation: typing.Any, rf: typing.Optional["_RestField"] = None): + if annotation is int and rf and rf._format == "str": + return _deserialize_int_as_str if rf and rf._format: return _DESERIALIZE_MAPPING_WITHFORMAT.get(rf._format) - return _DESERIALIZE_MAPPING.get(annotation) + return _DESERIALIZE_MAPPING.get(annotation) # pyright: ignore def _get_type_alias_type(module_name: str, alias_name: str): @@ -441,6 +451,10 @@ def _serialize(o, format: typing.Optional[str] = None): # pylint: disable=too-m return float(o) if isinstance(o, enum.Enum): return o.value + if isinstance(o, int): + if format == "str": + return str(o) + return o try: # First try datetime.datetime return _serialize_datetime(o, format) @@ -471,11 +485,16 @@ def _create_value(rf: typing.Optional["_RestField"], value: typing.Any) -> typin return value if rf._is_model: return _deserialize(rf._type, value) + if isinstance(value, ET.Element): + value = _deserialize(rf._type, value) return _serialize(value, rf._format) class Model(_MyMutableMapping): _is_model = True + # label whether current class's _attr_to_rest_field has been calculated + # could not see _attr_to_rest_field directly because subclass inherits it from parent class + _calculated: typing.Set[str] = set() def __init__(self, *args: typing.Any, **kwargs: typing.Any) -> None: class_name = self.__class__.__name__ @@ -486,10 +505,58 @@ def __init__(self, *args: typing.Any, **kwargs: typing.Any) -> None: for rest_field in self._attr_to_rest_field.values() if rest_field._default is not _UNSET } - if args: - dict_to_pass.update( - {k: _create_value(_get_rest_field(self._attr_to_rest_field, k), v) for k, v in args[0].items()} - ) + if args: # pylint: disable=too-many-nested-blocks + if isinstance(args[0], ET.Element): + existed_attr_keys = [] + model_meta = getattr(self, "_xml", {}) + + for rf in self._attr_to_rest_field.values(): + prop_meta = getattr(rf, "_xml", {}) + xml_name = prop_meta.get("name", rf._rest_name) + xml_ns = prop_meta.get("ns", model_meta.get("ns", None)) + if xml_ns: + xml_name = "{" + xml_ns + "}" + xml_name + + # attribute + if prop_meta.get("attribute", False) and args[0].get(xml_name) is not None: + existed_attr_keys.append(xml_name) + dict_to_pass[rf._rest_name] = _deserialize(rf._type, args[0].get(xml_name)) + continue + + # unwrapped element is array + if prop_meta.get("unwrapped", False): + # unwrapped array could either use prop items meta/prop meta + if prop_meta.get("itemsName"): + xml_name = prop_meta.get("itemsName") + xml_ns = prop_meta.get("itemNs") + if xml_ns: + xml_name = "{" + xml_ns + "}" + xml_name + items = args[0].findall(xml_name) # pyright: ignore + if len(items) > 0: + existed_attr_keys.append(xml_name) + dict_to_pass[rf._rest_name] = _deserialize(rf._type, items) + continue + + # text element is primitive type + if prop_meta.get("text", False): + if args[0].text is not None: + dict_to_pass[rf._rest_name] = _deserialize(rf._type, args[0].text) + continue + + # wrapped element could be normal property or array, it should only have one element + item = args[0].find(xml_name) + if item is not None: + existed_attr_keys.append(xml_name) + dict_to_pass[rf._rest_name] = _deserialize(rf._type, item) + + # rest thing is additional properties + for e in args[0]: + if e.tag not in existed_attr_keys: + dict_to_pass[e.tag] = _convert_element(e) + else: + dict_to_pass.update( + {k: _create_value(_get_rest_field(self._attr_to_rest_field, k), v) for k, v in args[0].items()} + ) else: non_attr_kwargs = [k for k in kwargs if k not in self._attr_to_rest_field] if non_attr_kwargs: @@ -507,55 +574,70 @@ def __init__(self, *args: typing.Any, **kwargs: typing.Any) -> None: def copy(self) -> "Model": return Model(self.__dict__) - def __new__(cls, *args: typing.Any, **kwargs: typing.Any) -> Self: # pylint: disable=unused-argument - # we know the last three classes in mro are going to be 'Model', 'dict', and 'object' - mros = cls.__mro__[:-3][::-1] # ignore model, dict, and object parents, and reverse the mro order - attr_to_rest_field: typing.Dict[str, _RestField] = { # map attribute name to rest_field property - k: v for mro_class in mros for k, v in mro_class.__dict__.items() if k[0] != "_" and hasattr(v, "_type") - } - annotations = { - k: v - for mro_class in mros - if hasattr(mro_class, "__annotations__") # pylint: disable=no-member - for k, v in mro_class.__annotations__.items() # pylint: disable=no-member - } - for attr, rf in attr_to_rest_field.items(): - rf._module = cls.__module__ - if not rf._type: - rf._type = rf._get_deserialize_callable_from_annotation(annotations.get(attr, None)) - if not rf._rest_name_input: - rf._rest_name_input = attr - cls._attr_to_rest_field: typing.Dict[str, _RestField] = dict(attr_to_rest_field.items()) + def __new__(cls, *args: typing.Any, **kwargs: typing.Any) -> Self: + if f"{cls.__module__}.{cls.__qualname__}" not in cls._calculated: + # we know the last nine classes in mro are going to be 'Model', '_MyMutableMapping', 'MutableMapping', + # 'Mapping', 'Collection', 'Sized', 'Iterable', 'Container' and 'object' + mros = cls.__mro__[:-9][::-1] # ignore parents, and reverse the mro order + attr_to_rest_field: typing.Dict[str, _RestField] = { # map attribute name to rest_field property + k: v for mro_class in mros for k, v in mro_class.__dict__.items() if k[0] != "_" and hasattr(v, "_type") + } + annotations = { + k: v + for mro_class in mros + if hasattr(mro_class, "__annotations__") + for k, v in mro_class.__annotations__.items() + } + for attr, rf in attr_to_rest_field.items(): + rf._module = cls.__module__ + if not rf._type: + rf._type = rf._get_deserialize_callable_from_annotation(annotations.get(attr, None)) + if not rf._rest_name_input: + rf._rest_name_input = attr + cls._attr_to_rest_field: typing.Dict[str, _RestField] = dict(attr_to_rest_field.items()) + cls._calculated.add(f"{cls.__module__}.{cls.__qualname__}") return super().__new__(cls) # pylint: disable=no-value-for-parameter def __init_subclass__(cls, discriminator: typing.Optional[str] = None) -> None: for base in cls.__bases__: - if hasattr(base, "__mapping__"): # pylint: disable=no-member - base.__mapping__[discriminator or cls.__name__] = cls # type: ignore # pylint: disable=no-member + if hasattr(base, "__mapping__"): + base.__mapping__[discriminator or cls.__name__] = cls # type: ignore @classmethod - def _get_discriminator(cls, exist_discriminators) -> typing.Optional[str]: + def _get_discriminator(cls, exist_discriminators) -> typing.Optional["_RestField"]: for v in cls.__dict__.values(): - if ( - isinstance(v, _RestField) and v._is_discriminator and v._rest_name not in exist_discriminators - ): # pylint: disable=protected-access - return v._rest_name # pylint: disable=protected-access + if isinstance(v, _RestField) and v._is_discriminator and v._rest_name not in exist_discriminators: + return v return None @classmethod def _deserialize(cls, data, exist_discriminators): - if not hasattr(cls, "__mapping__"): # pylint: disable=no-member + if not hasattr(cls, "__mapping__"): return cls(data) discriminator = cls._get_discriminator(exist_discriminators) - exist_discriminators.append(discriminator) - mapped_cls = cls.__mapping__.get(data.get(discriminator), cls) # pyright: ignore # pylint: disable=no-member - if mapped_cls == cls: + if discriminator is None: return cls(data) - return mapped_cls._deserialize(data, exist_discriminators) # pylint: disable=protected-access + exist_discriminators.append(discriminator._rest_name) + if isinstance(data, ET.Element): + model_meta = getattr(cls, "_xml", {}) + prop_meta = getattr(discriminator, "_xml", {}) + xml_name = prop_meta.get("name", discriminator._rest_name) + xml_ns = prop_meta.get("ns", model_meta.get("ns", None)) + if xml_ns: + xml_name = "{" + xml_ns + "}" + xml_name + + if data.get(xml_name) is not None: + discriminator_value = data.get(xml_name) + else: + discriminator_value = data.find(xml_name).text # pyright: ignore + else: + discriminator_value = data.get(discriminator._rest_name) + mapped_cls = cls.__mapping__.get(discriminator_value, cls) # pyright: ignore + return mapped_cls._deserialize(data, exist_discriminators) def as_dict(self, *, exclude_readonly: bool = False) -> typing.Dict[str, typing.Any]: - """Return a dict that can be JSONify using json.dump. + """Return a dict that can be turned into json using json.dump. :keyword bool exclude_readonly: Whether to remove the readonly properties. :returns: A dict JSON compatible object @@ -563,6 +645,7 @@ def as_dict(self, *, exclude_readonly: bool = False) -> typing.Dict[str, typing. """ result = {} + readonly_props = [] if exclude_readonly: readonly_props = [p._rest_name for p in self._attr_to_rest_field.values() if _is_readonly(p)] for k, v in self.items(): @@ -617,6 +700,8 @@ def _deserialize_dict( ): if obj is None: return obj + if isinstance(obj, ET.Element): + obj = {child.tag: child for child in obj} return {k: _deserialize(value_deserializer, v, module) for k, v in obj.items()} @@ -637,6 +722,8 @@ def _deserialize_sequence( ): if obj is None: return obj + if isinstance(obj, ET.Element): + obj = list(obj) return type(obj)(_deserialize(deserializer, entry, module) for entry in obj) @@ -647,12 +734,12 @@ def _sorted_annotations(types: typing.List[typing.Any]) -> typing.List[typing.An ) -def _get_deserialize_callable_from_annotation( # pylint: disable=R0911, R0915, R0912 +def _get_deserialize_callable_from_annotation( # pylint: disable=too-many-return-statements, too-many-branches annotation: typing.Any, module: typing.Optional[str], rf: typing.Optional["_RestField"] = None, ) -> typing.Optional[typing.Callable[[typing.Any], typing.Any]]: - if not annotation or annotation in [int, float]: + if not annotation: return None # is it a type alias? @@ -727,7 +814,6 @@ def _get_deserialize_callable_from_annotation( # pylint: disable=R0911, R0915, try: if annotation._name in ["List", "Set", "Tuple", "Sequence"]: # pyright: ignore if len(annotation.__args__) > 1: # pyright: ignore - entry_deserializers = [ _get_deserialize_callable_from_annotation(dt, module, rf) for dt in annotation.__args__ # pyright: ignore @@ -762,12 +848,23 @@ def _deserialize_default( def _deserialize_with_callable( deserializer: typing.Optional[typing.Callable[[typing.Any], typing.Any]], value: typing.Any, -): +): # pylint: disable=too-many-return-statements try: if value is None or isinstance(value, _Null): return None + if isinstance(value, ET.Element): + if deserializer is str: + return value.text or "" + if deserializer is int: + return int(value.text) if value.text else None + if deserializer is float: + return float(value.text) if value.text else None + if deserializer is bool: + return value.text == "true" if value.text else None if deserializer is None: return value + if deserializer in [int, float, bool]: + return deserializer(value) if isinstance(deserializer, CaseInsensitiveEnumMeta): try: return deserializer(value) @@ -808,6 +905,7 @@ def __init__( default: typing.Any = _UNSET, format: typing.Optional[str] = None, is_multipart_file_input: bool = False, + xml: typing.Optional[typing.Dict[str, typing.Any]] = None, ): self._type = type self._rest_name_input = name @@ -818,6 +916,7 @@ def __init__( self._default = default self._format = format self._is_multipart_file_input = is_multipart_file_input + self._xml = xml if xml is not None else {} @property def _class_type(self) -> typing.Any: @@ -868,6 +967,7 @@ def rest_field( default: typing.Any = _UNSET, format: typing.Optional[str] = None, is_multipart_file_input: bool = False, + xml: typing.Optional[typing.Dict[str, typing.Any]] = None, ) -> typing.Any: return _RestField( name=name, @@ -876,6 +976,7 @@ def rest_field( default=default, format=format, is_multipart_file_input=is_multipart_file_input, + xml=xml, ) @@ -884,5 +985,175 @@ def rest_discriminator( name: typing.Optional[str] = None, type: typing.Optional[typing.Callable] = None, # pylint: disable=redefined-builtin visibility: typing.Optional[typing.List[str]] = None, + xml: typing.Optional[typing.Dict[str, typing.Any]] = None, ) -> typing.Any: - return _RestField(name=name, type=type, is_discriminator=True, visibility=visibility) + return _RestField(name=name, type=type, is_discriminator=True, visibility=visibility, xml=xml) + + +def serialize_xml(model: Model, exclude_readonly: bool = False) -> str: + """Serialize a model to XML. + + :param Model model: The model to serialize. + :param bool exclude_readonly: Whether to exclude readonly properties. + :returns: The XML representation of the model. + :rtype: str + """ + return ET.tostring(_get_element(model, exclude_readonly), encoding="unicode") # type: ignore + + +def _get_element( + o: typing.Any, + exclude_readonly: bool = False, + parent_meta: typing.Optional[typing.Dict[str, typing.Any]] = None, + wrapped_element: typing.Optional[ET.Element] = None, +) -> typing.Union[ET.Element, typing.List[ET.Element]]: + if _is_model(o): + model_meta = getattr(o, "_xml", {}) + + # if prop is a model, then use the prop element directly, else generate a wrapper of model + if wrapped_element is None: + wrapped_element = _create_xml_element( + model_meta.get("name", o.__class__.__name__), + model_meta.get("prefix"), + model_meta.get("ns"), + ) + + readonly_props = [] + if exclude_readonly: + readonly_props = [p._rest_name for p in o._attr_to_rest_field.values() if _is_readonly(p)] + + for k, v in o.items(): + # do not serialize readonly properties + if exclude_readonly and k in readonly_props: + continue + + prop_rest_field = _get_rest_field(o._attr_to_rest_field, k) + if prop_rest_field: + prop_meta = getattr(prop_rest_field, "_xml").copy() + # use the wire name as xml name if no specific name is set + if prop_meta.get("name") is None: + prop_meta["name"] = k + else: + # additional properties will not have rest field, use the wire name as xml name + prop_meta = {"name": k} + + # if no ns for prop, use model's + if prop_meta.get("ns") is None and model_meta.get("ns"): + prop_meta["ns"] = model_meta.get("ns") + prop_meta["prefix"] = model_meta.get("prefix") + + if prop_meta.get("unwrapped", False): + # unwrapped could only set on array + wrapped_element.extend(_get_element(v, exclude_readonly, prop_meta)) + elif prop_meta.get("text", False): + # text could only set on primitive type + wrapped_element.text = _get_primitive_type_value(v) + elif prop_meta.get("attribute", False): + xml_name = prop_meta.get("name", k) + if prop_meta.get("ns"): + ET.register_namespace(prop_meta.get("prefix"), prop_meta.get("ns")) # pyright: ignore + xml_name = "{" + prop_meta.get("ns") + "}" + xml_name # pyright: ignore + # attribute should be primitive type + wrapped_element.set(xml_name, _get_primitive_type_value(v)) + else: + # other wrapped prop element + wrapped_element.append(_get_wrapped_element(v, exclude_readonly, prop_meta)) + return wrapped_element + if isinstance(o, list): + return [_get_element(x, exclude_readonly, parent_meta) for x in o] # type: ignore + if isinstance(o, dict): + result = [] + for k, v in o.items(): + result.append( + _get_wrapped_element( + v, + exclude_readonly, + { + "name": k, + "ns": parent_meta.get("ns") if parent_meta else None, + "prefix": parent_meta.get("prefix") if parent_meta else None, + }, + ) + ) + return result + + # primitive case need to create element based on parent_meta + if parent_meta: + return _get_wrapped_element( + o, + exclude_readonly, + { + "name": parent_meta.get("itemsName", parent_meta.get("name")), + "prefix": parent_meta.get("itemsPrefix", parent_meta.get("prefix")), + "ns": parent_meta.get("itemsNs", parent_meta.get("ns")), + }, + ) + + raise ValueError("Could not serialize value into xml: " + o) + + +def _get_wrapped_element( + v: typing.Any, + exclude_readonly: bool, + meta: typing.Optional[typing.Dict[str, typing.Any]], +) -> ET.Element: + wrapped_element = _create_xml_element( + meta.get("name") if meta else None, meta.get("prefix") if meta else None, meta.get("ns") if meta else None + ) + if isinstance(v, (dict, list)): + wrapped_element.extend(_get_element(v, exclude_readonly, meta)) + elif _is_model(v): + _get_element(v, exclude_readonly, meta, wrapped_element) + else: + wrapped_element.text = _get_primitive_type_value(v) + return wrapped_element + + +def _get_primitive_type_value(v) -> str: + if v is True: + return "true" + if v is False: + return "false" + if isinstance(v, _Null): + return "" + return str(v) + + +def _create_xml_element(tag, prefix=None, ns=None): + if prefix and ns: + ET.register_namespace(prefix, ns) + if ns: + return ET.Element("{" + ns + "}" + tag) + return ET.Element(tag) + + +def _deserialize_xml( + deserializer: typing.Any, + value: str, +) -> typing.Any: + element = ET.fromstring(value) # nosec + return _deserialize(deserializer, element) + + +def _convert_element(e: ET.Element): + # dict case + if len(e.attrib) > 0 or len({child.tag for child in e}) > 1: + dict_result: typing.Dict[str, typing.Any] = {} + for child in e: + if dict_result.get(child.tag) is not None: + if isinstance(dict_result[child.tag], list): + dict_result[child.tag].append(_convert_element(child)) + else: + dict_result[child.tag] = [dict_result[child.tag], _convert_element(child)] + else: + dict_result[child.tag] = _convert_element(child) + dict_result.update(e.attrib) + return dict_result + # array case + if len(e) > 0: + array_result: typing.List[typing.Any] = [] + for child in e: + array_result.append(_convert_element(child)) + return array_result + # primitive case + return e.text diff --git a/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/_operations/__init__.py b/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/_operations/__init__.py index f30b11092e89..8a3952cdf768 100644 --- a/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/_operations/__init__.py +++ b/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/_operations/__init__.py @@ -5,15 +5,21 @@ # Code generated by Microsoft (R) Python Code Generator. # Changes may cause incorrect behavior and will be lost if the code is regenerated. # -------------------------------------------------------------------------- +# pylint: disable=wrong-import-position -from ._operations import DeidentificationClientOperationsMixin +from typing import TYPE_CHECKING + +if TYPE_CHECKING: + from ._patch import * # pylint: disable=unused-wildcard-import + +from ._operations import DeidentificationClientOperationsMixin # type: ignore from ._patch import __all__ as _patch_all -from ._patch import * # pylint: disable=unused-wildcard-import +from ._patch import * from ._patch import patch_sdk as _patch_sdk __all__ = [ "DeidentificationClientOperationsMixin", ] -__all__.extend([p for p in _patch_all if p not in __all__]) +__all__.extend([p for p in _patch_all if p not in __all__]) # pyright: ignore _patch_sdk() diff --git a/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/_operations/_operations.py b/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/_operations/_operations.py index f1aea8456604..06be6478e6db 100644 --- a/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/_operations/_operations.py +++ b/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/_operations/_operations.py @@ -1,4 +1,3 @@ -# pylint: disable=too-many-lines,too-many-statements # coding=utf-8 # -------------------------------------------------------------------------- # Copyright (c) Microsoft Corporation. All rights reserved. @@ -9,7 +8,7 @@ from io import IOBase import json import sys -from typing import Any, Callable, Dict, IO, Iterable, Iterator, List, Optional, Type, TypeVar, Union, cast, overload +from typing import Any, Callable, Dict, IO, Iterable, Iterator, List, Optional, TypeVar, Union, cast, overload import urllib.parse from azure.core.exceptions import ( @@ -18,6 +17,8 @@ ResourceExistsError, ResourceNotFoundError, ResourceNotModifiedError, + StreamClosedError, + StreamConsumedError, map_error, ) from azure.core.paging import ItemPaged @@ -36,7 +37,7 @@ if sys.version_info >= (3, 9): from collections.abc import MutableMapping else: - from typing import MutableMapping # type: ignore # pylint: disable=ungrouped-imports + from typing import MutableMapping # type: ignore JSON = MutableMapping[str, Any] # pylint: disable=unsubscriptable-object T = TypeVar("T") ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]] @@ -49,7 +50,7 @@ def build_deidentification_get_job_request(name: str, **kwargs: Any) -> HttpRequ _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) - api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2024-07-12-preview")) + api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2024-11-15")) accept = _headers.pop("Accept", "application/json") # Construct URL @@ -69,12 +70,14 @@ def build_deidentification_get_job_request(name: str, **kwargs: Any) -> HttpRequ return HttpRequest(method="GET", url=_url, params=_params, headers=_headers, **kwargs) -def build_deidentification_create_job_request(name: str, **kwargs: Any) -> HttpRequest: # pylint: disable=name-too-long +def build_deidentification_deidentify_documents_request( # pylint: disable=name-too-long + name: str, **kwargs: Any +) -> HttpRequest: _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) content_type: Optional[str] = kwargs.pop("content_type", _headers.pop("Content-Type", None)) - api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2024-07-12-preview")) + api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2024-11-15")) accept = _headers.pop("Accept", "application/json") # Construct URL @@ -96,13 +99,13 @@ def build_deidentification_create_job_request(name: str, **kwargs: Any) -> HttpR return HttpRequest(method="PUT", url=_url, params=_params, headers=_headers, **kwargs) -def build_deidentification_list_jobs_request( +def build_deidentification_list_jobs_internal_request( # pylint: disable=name-too-long *, maxpagesize: Optional[int] = None, continuation_token_parameter: Optional[str] = None, **kwargs: Any ) -> HttpRequest: _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) - api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2024-07-12-preview")) + api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2024-11-15")) accept = _headers.pop("Accept", "application/json") # Construct URL @@ -123,19 +126,23 @@ def build_deidentification_list_jobs_request( return HttpRequest(method="GET", url=_url, params=_params, headers=_headers, **kwargs) -def build_deidentification_list_job_documents_request( # pylint: disable=name-too-long - name: str, *, maxpagesize: Optional[int] = None, continuation_token_parameter: Optional[str] = None, **kwargs: Any +def build_deidentification_list_job_documents_internal_request( # pylint: disable=name-too-long + job_name: str, + *, + maxpagesize: Optional[int] = None, + continuation_token_parameter: Optional[str] = None, + **kwargs: Any ) -> HttpRequest: _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) - api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2024-07-12-preview")) + api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2024-11-15")) accept = _headers.pop("Accept", "application/json") # Construct URL - _url = "/jobs/{name}/documents" + _url = "/jobs/{jobName}/documents" path_format_arguments = { - "name": _SERIALIZER.url("name", name, "str"), + "jobName": _SERIALIZER.url("job_name", job_name, "str"), } _url: str = _url.format(**path_format_arguments) # type: ignore @@ -159,7 +166,7 @@ def build_deidentification_cancel_job_request(name: str, **kwargs: Any) -> HttpR _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) - api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2024-07-12-preview")) + api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2024-11-15")) accept = _headers.pop("Accept", "application/json") # Construct URL @@ -183,7 +190,7 @@ def build_deidentification_delete_job_request(name: str, **kwargs: Any) -> HttpR _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) - api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2024-07-12-preview")) + api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2024-11-15")) accept = _headers.pop("Accept", "application/json") # Construct URL @@ -203,12 +210,12 @@ def build_deidentification_delete_job_request(name: str, **kwargs: Any) -> HttpR return HttpRequest(method="DELETE", url=_url, params=_params, headers=_headers, **kwargs) -def build_deidentification_deidentify_request(**kwargs: Any) -> HttpRequest: # pylint: disable=name-too-long +def build_deidentification_deidentify_text_request(**kwargs: Any) -> HttpRequest: # pylint: disable=name-too-long _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) content_type: Optional[str] = kwargs.pop("content_type", _headers.pop("Content-Type", None)) - api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2024-07-12-preview")) + api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2024-11-15")) accept = _headers.pop("Accept", "application/json") # Construct URL @@ -238,53 +245,8 @@ def get_job(self, name: str, **kwargs: Any) -> _models.DeidentificationJob: :return: DeidentificationJob. The DeidentificationJob is compatible with MutableMapping :rtype: ~azure.health.deidentification.models.DeidentificationJob :raises ~azure.core.exceptions.HttpResponseError: - - Example: - .. code-block:: python - - # response body for status code(s): 200 - response == { - "createdAt": "2020-02-20 00:00:00", - "lastUpdatedAt": "2020-02-20 00:00:00", - "name": "str", - "sourceLocation": { - "location": "str", - "prefix": "str", - "extensions": [ - "str" - ] - }, - "status": "str", - "targetLocation": { - "location": "str", - "prefix": "str" - }, - "dataType": "str", - "error": { - "code": "str", - "message": "str", - "details": [ - ... - ], - "innererror": { - "code": "str", - "innererror": ... - }, - "target": "str" - }, - "operation": "str", - "redactionFormat": "str", - "startedAt": "2020-02-20 00:00:00", - "summary": { - "bytesProcessed": 0, - "canceled": 0, - "failed": 0, - "successful": 0, - "total": 0 - } - } """ - error_map: MutableMapping[int, Type[HttpResponseError]] = { + error_map: MutableMapping = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError, @@ -304,7 +266,7 @@ def get_job(self, name: str, **kwargs: Any) -> _models.DeidentificationJob: params=_params, ) path_format_arguments = { - "endpoint": self._serialize.url("self._config.endpoint", self._config.endpoint, "str"), + "endpoint": self._serialize.url("self._config.endpoint", self._config.endpoint, "str", skip_quote=True), } _request.url = self._client.format_url(_request.url, **path_format_arguments) @@ -317,7 +279,10 @@ def get_job(self, name: str, **kwargs: Any) -> _models.DeidentificationJob: if response.status_code not in [200]: if _stream: - response.read() # Load the body in memory and close the socket + try: + response.read() # Load the body in memory and close the socket + except (StreamConsumedError, StreamClosedError): + pass map_error(status_code=response.status_code, response=response, error_map=error_map) raise HttpResponseError(response=response) @@ -336,10 +301,10 @@ def get_job(self, name: str, **kwargs: Any) -> _models.DeidentificationJob: return deserialized # type: ignore - def _create_job_initial( + def _deidentify_documents_initial( self, name: str, resource: Union[_models.DeidentificationJob, JSON, IO[bytes]], **kwargs: Any ) -> Iterator[bytes]: - error_map: MutableMapping[int, Type[HttpResponseError]] = { + error_map: MutableMapping = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError, @@ -360,7 +325,7 @@ def _create_job_initial( else: _content = json.dumps(resource, cls=SdkJSONEncoder, exclude_readonly=True) # type: ignore - _request = build_deidentification_create_job_request( + _request = build_deidentification_deidentify_documents_request( name=name, content_type=content_type, api_version=self._config.api_version, @@ -369,7 +334,7 @@ def _create_job_initial( params=_params, ) path_format_arguments = { - "endpoint": self._serialize.url("self._config.endpoint", self._config.endpoint, "str"), + "endpoint": self._serialize.url("self._config.endpoint", self._config.endpoint, "str", skip_quote=True), } _request.url = self._client.format_url(_request.url, **path_format_arguments) @@ -381,30 +346,20 @@ def _create_job_initial( response = pipeline_response.http_response if response.status_code not in [200, 201]: - response.read() # Load the body in memory and close the socket + try: + response.read() # Load the body in memory and close the socket + except (StreamConsumedError, StreamClosedError): + pass map_error(status_code=response.status_code, response=response, error_map=error_map) raise HttpResponseError(response=response) response_headers = {} - if response.status_code == 200: - response_headers["x-ms-client-request-id"] = self._deserialize( - "str", response.headers.get("x-ms-client-request-id") - ) - response_headers["Operation-Location"] = self._deserialize( - "str", response.headers.get("Operation-Location") - ) - - deserialized = response.iter_bytes() - - if response.status_code == 201: - response_headers["x-ms-client-request-id"] = self._deserialize( - "str", response.headers.get("x-ms-client-request-id") - ) - response_headers["Operation-Location"] = self._deserialize( - "str", response.headers.get("Operation-Location") - ) + response_headers["x-ms-client-request-id"] = self._deserialize( + "str", response.headers.get("x-ms-client-request-id") + ) + response_headers["Operation-Location"] = self._deserialize("str", response.headers.get("Operation-Location")) - deserialized = response.iter_bytes() + deserialized = response.iter_bytes() if cls: return cls(pipeline_response, deserialized, response_headers) # type: ignore @@ -412,7 +367,7 @@ def _create_job_initial( return deserialized # type: ignore @overload - def begin_create_job( + def begin_deidentify_documents( self, name: str, resource: _models.DeidentificationJob, *, content_type: str = "application/json", **kwargs: Any ) -> LROPoller[_models.DeidentificationJob]: """Create a de-identification job. @@ -431,97 +386,10 @@ def begin_create_job( :rtype: ~azure.core.polling.LROPoller[~azure.health.deidentification.models.DeidentificationJob] :raises ~azure.core.exceptions.HttpResponseError: - - Example: - .. code-block:: python - - # JSON input template you can fill out and use as your body input. - resource = { - "createdAt": "2020-02-20 00:00:00", - "lastUpdatedAt": "2020-02-20 00:00:00", - "name": "str", - "sourceLocation": { - "location": "str", - "prefix": "str", - "extensions": [ - "str" - ] - }, - "status": "str", - "targetLocation": { - "location": "str", - "prefix": "str" - }, - "dataType": "str", - "error": { - "code": "str", - "message": "str", - "details": [ - ... - ], - "innererror": { - "code": "str", - "innererror": ... - }, - "target": "str" - }, - "operation": "str", - "redactionFormat": "str", - "startedAt": "2020-02-20 00:00:00", - "summary": { - "bytesProcessed": 0, - "canceled": 0, - "failed": 0, - "successful": 0, - "total": 0 - } - } - - # response body for status code(s): 201, 200 - response == { - "createdAt": "2020-02-20 00:00:00", - "lastUpdatedAt": "2020-02-20 00:00:00", - "name": "str", - "sourceLocation": { - "location": "str", - "prefix": "str", - "extensions": [ - "str" - ] - }, - "status": "str", - "targetLocation": { - "location": "str", - "prefix": "str" - }, - "dataType": "str", - "error": { - "code": "str", - "message": "str", - "details": [ - ... - ], - "innererror": { - "code": "str", - "innererror": ... - }, - "target": "str" - }, - "operation": "str", - "redactionFormat": "str", - "startedAt": "2020-02-20 00:00:00", - "summary": { - "bytesProcessed": 0, - "canceled": 0, - "failed": 0, - "successful": 0, - "total": 0 - } - } """ @overload - def begin_create_job( + def begin_deidentify_documents( self, name: str, resource: JSON, *, content_type: str = "application/json", **kwargs: Any ) -> LROPoller[_models.DeidentificationJob]: """Create a de-identification job. @@ -540,55 +408,10 @@ def begin_create_job( :rtype: ~azure.core.polling.LROPoller[~azure.health.deidentification.models.DeidentificationJob] :raises ~azure.core.exceptions.HttpResponseError: - - Example: - .. code-block:: python - - # response body for status code(s): 201, 200 - response == { - "createdAt": "2020-02-20 00:00:00", - "lastUpdatedAt": "2020-02-20 00:00:00", - "name": "str", - "sourceLocation": { - "location": "str", - "prefix": "str", - "extensions": [ - "str" - ] - }, - "status": "str", - "targetLocation": { - "location": "str", - "prefix": "str" - }, - "dataType": "str", - "error": { - "code": "str", - "message": "str", - "details": [ - ... - ], - "innererror": { - "code": "str", - "innererror": ... - }, - "target": "str" - }, - "operation": "str", - "redactionFormat": "str", - "startedAt": "2020-02-20 00:00:00", - "summary": { - "bytesProcessed": 0, - "canceled": 0, - "failed": 0, - "successful": 0, - "total": 0 - } - } """ @overload - def begin_create_job( + def begin_deidentify_documents( self, name: str, resource: IO[bytes], *, content_type: str = "application/json", **kwargs: Any ) -> LROPoller[_models.DeidentificationJob]: """Create a de-identification job. @@ -607,55 +430,10 @@ def begin_create_job( :rtype: ~azure.core.polling.LROPoller[~azure.health.deidentification.models.DeidentificationJob] :raises ~azure.core.exceptions.HttpResponseError: - - Example: - .. code-block:: python - - # response body for status code(s): 201, 200 - response == { - "createdAt": "2020-02-20 00:00:00", - "lastUpdatedAt": "2020-02-20 00:00:00", - "name": "str", - "sourceLocation": { - "location": "str", - "prefix": "str", - "extensions": [ - "str" - ] - }, - "status": "str", - "targetLocation": { - "location": "str", - "prefix": "str" - }, - "dataType": "str", - "error": { - "code": "str", - "message": "str", - "details": [ - ... - ], - "innererror": { - "code": "str", - "innererror": ... - }, - "target": "str" - }, - "operation": "str", - "redactionFormat": "str", - "startedAt": "2020-02-20 00:00:00", - "summary": { - "bytesProcessed": 0, - "canceled": 0, - "failed": 0, - "successful": 0, - "total": 0 - } - } """ @distributed_trace - def begin_create_job( + def begin_deidentify_documents( self, name: str, resource: Union[_models.DeidentificationJob, JSON, IO[bytes]], **kwargs: Any ) -> LROPoller[_models.DeidentificationJob]: """Create a de-identification job. @@ -672,93 +450,6 @@ def begin_create_job( :rtype: ~azure.core.polling.LROPoller[~azure.health.deidentification.models.DeidentificationJob] :raises ~azure.core.exceptions.HttpResponseError: - - Example: - .. code-block:: python - - # JSON input template you can fill out and use as your body input. - resource = { - "createdAt": "2020-02-20 00:00:00", - "lastUpdatedAt": "2020-02-20 00:00:00", - "name": "str", - "sourceLocation": { - "location": "str", - "prefix": "str", - "extensions": [ - "str" - ] - }, - "status": "str", - "targetLocation": { - "location": "str", - "prefix": "str" - }, - "dataType": "str", - "error": { - "code": "str", - "message": "str", - "details": [ - ... - ], - "innererror": { - "code": "str", - "innererror": ... - }, - "target": "str" - }, - "operation": "str", - "redactionFormat": "str", - "startedAt": "2020-02-20 00:00:00", - "summary": { - "bytesProcessed": 0, - "canceled": 0, - "failed": 0, - "successful": 0, - "total": 0 - } - } - - # response body for status code(s): 201, 200 - response == { - "createdAt": "2020-02-20 00:00:00", - "lastUpdatedAt": "2020-02-20 00:00:00", - "name": "str", - "sourceLocation": { - "location": "str", - "prefix": "str", - "extensions": [ - "str" - ] - }, - "status": "str", - "targetLocation": { - "location": "str", - "prefix": "str" - }, - "dataType": "str", - "error": { - "code": "str", - "message": "str", - "details": [ - ... - ], - "innererror": { - "code": "str", - "innererror": ... - }, - "target": "str" - }, - "operation": "str", - "redactionFormat": "str", - "startedAt": "2020-02-20 00:00:00", - "summary": { - "bytesProcessed": 0, - "canceled": 0, - "failed": 0, - "successful": 0, - "total": 0 - } - } """ _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) _params = kwargs.pop("params", {}) or {} @@ -769,7 +460,7 @@ def begin_create_job( lro_delay = kwargs.pop("polling_interval", self._config.polling_interval) cont_token: Optional[str] = kwargs.pop("continuation_token", None) if cont_token is None: - raw_result = self._create_job_initial( + raw_result = self._deidentify_documents_initial( name=name, resource=resource, content_type=content_type, @@ -797,7 +488,7 @@ def get_long_running_output(pipeline_response): return deserialized path_format_arguments = { - "endpoint": self._serialize.url("self._config.endpoint", self._config.endpoint, "str"), + "endpoint": self._serialize.url("self._config.endpoint", self._config.endpoint, "str", skip_quote=True), } if polling is True: @@ -820,7 +511,7 @@ def get_long_running_output(pipeline_response): ) @distributed_trace - def list_jobs( + def _list_jobs_internal( self, *, continuation_token_parameter: Optional[str] = None, **kwargs: Any ) -> Iterable["_models.DeidentificationJob"]: """List de-identification jobs. @@ -833,51 +524,6 @@ def list_jobs( :return: An iterator like instance of DeidentificationJob :rtype: ~azure.core.paging.ItemPaged[~azure.health.deidentification.models.DeidentificationJob] :raises ~azure.core.exceptions.HttpResponseError: - - Example: - .. code-block:: python - - # response body for status code(s): 200 - response == { - "createdAt": "2020-02-20 00:00:00", - "lastUpdatedAt": "2020-02-20 00:00:00", - "name": "str", - "sourceLocation": { - "location": "str", - "prefix": "str", - "extensions": [ - "str" - ] - }, - "status": "str", - "targetLocation": { - "location": "str", - "prefix": "str" - }, - "dataType": "str", - "error": { - "code": "str", - "message": "str", - "details": [ - ... - ], - "innererror": { - "code": "str", - "innererror": ... - }, - "target": "str" - }, - "operation": "str", - "redactionFormat": "str", - "startedAt": "2020-02-20 00:00:00", - "summary": { - "bytesProcessed": 0, - "canceled": 0, - "failed": 0, - "successful": 0, - "total": 0 - } - } """ _headers = kwargs.pop("headers", {}) or {} _params = kwargs.pop("params", {}) or {} @@ -885,7 +531,7 @@ def list_jobs( maxpagesize = kwargs.pop("maxpagesize", None) cls: ClsType[List[_models.DeidentificationJob]] = kwargs.pop("cls", None) - error_map: MutableMapping[int, Type[HttpResponseError]] = { + error_map: MutableMapping = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError, @@ -896,7 +542,7 @@ def list_jobs( def prepare_request(next_link=None): if not next_link: - _request = build_deidentification_list_jobs_request( + _request = build_deidentification_list_jobs_internal_request( maxpagesize=maxpagesize, continuation_token_parameter=continuation_token_parameter, api_version=self._config.api_version, @@ -904,7 +550,9 @@ def prepare_request(next_link=None): params=_params, ) path_format_arguments = { - "endpoint": self._serialize.url("self._config.endpoint", self._config.endpoint, "str"), + "endpoint": self._serialize.url( + "self._config.endpoint", self._config.endpoint, "str", skip_quote=True + ), } _request.url = self._client.format_url(_request.url, **path_format_arguments) @@ -922,7 +570,9 @@ def prepare_request(next_link=None): "GET", urllib.parse.urljoin(next_link, _parsed_next_link.path), params=_next_request_params ) path_format_arguments = { - "endpoint": self._serialize.url("self._config.endpoint", self._config.endpoint, "str"), + "endpoint": self._serialize.url( + "self._config.endpoint", self._config.endpoint, "str", skip_quote=True + ), } _request.url = self._client.format_url(_request.url, **path_format_arguments) @@ -953,58 +603,30 @@ def get_next(next_link=None): return ItemPaged(get_next, extract_data) @distributed_trace - def list_job_documents( - self, name: str, *, continuation_token_parameter: Optional[str] = None, **kwargs: Any - ) -> Iterable["_models.DocumentDetails"]: + def _list_job_documents_internal( + self, job_name: str, *, continuation_token_parameter: Optional[str] = None, **kwargs: Any + ) -> Iterable["_models.DeidentificationDocumentDetails"]: """List processed documents within a job. - Resource list operation template. + The most basic operation. - :param name: The name of a job. Required. - :type name: str + :param job_name: The name of a job. Required. + :type job_name: str :keyword continuation_token_parameter: Token to continue a previous query. Default value is None. :paramtype continuation_token_parameter: str - :return: An iterator like instance of DocumentDetails - :rtype: ~azure.core.paging.ItemPaged[~azure.health.deidentification.models.DocumentDetails] + :return: An iterator like instance of DeidentificationDocumentDetails + :rtype: + ~azure.core.paging.ItemPaged[~azure.health.deidentification.models.DeidentificationDocumentDetails] :raises ~azure.core.exceptions.HttpResponseError: - - Example: - .. code-block:: python - - # response body for status code(s): 200 - response == { - "id": "str", - "input": { - "etag": "str", - "path": "str" - }, - "status": "str", - "error": { - "code": "str", - "message": "str", - "details": [ - ... - ], - "innererror": { - "code": "str", - "innererror": ... - }, - "target": "str" - }, - "output": { - "etag": "str", - "path": "str" - } - } """ _headers = kwargs.pop("headers", {}) or {} _params = kwargs.pop("params", {}) or {} maxpagesize = kwargs.pop("maxpagesize", None) - cls: ClsType[List[_models.DocumentDetails]] = kwargs.pop("cls", None) + cls: ClsType[List[_models.DeidentificationDocumentDetails]] = kwargs.pop("cls", None) - error_map: MutableMapping[int, Type[HttpResponseError]] = { + error_map: MutableMapping = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError, @@ -1015,8 +637,8 @@ def list_job_documents( def prepare_request(next_link=None): if not next_link: - _request = build_deidentification_list_job_documents_request( - name=name, + _request = build_deidentification_list_job_documents_internal_request( + job_name=job_name, maxpagesize=maxpagesize, continuation_token_parameter=continuation_token_parameter, api_version=self._config.api_version, @@ -1024,7 +646,9 @@ def prepare_request(next_link=None): params=_params, ) path_format_arguments = { - "endpoint": self._serialize.url("self._config.endpoint", self._config.endpoint, "str"), + "endpoint": self._serialize.url( + "self._config.endpoint", self._config.endpoint, "str", skip_quote=True + ), } _request.url = self._client.format_url(_request.url, **path_format_arguments) @@ -1042,7 +666,9 @@ def prepare_request(next_link=None): "GET", urllib.parse.urljoin(next_link, _parsed_next_link.path), params=_next_request_params ) path_format_arguments = { - "endpoint": self._serialize.url("self._config.endpoint", self._config.endpoint, "str"), + "endpoint": self._serialize.url( + "self._config.endpoint", self._config.endpoint, "str", skip_quote=True + ), } _request.url = self._client.format_url(_request.url, **path_format_arguments) @@ -1050,7 +676,7 @@ def prepare_request(next_link=None): def extract_data(pipeline_response): deserialized = pipeline_response.http_response.json() - list_of_elem = _deserialize(List[_models.DocumentDetails], deserialized["value"]) + list_of_elem = _deserialize(List[_models.DeidentificationDocumentDetails], deserialized["value"]) if cls: list_of_elem = cls(list_of_elem) # type: ignore return deserialized.get("nextLink") or None, iter(list_of_elem) @@ -1088,53 +714,8 @@ def cancel_job(self, name: str, **kwargs: Any) -> _models.DeidentificationJob: :return: DeidentificationJob. The DeidentificationJob is compatible with MutableMapping :rtype: ~azure.health.deidentification.models.DeidentificationJob :raises ~azure.core.exceptions.HttpResponseError: - - Example: - .. code-block:: python - - # response body for status code(s): 200 - response == { - "createdAt": "2020-02-20 00:00:00", - "lastUpdatedAt": "2020-02-20 00:00:00", - "name": "str", - "sourceLocation": { - "location": "str", - "prefix": "str", - "extensions": [ - "str" - ] - }, - "status": "str", - "targetLocation": { - "location": "str", - "prefix": "str" - }, - "dataType": "str", - "error": { - "code": "str", - "message": "str", - "details": [ - ... - ], - "innererror": { - "code": "str", - "innererror": ... - }, - "target": "str" - }, - "operation": "str", - "redactionFormat": "str", - "startedAt": "2020-02-20 00:00:00", - "summary": { - "bytesProcessed": 0, - "canceled": 0, - "failed": 0, - "successful": 0, - "total": 0 - } - } """ - error_map: MutableMapping[int, Type[HttpResponseError]] = { + error_map: MutableMapping = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError, @@ -1154,7 +735,7 @@ def cancel_job(self, name: str, **kwargs: Any) -> _models.DeidentificationJob: params=_params, ) path_format_arguments = { - "endpoint": self._serialize.url("self._config.endpoint", self._config.endpoint, "str"), + "endpoint": self._serialize.url("self._config.endpoint", self._config.endpoint, "str", skip_quote=True), } _request.url = self._client.format_url(_request.url, **path_format_arguments) @@ -1167,7 +748,10 @@ def cancel_job(self, name: str, **kwargs: Any) -> _models.DeidentificationJob: if response.status_code not in [200]: if _stream: - response.read() # Load the body in memory and close the socket + try: + response.read() # Load the body in memory and close the socket + except (StreamConsumedError, StreamClosedError): + pass map_error(status_code=response.status_code, response=response, error_map=error_map) raise HttpResponseError(response=response) @@ -1198,7 +782,7 @@ def delete_job(self, name: str, **kwargs: Any) -> None: # pylint: disable=incon :rtype: None :raises ~azure.core.exceptions.HttpResponseError: """ - error_map: MutableMapping[int, Type[HttpResponseError]] = { + error_map: MutableMapping = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError, @@ -1218,7 +802,7 @@ def delete_job(self, name: str, **kwargs: Any) -> None: # pylint: disable=incon params=_params, ) path_format_arguments = { - "endpoint": self._serialize.url("self._config.endpoint", self._config.endpoint, "str"), + "endpoint": self._serialize.url("self._config.endpoint", self._config.endpoint, "str", skip_quote=True), } _request.url = self._client.format_url(_request.url, **path_format_arguments) @@ -1242,7 +826,7 @@ def delete_job(self, name: str, **kwargs: Any) -> None: # pylint: disable=incon return cls(pipeline_response, None, response_headers) # type: ignore @overload - def deidentify( + def deidentify_text( self, body: _models.DeidentificationContent, *, content_type: str = "application/json", **kwargs: Any ) -> _models.DeidentificationResult: """De-identify text. @@ -1257,47 +841,10 @@ def deidentify( :return: DeidentificationResult. The DeidentificationResult is compatible with MutableMapping :rtype: ~azure.health.deidentification.models.DeidentificationResult :raises ~azure.core.exceptions.HttpResponseError: - - Example: - .. code-block:: python - - # JSON input template you can fill out and use as your body input. - body = { - "inputText": "str", - "dataType": "str", - "operation": "str", - "redactionFormat": "str" - } - - # response body for status code(s): 200 - response == { - "outputText": "str", - "taggerResult": { - "entities": [ - { - "category": "str", - "length": { - "codePoint": 0, - "utf16": 0, - "utf8": 0 - }, - "offset": { - "codePoint": 0, - "utf16": 0, - "utf8": 0 - }, - "confidenceScore": 0.0, - "text": "str" - } - ], - "etag": "str", - "path": "str" - } - } """ @overload - def deidentify( + def deidentify_text( self, body: JSON, *, content_type: str = "application/json", **kwargs: Any ) -> _models.DeidentificationResult: """De-identify text. @@ -1312,39 +859,10 @@ def deidentify( :return: DeidentificationResult. The DeidentificationResult is compatible with MutableMapping :rtype: ~azure.health.deidentification.models.DeidentificationResult :raises ~azure.core.exceptions.HttpResponseError: - - Example: - .. code-block:: python - - # response body for status code(s): 200 - response == { - "outputText": "str", - "taggerResult": { - "entities": [ - { - "category": "str", - "length": { - "codePoint": 0, - "utf16": 0, - "utf8": 0 - }, - "offset": { - "codePoint": 0, - "utf16": 0, - "utf8": 0 - }, - "confidenceScore": 0.0, - "text": "str" - } - ], - "etag": "str", - "path": "str" - } - } """ @overload - def deidentify( + def deidentify_text( self, body: IO[bytes], *, content_type: str = "application/json", **kwargs: Any ) -> _models.DeidentificationResult: """De-identify text. @@ -1359,39 +877,10 @@ def deidentify( :return: DeidentificationResult. The DeidentificationResult is compatible with MutableMapping :rtype: ~azure.health.deidentification.models.DeidentificationResult :raises ~azure.core.exceptions.HttpResponseError: - - Example: - .. code-block:: python - - # response body for status code(s): 200 - response == { - "outputText": "str", - "taggerResult": { - "entities": [ - { - "category": "str", - "length": { - "codePoint": 0, - "utf16": 0, - "utf8": 0 - }, - "offset": { - "codePoint": 0, - "utf16": 0, - "utf8": 0 - }, - "confidenceScore": 0.0, - "text": "str" - } - ], - "etag": "str", - "path": "str" - } - } """ @distributed_trace - def deidentify( + def deidentify_text( self, body: Union[_models.DeidentificationContent, JSON, IO[bytes]], **kwargs: Any ) -> _models.DeidentificationResult: """De-identify text. @@ -1404,45 +893,8 @@ def deidentify( :return: DeidentificationResult. The DeidentificationResult is compatible with MutableMapping :rtype: ~azure.health.deidentification.models.DeidentificationResult :raises ~azure.core.exceptions.HttpResponseError: - - Example: - .. code-block:: python - - # JSON input template you can fill out and use as your body input. - body = { - "inputText": "str", - "dataType": "str", - "operation": "str", - "redactionFormat": "str" - } - - # response body for status code(s): 200 - response == { - "outputText": "str", - "taggerResult": { - "entities": [ - { - "category": "str", - "length": { - "codePoint": 0, - "utf16": 0, - "utf8": 0 - }, - "offset": { - "codePoint": 0, - "utf16": 0, - "utf8": 0 - }, - "confidenceScore": 0.0, - "text": "str" - } - ], - "etag": "str", - "path": "str" - } - } """ - error_map: MutableMapping[int, Type[HttpResponseError]] = { + error_map: MutableMapping = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError, @@ -1463,7 +915,7 @@ def deidentify( else: _content = json.dumps(body, cls=SdkJSONEncoder, exclude_readonly=True) # type: ignore - _request = build_deidentification_deidentify_request( + _request = build_deidentification_deidentify_text_request( content_type=content_type, api_version=self._config.api_version, content=_content, @@ -1471,7 +923,7 @@ def deidentify( params=_params, ) path_format_arguments = { - "endpoint": self._serialize.url("self._config.endpoint", self._config.endpoint, "str"), + "endpoint": self._serialize.url("self._config.endpoint", self._config.endpoint, "str", skip_quote=True), } _request.url = self._client.format_url(_request.url, **path_format_arguments) @@ -1484,16 +936,24 @@ def deidentify( if response.status_code not in [200]: if _stream: - response.read() # Load the body in memory and close the socket + try: + response.read() # Load the body in memory and close the socket + except (StreamConsumedError, StreamClosedError): + pass map_error(status_code=response.status_code, response=response, error_map=error_map) raise HttpResponseError(response=response) + response_headers = {} + response_headers["x-ms-client-request-id"] = self._deserialize( + "str", response.headers.get("x-ms-client-request-id") + ) + if _stream: deserialized = response.iter_bytes() else: deserialized = _deserialize(_models.DeidentificationResult, response.json()) if cls: - return cls(pipeline_response, deserialized, {}) # type: ignore + return cls(pipeline_response, deserialized, response_headers) # type: ignore return deserialized # type: ignore diff --git a/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/_patch.py b/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/_patch.py index f7dd32510333..bb09ee2cbf77 100644 --- a/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/_patch.py +++ b/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/_patch.py @@ -6,9 +6,41 @@ Follow our quickstart for examples: https://aka.ms/azsdk/python/dpcodegen/python/customize """ -from typing import List +from typing import Any, Iterable, List +from azure.core.tracing.decorator import distributed_trace +from azure.health.deidentification.models import DeidentificationDocumentDetails, DeidentificationJob +from ._client import DeidentificationClient as DeidentificationClientGenerated -__all__: List[str] = [] # Add all objects you want publicly available to users at this package level +__all__: List[str] = [ + "DeidentificationClient", +] # Add all objects you want publicly available to users at this package level + + +class DeidentificationClient(DeidentificationClientGenerated): + + @distributed_trace + def list_jobs(self, **kwargs: Any) -> Iterable[DeidentificationJob]: + """ + List de-identification jobs. + + :return: An iterator like instance of DeidentificationJob + :rtype: ~azure.core.paging.ItemPaged[~azure.health.deidentification.models.DeidentificationJob] + :raises ~azure.core.exceptions.HttpResponseError: + """ + return super()._list_jobs_internal(continuation_token_parameter=None, **kwargs) + + @distributed_trace + def list_job_documents(self, job_name: str, **kwargs: Any) -> Iterable[DeidentificationDocumentDetails]: + """ + List processed documents within a job. + + :param job_name: The name of a job. Required. + :type job_name: str + :return: An iterator like instance of DocumentDetails + :rtype: ~azure.core.paging.ItemPaged[~azure.health.deidentification.models.DocumentDetails] + :raises ~azure.core.exceptions.HttpResponseError: + """ + return super()._list_job_documents_internal(job_name=job_name, continuation_token_parameter=None, **kwargs) def patch_sdk(): diff --git a/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/_serialization.py b/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/_serialization.py index 8139854b97bb..ce17d1798ce7 100644 --- a/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/_serialization.py +++ b/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/_serialization.py @@ -1,3 +1,4 @@ +# pylint: disable=too-many-lines # -------------------------------------------------------------------------- # # Copyright (c) Microsoft Corporation. All rights reserved. @@ -24,7 +25,6 @@ # # -------------------------------------------------------------------------- -# pylint: skip-file # pyright: reportUnnecessaryTypeIgnoreComment=false from base64 import b64decode, b64encode @@ -52,7 +52,6 @@ MutableMapping, Type, List, - Mapping, ) try: @@ -91,6 +90,8 @@ def deserialize_from_text(cls, data: Optional[Union[AnyStr, IO]], content_type: :param data: Input, could be bytes or stream (will be decoded with UTF8) or text :type data: str or bytes or IO :param str content_type: The content type. + :return: The deserialized data. + :rtype: object """ if hasattr(data, "read"): # Assume a stream @@ -112,7 +113,7 @@ def deserialize_from_text(cls, data: Optional[Union[AnyStr, IO]], content_type: try: return json.loads(data_as_str) except ValueError as err: - raise DeserializationError("JSON is invalid: {}".format(err), err) + raise DeserializationError("JSON is invalid: {}".format(err), err) from err elif "xml" in (content_type or []): try: @@ -155,6 +156,11 @@ def deserialize_from_http_generics(cls, body_bytes: Optional[Union[AnyStr, IO]], Use bytes and headers to NOT use any requests/aiohttp or whatever specific implementation. Headers will tested for "content-type" + + :param bytes body_bytes: The body of the response. + :param dict headers: The headers of the response. + :returns: The deserialized data. + :rtype: object """ # Try to use content-type from headers if available content_type = None @@ -184,15 +190,30 @@ class UTC(datetime.tzinfo): """Time Zone info for handling UTC""" def utcoffset(self, dt): - """UTF offset for UTC is 0.""" + """UTF offset for UTC is 0. + + :param datetime.datetime dt: The datetime + :returns: The offset + :rtype: datetime.timedelta + """ return datetime.timedelta(0) def tzname(self, dt): - """Timestamp representation.""" + """Timestamp representation. + + :param datetime.datetime dt: The datetime + :returns: The timestamp representation + :rtype: str + """ return "Z" def dst(self, dt): - """No daylight saving for UTC.""" + """No daylight saving for UTC. + + :param datetime.datetime dt: The datetime + :returns: The daylight saving time + :rtype: datetime.timedelta + """ return datetime.timedelta(hours=1) @@ -206,7 +227,7 @@ class _FixedOffset(datetime.tzinfo): # type: ignore :param datetime.timedelta offset: offset in timedelta format """ - def __init__(self, offset): + def __init__(self, offset) -> None: self.__offset = offset def utcoffset(self, dt): @@ -235,24 +256,26 @@ def __getinitargs__(self): _FLATTEN = re.compile(r"(? None: self.additional_properties: Optional[Dict[str, Any]] = {} - for k in kwargs: + for k in kwargs: # pylint: disable=consider-using-dict-items if k not in self._attribute_map: _LOGGER.warning("%s is not a known attribute of class %s and will be ignored", k, self.__class__) elif k in self._validation and self._validation[k].get("readonly", False): @@ -300,13 +330,23 @@ def __init__(self, **kwargs: Any) -> None: setattr(self, k, kwargs[k]) def __eq__(self, other: Any) -> bool: - """Compare objects by comparing all attributes.""" + """Compare objects by comparing all attributes. + + :param object other: The object to compare + :returns: True if objects are equal + :rtype: bool + """ if isinstance(other, self.__class__): return self.__dict__ == other.__dict__ return False def __ne__(self, other: Any) -> bool: - """Compare objects by comparing all attributes.""" + """Compare objects by comparing all attributes. + + :param object other: The object to compare + :returns: True if objects are not equal + :rtype: bool + """ return not self.__eq__(other) def __str__(self) -> str: @@ -326,7 +366,11 @@ def is_xml_model(cls) -> bool: @classmethod def _create_xml_node(cls): - """Create XML node.""" + """Create XML node. + + :returns: The XML node + :rtype: xml.etree.ElementTree.Element + """ try: xml_map = cls._xml_map # type: ignore except AttributeError: @@ -346,7 +390,9 @@ def serialize(self, keep_readonly: bool = False, **kwargs: Any) -> JSON: :rtype: dict """ serializer = Serializer(self._infer_class_models()) - return serializer._serialize(self, keep_readonly=keep_readonly, **kwargs) # type: ignore + return serializer._serialize( # type: ignore # pylint: disable=protected-access + self, keep_readonly=keep_readonly, **kwargs + ) def as_dict( self, @@ -380,12 +426,15 @@ def my_key_transformer(key, attr_desc, value): If you want XML serialization, you can pass the kwargs is_xml=True. + :param bool keep_readonly: If you want to serialize the readonly attributes :param function key_transformer: A key transformer function. :returns: A dict JSON compatible object :rtype: dict """ serializer = Serializer(self._infer_class_models()) - return serializer._serialize(self, key_transformer=key_transformer, keep_readonly=keep_readonly, **kwargs) # type: ignore + return serializer._serialize( # type: ignore # pylint: disable=protected-access + self, key_transformer=key_transformer, keep_readonly=keep_readonly, **kwargs + ) @classmethod def _infer_class_models(cls): @@ -395,7 +444,7 @@ def _infer_class_models(cls): client_models = {k: v for k, v in models.__dict__.items() if isinstance(v, type)} if cls.__name__ not in client_models: raise ValueError("Not Autorest generated code") - except Exception: + except Exception: # pylint: disable=broad-exception-caught # Assume it's not Autorest generated (tests?). Add ourselves as dependencies. client_models = {cls.__name__: cls} return client_models @@ -408,6 +457,7 @@ def deserialize(cls: Type[ModelType], data: Any, content_type: Optional[str] = N :param str content_type: JSON by default, set application/xml if XML. :returns: An instance of this model :raises: DeserializationError if something went wrong + :rtype: ModelType """ deserializer = Deserializer(cls._infer_class_models()) return deserializer(cls.__name__, data, content_type=content_type) # type: ignore @@ -426,9 +476,11 @@ def from_dict( and last_rest_key_case_insensitive_extractor) :param dict data: A dict using RestAPI structure + :param function key_extractors: A key extractor function. :param str content_type: JSON by default, set application/xml if XML. :returns: An instance of this model :raises: DeserializationError if something went wrong + :rtype: ModelType """ deserializer = Deserializer(cls._infer_class_models()) deserializer.key_extractors = ( # type: ignore @@ -448,21 +500,25 @@ def _flatten_subtype(cls, key, objects): return {} result = dict(cls._subtype_map[key]) for valuetype in cls._subtype_map[key].values(): - result.update(objects[valuetype]._flatten_subtype(key, objects)) + result.update(objects[valuetype]._flatten_subtype(key, objects)) # pylint: disable=protected-access return result @classmethod def _classify(cls, response, objects): """Check the class _subtype_map for any child classes. We want to ignore any inherited _subtype_maps. - Remove the polymorphic key from the initial data. + + :param dict response: The initial data + :param dict objects: The class objects + :returns: The class to be used + :rtype: class """ for subtype_key in cls.__dict__.get("_subtype_map", {}).keys(): subtype_value = None if not isinstance(response, ET.Element): rest_api_response_key = cls._get_rest_key_parts(subtype_key)[-1] - subtype_value = response.pop(rest_api_response_key, None) or response.pop(subtype_key, None) + subtype_value = response.get(rest_api_response_key, None) or response.get(subtype_key, None) else: subtype_value = xml_key_extractor(subtype_key, cls._attribute_map[subtype_key], response) if subtype_value: @@ -501,11 +557,13 @@ def _decode_attribute_map_key(key): inside the received data. :param str key: A key string from the generated code + :returns: The decoded key + :rtype: str """ return key.replace("\\.", ".") -class Serializer(object): +class Serializer(object): # pylint: disable=too-many-public-methods """Request object model serializer.""" basic_types = {str: "str", int: "int", bool: "bool", float: "float"} @@ -540,7 +598,7 @@ class Serializer(object): "multiple": lambda x, y: x % y != 0, } - def __init__(self, classes: Optional[Mapping[str, type]] = None): + def __init__(self, classes: Optional[Mapping[str, type]] = None) -> None: self.serialize_type = { "iso-8601": Serializer.serialize_iso, "rfc-1123": Serializer.serialize_rfc, @@ -560,13 +618,16 @@ def __init__(self, classes: Optional[Mapping[str, type]] = None): self.key_transformer = full_restapi_key_transformer self.client_side_validation = True - def _serialize(self, target_obj, data_type=None, **kwargs): + def _serialize( # pylint: disable=too-many-nested-blocks, too-many-branches, too-many-statements, too-many-locals + self, target_obj, data_type=None, **kwargs + ): """Serialize data into a string according to type. - :param target_obj: The data to be serialized. + :param object target_obj: The data to be serialized. :param str data_type: The type to be serialized from. :rtype: str, dict :raises: SerializationError if serialization fails. + :returns: The serialized data. """ key_transformer = kwargs.get("key_transformer", self.key_transformer) keep_readonly = kwargs.get("keep_readonly", False) @@ -592,12 +653,14 @@ def _serialize(self, target_obj, data_type=None, **kwargs): serialized = {} if is_xml_model_serialization: - serialized = target_obj._create_xml_node() + serialized = target_obj._create_xml_node() # pylint: disable=protected-access try: - attributes = target_obj._attribute_map + attributes = target_obj._attribute_map # pylint: disable=protected-access for attr, attr_desc in attributes.items(): attr_name = attr - if not keep_readonly and target_obj._validation.get(attr_name, {}).get("readonly", False): + if not keep_readonly and target_obj._validation.get( # pylint: disable=protected-access + attr_name, {} + ).get("readonly", False): continue if attr_name == "additional_properties" and attr_desc["key"] == "": @@ -633,7 +696,8 @@ def _serialize(self, target_obj, data_type=None, **kwargs): if isinstance(new_attr, list): serialized.extend(new_attr) # type: ignore elif isinstance(new_attr, ET.Element): - # If the down XML has no XML/Name, we MUST replace the tag with the local tag. But keeping the namespaces. + # If the down XML has no XML/Name, + # we MUST replace the tag with the local tag. But keeping the namespaces. if "name" not in getattr(orig_attr, "_xml_map", {}): splitted_tag = new_attr.tag.split("}") if len(splitted_tag) == 2: # Namespace @@ -664,17 +728,17 @@ def _serialize(self, target_obj, data_type=None, **kwargs): except (AttributeError, KeyError, TypeError) as err: msg = "Attribute {} in object {} cannot be serialized.\n{}".format(attr_name, class_name, str(target_obj)) raise SerializationError(msg) from err - else: - return serialized + return serialized def body(self, data, data_type, **kwargs): """Serialize data intended for a request body. - :param data: The data to be serialized. + :param object data: The data to be serialized. :param str data_type: The type to be serialized from. :rtype: dict :raises: SerializationError if serialization fails. :raises: ValueError if data is None + :returns: The serialized request body """ # Just in case this is a dict @@ -703,7 +767,7 @@ def body(self, data, data_type, **kwargs): attribute_key_case_insensitive_extractor, last_rest_key_case_insensitive_extractor, ] - data = deserializer._deserialize(data_type, data) + data = deserializer._deserialize(data_type, data) # pylint: disable=protected-access except DeserializationError as err: raise SerializationError("Unable to build a model: " + str(err)) from err @@ -712,9 +776,11 @@ def body(self, data, data_type, **kwargs): def url(self, name, data, data_type, **kwargs): """Serialize data intended for a URL path. - :param data: The data to be serialized. + :param str name: The name of the URL path parameter. + :param object data: The data to be serialized. :param str data_type: The type to be serialized from. :rtype: str + :returns: The serialized URL path :raises: TypeError if serialization fails. :raises: ValueError if data is None """ @@ -728,21 +794,20 @@ def url(self, name, data, data_type, **kwargs): output = output.replace("{", quote("{")).replace("}", quote("}")) else: output = quote(str(output), safe="") - except SerializationError: - raise TypeError("{} must be type {}.".format(name, data_type)) - else: - return output + except SerializationError as exc: + raise TypeError("{} must be type {}.".format(name, data_type)) from exc + return output def query(self, name, data, data_type, **kwargs): """Serialize data intended for a URL query. - :param data: The data to be serialized. + :param str name: The name of the query parameter. + :param object data: The data to be serialized. :param str data_type: The type to be serialized from. - :keyword bool skip_quote: Whether to skip quote the serialized result. - Defaults to False. :rtype: str, list :raises: TypeError if serialization fails. :raises: ValueError if data is None + :returns: The serialized query parameter """ try: # Treat the list aside, since we don't want to encode the div separator @@ -759,19 +824,20 @@ def query(self, name, data, data_type, **kwargs): output = str(output) else: output = quote(str(output), safe="") - except SerializationError: - raise TypeError("{} must be type {}.".format(name, data_type)) - else: - return str(output) + except SerializationError as exc: + raise TypeError("{} must be type {}.".format(name, data_type)) from exc + return str(output) def header(self, name, data, data_type, **kwargs): """Serialize data intended for a request header. - :param data: The data to be serialized. + :param str name: The name of the header. + :param object data: The data to be serialized. :param str data_type: The type to be serialized from. :rtype: str :raises: TypeError if serialization fails. :raises: ValueError if data is None + :returns: The serialized header """ try: if data_type in ["[str]"]: @@ -780,21 +846,20 @@ def header(self, name, data, data_type, **kwargs): output = self.serialize_data(data, data_type, **kwargs) if data_type == "bool": output = json.dumps(output) - except SerializationError: - raise TypeError("{} must be type {}.".format(name, data_type)) - else: - return str(output) + except SerializationError as exc: + raise TypeError("{} must be type {}.".format(name, data_type)) from exc + return str(output) def serialize_data(self, data, data_type, **kwargs): """Serialize generic data according to supplied data type. - :param data: The data to be serialized. + :param object data: The data to be serialized. :param str data_type: The type to be serialized from. - :param bool required: Whether it's essential that the data not be - empty or None :raises: AttributeError if required data is None. :raises: ValueError if data is None :raises: SerializationError if serialization fails. + :returns: The serialized data. + :rtype: str, int, float, bool, dict, list """ if data is None: raise ValueError("No value for given attribute") @@ -805,7 +870,7 @@ def serialize_data(self, data, data_type, **kwargs): if data_type in self.basic_types.values(): return self.serialize_basic(data, data_type, **kwargs) - elif data_type in self.serialize_type: + if data_type in self.serialize_type: return self.serialize_type[data_type](data, **kwargs) # If dependencies is empty, try with current data class @@ -821,11 +886,10 @@ def serialize_data(self, data, data_type, **kwargs): except (ValueError, TypeError) as err: msg = "Unable to serialize value: {!r} as type: {!r}." raise SerializationError(msg.format(data, data_type)) from err - else: - return self._serialize(data, **kwargs) + return self._serialize(data, **kwargs) @classmethod - def _get_custom_serializers(cls, data_type, **kwargs): + def _get_custom_serializers(cls, data_type, **kwargs): # pylint: disable=inconsistent-return-statements custom_serializer = kwargs.get("basic_types_serializers", {}).get(data_type) if custom_serializer: return custom_serializer @@ -841,23 +905,26 @@ def serialize_basic(cls, data, data_type, **kwargs): - basic_types_serializers dict[str, callable] : If set, use the callable as serializer - is_xml bool : If set, use xml_basic_types_serializers - :param data: Object to be serialized. + :param obj data: Object to be serialized. :param str data_type: Type of object in the iterable. + :rtype: str, int, float, bool + :return: serialized object """ custom_serializer = cls._get_custom_serializers(data_type, **kwargs) if custom_serializer: return custom_serializer(data) if data_type == "str": return cls.serialize_unicode(data) - return eval(data_type)(data) # nosec + return eval(data_type)(data) # nosec # pylint: disable=eval-used @classmethod def serialize_unicode(cls, data): """Special handling for serializing unicode strings in Py2. Encode to UTF-8 if unicode, otherwise handle as a str. - :param data: Object to be serialized. + :param str data: Object to be serialized. :rtype: str + :return: serialized object """ try: # If I received an enum, return its value return data.value @@ -871,8 +938,7 @@ def serialize_unicode(cls, data): return data except NameError: return str(data) - else: - return str(data) + return str(data) def serialize_iter(self, data, iter_type, div=None, **kwargs): """Serialize iterable. @@ -882,15 +948,13 @@ def serialize_iter(self, data, iter_type, div=None, **kwargs): serialization_ctxt['type'] should be same as data_type. - is_xml bool : If set, serialize as XML - :param list attr: Object to be serialized. + :param list data: Object to be serialized. :param str iter_type: Type of object in the iterable. - :param bool required: Whether the objects in the iterable must - not be None or empty. :param str div: If set, this str will be used to combine the elements in the iterable into a combined string. Default is 'None'. - :keyword bool do_quote: Whether to quote the serialized result of each iterable element. Defaults to False. :rtype: list, str + :return: serialized iterable """ if isinstance(data, str): raise SerializationError("Refuse str type as a valid iter type.") @@ -945,9 +1009,8 @@ def serialize_dict(self, attr, dict_type, **kwargs): :param dict attr: Object to be serialized. :param str dict_type: Type of object in the dictionary. - :param bool required: Whether the objects in the dictionary must - not be None or empty. :rtype: dict + :return: serialized dictionary """ serialization_ctxt = kwargs.get("serialization_ctxt", {}) serialized = {} @@ -971,7 +1034,7 @@ def serialize_dict(self, attr, dict_type, **kwargs): return serialized - def serialize_object(self, attr, **kwargs): + def serialize_object(self, attr, **kwargs): # pylint: disable=too-many-return-statements """Serialize a generic object. This will be handled as a dictionary. If object passed in is not a basic type (str, int, float, dict, list) it will simply be @@ -979,6 +1042,7 @@ def serialize_object(self, attr, **kwargs): :param dict attr: Object to be serialized. :rtype: dict or str + :return: serialized object """ if attr is None: return None @@ -1003,7 +1067,7 @@ def serialize_object(self, attr, **kwargs): return self.serialize_decimal(attr) # If it's a model or I know this dependency, serialize as a Model - elif obj_type in self.dependencies.values() or isinstance(attr, Model): + if obj_type in self.dependencies.values() or isinstance(attr, Model): return self._serialize(attr) if obj_type == dict: @@ -1034,56 +1098,61 @@ def serialize_enum(attr, enum_obj=None): try: enum_obj(result) # type: ignore return result - except ValueError: + except ValueError as exc: for enum_value in enum_obj: # type: ignore if enum_value.value.lower() == str(attr).lower(): return enum_value.value error = "{!r} is not valid value for enum {!r}" - raise SerializationError(error.format(attr, enum_obj)) + raise SerializationError(error.format(attr, enum_obj)) from exc @staticmethod - def serialize_bytearray(attr, **kwargs): + def serialize_bytearray(attr, **kwargs): # pylint: disable=unused-argument """Serialize bytearray into base-64 string. - :param attr: Object to be serialized. + :param str attr: Object to be serialized. :rtype: str + :return: serialized base64 """ return b64encode(attr).decode() @staticmethod - def serialize_base64(attr, **kwargs): + def serialize_base64(attr, **kwargs): # pylint: disable=unused-argument """Serialize str into base-64 string. - :param attr: Object to be serialized. + :param str attr: Object to be serialized. :rtype: str + :return: serialized base64 """ encoded = b64encode(attr).decode("ascii") return encoded.strip("=").replace("+", "-").replace("/", "_") @staticmethod - def serialize_decimal(attr, **kwargs): + def serialize_decimal(attr, **kwargs): # pylint: disable=unused-argument """Serialize Decimal object to float. - :param attr: Object to be serialized. + :param decimal attr: Object to be serialized. :rtype: float + :return: serialized decimal """ return float(attr) @staticmethod - def serialize_long(attr, **kwargs): + def serialize_long(attr, **kwargs): # pylint: disable=unused-argument """Serialize long (Py2) or int (Py3). - :param attr: Object to be serialized. + :param int attr: Object to be serialized. :rtype: int/long + :return: serialized long """ return _long_type(attr) @staticmethod - def serialize_date(attr, **kwargs): + def serialize_date(attr, **kwargs): # pylint: disable=unused-argument """Serialize Date object into ISO-8601 formatted string. :param Date attr: Object to be serialized. :rtype: str + :return: serialized date """ if isinstance(attr, str): attr = isodate.parse_date(attr) @@ -1091,11 +1160,12 @@ def serialize_date(attr, **kwargs): return t @staticmethod - def serialize_time(attr, **kwargs): + def serialize_time(attr, **kwargs): # pylint: disable=unused-argument """Serialize Time object into ISO-8601 formatted string. :param datetime.time attr: Object to be serialized. :rtype: str + :return: serialized time """ if isinstance(attr, str): attr = isodate.parse_time(attr) @@ -1105,30 +1175,32 @@ def serialize_time(attr, **kwargs): return t @staticmethod - def serialize_duration(attr, **kwargs): + def serialize_duration(attr, **kwargs): # pylint: disable=unused-argument """Serialize TimeDelta object into ISO-8601 formatted string. :param TimeDelta attr: Object to be serialized. :rtype: str + :return: serialized duration """ if isinstance(attr, str): attr = isodate.parse_duration(attr) return isodate.duration_isoformat(attr) @staticmethod - def serialize_rfc(attr, **kwargs): + def serialize_rfc(attr, **kwargs): # pylint: disable=unused-argument """Serialize Datetime object into RFC-1123 formatted string. :param Datetime attr: Object to be serialized. :rtype: str :raises: TypeError if format invalid. + :return: serialized rfc """ try: if not attr.tzinfo: _LOGGER.warning("Datetime with no tzinfo will be considered UTC.") utc = attr.utctimetuple() - except AttributeError: - raise TypeError("RFC1123 object must be valid Datetime object.") + except AttributeError as exc: + raise TypeError("RFC1123 object must be valid Datetime object.") from exc return "{}, {:02} {} {:04} {:02}:{:02}:{:02} GMT".format( Serializer.days[utc.tm_wday], @@ -1141,12 +1213,13 @@ def serialize_rfc(attr, **kwargs): ) @staticmethod - def serialize_iso(attr, **kwargs): + def serialize_iso(attr, **kwargs): # pylint: disable=unused-argument """Serialize Datetime object into ISO-8601 formatted string. :param Datetime attr: Object to be serialized. :rtype: str :raises: SerializationError if format invalid. + :return: serialized iso """ if isinstance(attr, str): attr = isodate.parse_datetime(attr) @@ -1172,13 +1245,14 @@ def serialize_iso(attr, **kwargs): raise TypeError(msg) from err @staticmethod - def serialize_unix(attr, **kwargs): + def serialize_unix(attr, **kwargs): # pylint: disable=unused-argument """Serialize Datetime object into IntTime format. This is represented as seconds. :param Datetime attr: Object to be serialized. :rtype: int :raises: SerializationError if format invalid + :return: serialied unix """ if isinstance(attr, int): return attr @@ -1186,11 +1260,11 @@ def serialize_unix(attr, **kwargs): if not attr.tzinfo: _LOGGER.warning("Datetime with no tzinfo will be considered UTC.") return int(calendar.timegm(attr.utctimetuple())) - except AttributeError: - raise TypeError("Unix time object must be valid Datetime object.") + except AttributeError as exc: + raise TypeError("Unix time object must be valid Datetime object.") from exc -def rest_key_extractor(attr, attr_desc, data): +def rest_key_extractor(attr, attr_desc, data): # pylint: disable=unused-argument key = attr_desc["key"] working_data = data @@ -1211,7 +1285,9 @@ def rest_key_extractor(attr, attr_desc, data): return working_data.get(key) -def rest_key_case_insensitive_extractor(attr, attr_desc, data): +def rest_key_case_insensitive_extractor( # pylint: disable=unused-argument, inconsistent-return-statements + attr, attr_desc, data +): key = attr_desc["key"] working_data = data @@ -1232,17 +1308,29 @@ def rest_key_case_insensitive_extractor(attr, attr_desc, data): return attribute_key_case_insensitive_extractor(key, None, working_data) -def last_rest_key_extractor(attr, attr_desc, data): - """Extract the attribute in "data" based on the last part of the JSON path key.""" +def last_rest_key_extractor(attr, attr_desc, data): # pylint: disable=unused-argument + """Extract the attribute in "data" based on the last part of the JSON path key. + + :param str attr: The attribute to extract + :param dict attr_desc: The attribute description + :param dict data: The data to extract from + :rtype: object + :returns: The extracted attribute + """ key = attr_desc["key"] dict_keys = _FLATTEN.split(key) return attribute_key_extractor(dict_keys[-1], None, data) -def last_rest_key_case_insensitive_extractor(attr, attr_desc, data): +def last_rest_key_case_insensitive_extractor(attr, attr_desc, data): # pylint: disable=unused-argument """Extract the attribute in "data" based on the last part of the JSON path key. This is the case insensitive version of "last_rest_key_extractor" + :param str attr: The attribute to extract + :param dict attr_desc: The attribute description + :param dict data: The data to extract from + :rtype: object + :returns: The extracted attribute """ key = attr_desc["key"] dict_keys = _FLATTEN.split(key) @@ -1279,7 +1367,7 @@ def _extract_name_from_internal_type(internal_type): return xml_name -def xml_key_extractor(attr, attr_desc, data): +def xml_key_extractor(attr, attr_desc, data): # pylint: disable=unused-argument,too-many-return-statements if isinstance(data, dict): return None @@ -1331,22 +1419,21 @@ def xml_key_extractor(attr, attr_desc, data): if is_iter_type: if is_wrapped: return None # is_wrapped no node, we want None - else: - return [] # not wrapped, assume empty list + return [] # not wrapped, assume empty list return None # Assume it's not there, maybe an optional node. # If is_iter_type and not wrapped, return all found children if is_iter_type: if not is_wrapped: return children - else: # Iter and wrapped, should have found one node only (the wrap one) - if len(children) != 1: - raise DeserializationError( - "Tried to deserialize an array not wrapped, and found several nodes '{}'. Maybe you should declare this array as wrapped?".format( - xml_name - ) + # Iter and wrapped, should have found one node only (the wrap one) + if len(children) != 1: + raise DeserializationError( + "Tried to deserialize an array not wrapped, and found several nodes '{}'. Maybe you should declare this array as wrapped?".format( # pylint: disable=line-too-long + xml_name ) - return list(children[0]) # Might be empty list and that's ok. + ) + return list(children[0]) # Might be empty list and that's ok. # Here it's not a itertype, we should have found one element only or empty if len(children) > 1: @@ -1363,9 +1450,9 @@ class Deserializer(object): basic_types = {str: "str", int: "int", bool: "bool", float: "float"} - valid_date = re.compile(r"\d{4}[-]\d{2}[-]\d{2}T\d{2}:\d{2}:\d{2}" r"\.?\d*Z?[-+]?[\d{2}]?:?[\d{2}]?") + valid_date = re.compile(r"\d{4}[-]\d{2}[-]\d{2}T\d{2}:\d{2}:\d{2}\.?\d*Z?[-+]?[\d{2}]?:?[\d{2}]?") - def __init__(self, classes: Optional[Mapping[str, type]] = None): + def __init__(self, classes: Optional[Mapping[str, type]] = None) -> None: self.deserialize_type = { "iso-8601": Deserializer.deserialize_iso, "rfc-1123": Deserializer.deserialize_rfc, @@ -1403,11 +1490,12 @@ def __call__(self, target_obj, response_data, content_type=None): :param str content_type: Swagger "produces" if available. :raises: DeserializationError if deserialization fails. :return: Deserialized object. + :rtype: object """ data = self._unpack_content(response_data, content_type) return self._deserialize(target_obj, data) - def _deserialize(self, target_obj, data): + def _deserialize(self, target_obj, data): # pylint: disable=inconsistent-return-statements """Call the deserializer on a model. Data needs to be already deserialized as JSON or XML ElementTree @@ -1416,12 +1504,13 @@ def _deserialize(self, target_obj, data): :param object data: Object to deserialize. :raises: DeserializationError if deserialization fails. :return: Deserialized object. + :rtype: object """ # This is already a model, go recursive just in case if hasattr(data, "_attribute_map"): constants = [name for name, config in getattr(data, "_validation", {}).items() if config.get("constant")] try: - for attr, mapconfig in data._attribute_map.items(): + for attr, mapconfig in data._attribute_map.items(): # pylint: disable=protected-access if attr in constants: continue value = getattr(data, attr) @@ -1440,13 +1529,13 @@ def _deserialize(self, target_obj, data): if isinstance(response, str): return self.deserialize_data(data, response) - elif isinstance(response, type) and issubclass(response, Enum): + if isinstance(response, type) and issubclass(response, Enum): return self.deserialize_enum(data, response) if data is None or data is CoreNull: return data try: - attributes = response._attribute_map # type: ignore + attributes = response._attribute_map # type: ignore # pylint: disable=protected-access d_attrs = {} for attr, attr_desc in attributes.items(): # Check empty string. If it's not empty, someone has a real "additionalProperties"... @@ -1476,9 +1565,8 @@ def _deserialize(self, target_obj, data): except (AttributeError, TypeError, KeyError) as err: msg = "Unable to deserialize to object: " + class_name # type: ignore raise DeserializationError(msg) from err - else: - additional_properties = self._build_additional_properties(attributes, data) - return self._instantiate_model(response, d_attrs, additional_properties) + additional_properties = self._build_additional_properties(attributes, data) + return self._instantiate_model(response, d_attrs, additional_properties) def _build_additional_properties(self, attribute_map, data): if not self.additional_properties_detection: @@ -1505,6 +1593,8 @@ def _classify_target(self, target, data): :param str target: The target object type to deserialize to. :param str/dict data: The response data to deserialize. + :return: The classified target object and its class name. + :rtype: tuple """ if target is None: return None, None @@ -1516,7 +1606,7 @@ def _classify_target(self, target, data): return target, target try: - target = target._classify(data, self.dependencies) # type: ignore + target = target._classify(data, self.dependencies) # type: ignore # pylint: disable=protected-access except AttributeError: pass # Target is not a Model, no classify return target, target.__class__.__name__ # type: ignore @@ -1531,10 +1621,12 @@ def failsafe_deserialize(self, target_obj, data, content_type=None): :param str target_obj: The target object type to deserialize to. :param str/dict data: The response data to deserialize. :param str content_type: Swagger "produces" if available. + :return: Deserialized object. + :rtype: object """ try: return self(target_obj, data, content_type=content_type) - except: + except: # pylint: disable=bare-except _LOGGER.debug( "Ran into a deserialization error. Ignoring since this is failsafe deserialization", exc_info=True ) @@ -1552,10 +1644,12 @@ def _unpack_content(raw_data, content_type=None): If raw_data is something else, bypass all logic and return it directly. - :param raw_data: Data to be processed. - :param content_type: How to parse if raw_data is a string/bytes. + :param obj raw_data: Data to be processed. + :param str content_type: How to parse if raw_data is a string/bytes. :raises JSONDecodeError: If JSON is requested and parsing is impossible. :raises UnicodeDecodeError: If bytes is not UTF8 + :rtype: object + :return: Unpacked content. """ # Assume this is enough to detect a Pipeline Response without importing it context = getattr(raw_data, "context", {}) @@ -1579,14 +1673,21 @@ def _unpack_content(raw_data, content_type=None): def _instantiate_model(self, response, attrs, additional_properties=None): """Instantiate a response model passing in deserialized args. - :param response: The response model class. - :param d_attrs: The deserialized response attributes. + :param Response response: The response model class. + :param dict attrs: The deserialized response attributes. + :param dict additional_properties: Additional properties to be set. + :rtype: Response + :return: The instantiated response model. """ if callable(response): subtype = getattr(response, "_subtype_map", {}) try: - readonly = [k for k, v in response._validation.items() if v.get("readonly")] - const = [k for k, v in response._validation.items() if v.get("constant")] + readonly = [ + k for k, v in response._validation.items() if v.get("readonly") # pylint: disable=protected-access + ] + const = [ + k for k, v in response._validation.items() if v.get("constant") # pylint: disable=protected-access + ] kwargs = {k: v for k, v in attrs.items() if k not in subtype and k not in readonly + const} response_obj = response(**kwargs) for attr in readonly: @@ -1596,7 +1697,7 @@ def _instantiate_model(self, response, attrs, additional_properties=None): return response_obj except TypeError as err: msg = "Unable to deserialize {} into model {}. ".format(kwargs, response) # type: ignore - raise DeserializationError(msg + str(err)) + raise DeserializationError(msg + str(err)) from err else: try: for attr, value in attrs.items(): @@ -1605,15 +1706,16 @@ def _instantiate_model(self, response, attrs, additional_properties=None): except Exception as exp: msg = "Unable to populate response model. " msg += "Type: {}, Error: {}".format(type(response), exp) - raise DeserializationError(msg) + raise DeserializationError(msg) from exp - def deserialize_data(self, data, data_type): + def deserialize_data(self, data, data_type): # pylint: disable=too-many-return-statements """Process data for deserialization according to data type. :param str data: The response string to be deserialized. :param str data_type: The type to deserialize to. :raises: DeserializationError if deserialization fails. :return: Deserialized object. + :rtype: object """ if data is None: return data @@ -1627,7 +1729,11 @@ def deserialize_data(self, data, data_type): if isinstance(data, self.deserialize_expected_types.get(data_type, tuple())): return data - is_a_text_parsing_type = lambda x: x not in ["object", "[]", r"{}"] + is_a_text_parsing_type = lambda x: x not in [ # pylint: disable=unnecessary-lambda-assignment + "object", + "[]", + r"{}", + ] if isinstance(data, ET.Element) and is_a_text_parsing_type(data_type) and not data.text: return None data_val = self.deserialize_type[data_type](data) @@ -1647,14 +1753,14 @@ def deserialize_data(self, data, data_type): msg = "Unable to deserialize response data." msg += " Data: {}, {}".format(data, data_type) raise DeserializationError(msg) from err - else: - return self._deserialize(obj_type, data) + return self._deserialize(obj_type, data) def deserialize_iter(self, attr, iter_type): """Deserialize an iterable. :param list attr: Iterable to be deserialized. :param str iter_type: The type of object in the iterable. + :return: Deserialized iterable. :rtype: list """ if attr is None: @@ -1671,6 +1777,7 @@ def deserialize_dict(self, attr, dict_type): :param dict/list attr: Dictionary to be deserialized. Also accepts a list of key, value pairs. :param str dict_type: The object type of the items in the dictionary. + :return: Deserialized dictionary. :rtype: dict """ if isinstance(attr, list): @@ -1681,11 +1788,12 @@ def deserialize_dict(self, attr, dict_type): attr = {el.tag: el.text for el in attr} return {k: self.deserialize_data(v, dict_type) for k, v in attr.items()} - def deserialize_object(self, attr, **kwargs): + def deserialize_object(self, attr, **kwargs): # pylint: disable=too-many-return-statements """Deserialize a generic object. This will be handled as a dictionary. :param dict attr: Dictionary to be deserialized. + :return: Deserialized object. :rtype: dict :raises: TypeError if non-builtin datatype encountered. """ @@ -1720,11 +1828,10 @@ def deserialize_object(self, attr, **kwargs): pass return deserialized - else: - error = "Cannot deserialize generic object with type: " - raise TypeError(error + str(obj_type)) + error = "Cannot deserialize generic object with type: " + raise TypeError(error + str(obj_type)) - def deserialize_basic(self, attr, data_type): + def deserialize_basic(self, attr, data_type): # pylint: disable=too-many-return-statements """Deserialize basic builtin data type from string. Will attempt to convert to str, int, float and bool. This function will also accept '1', '0', 'true' and 'false' as @@ -1732,6 +1839,7 @@ def deserialize_basic(self, attr, data_type): :param str attr: response string to be deserialized. :param str data_type: deserialization data type. + :return: Deserialized basic type. :rtype: str, int, float or bool :raises: TypeError if string format is not valid. """ @@ -1743,24 +1851,23 @@ def deserialize_basic(self, attr, data_type): if data_type == "str": # None or '', node is empty string. return "" - else: - # None or '', node with a strong type is None. - # Don't try to model "empty bool" or "empty int" - return None + # None or '', node with a strong type is None. + # Don't try to model "empty bool" or "empty int" + return None if data_type == "bool": if attr in [True, False, 1, 0]: return bool(attr) - elif isinstance(attr, str): + if isinstance(attr, str): if attr.lower() in ["true", "1"]: return True - elif attr.lower() in ["false", "0"]: + if attr.lower() in ["false", "0"]: return False raise TypeError("Invalid boolean value: {}".format(attr)) if data_type == "str": return self.deserialize_unicode(attr) - return eval(data_type)(attr) # nosec + return eval(data_type)(attr) # nosec # pylint: disable=eval-used @staticmethod def deserialize_unicode(data): @@ -1768,6 +1875,7 @@ def deserialize_unicode(data): as a string. :param str data: response string to be deserialized. + :return: Deserialized string. :rtype: str or unicode """ # We might be here because we have an enum modeled as string, @@ -1781,8 +1889,7 @@ def deserialize_unicode(data): return data except NameError: return str(data) - else: - return str(data) + return str(data) @staticmethod def deserialize_enum(data, enum_obj): @@ -1794,6 +1901,7 @@ def deserialize_enum(data, enum_obj): :param str data: Response string to be deserialized. If this value is None or invalid it will be returned as-is. :param Enum enum_obj: Enum object to deserialize to. + :return: Deserialized enum object. :rtype: Enum """ if isinstance(data, enum_obj) or data is None: @@ -1804,9 +1912,9 @@ def deserialize_enum(data, enum_obj): # Workaround. We might consider remove it in the future. try: return list(enum_obj.__members__.values())[data] - except IndexError: + except IndexError as exc: error = "{!r} is not a valid index for enum {!r}" - raise DeserializationError(error.format(data, enum_obj)) + raise DeserializationError(error.format(data, enum_obj)) from exc try: return enum_obj(str(data)) except ValueError: @@ -1822,6 +1930,7 @@ def deserialize_bytearray(attr): """Deserialize string into bytearray. :param str attr: response string to be deserialized. + :return: Deserialized bytearray :rtype: bytearray :raises: TypeError if string format invalid. """ @@ -1834,6 +1943,7 @@ def deserialize_base64(attr): """Deserialize base64 encoded string into string. :param str attr: response string to be deserialized. + :return: Deserialized base64 string :rtype: bytearray :raises: TypeError if string format invalid. """ @@ -1849,8 +1959,9 @@ def deserialize_decimal(attr): """Deserialize string into Decimal object. :param str attr: response string to be deserialized. - :rtype: Decimal + :return: Deserialized decimal :raises: DeserializationError if string format invalid. + :rtype: decimal """ if isinstance(attr, ET.Element): attr = attr.text @@ -1865,6 +1976,7 @@ def deserialize_long(attr): """Deserialize string into long (Py2) or int (Py3). :param str attr: response string to be deserialized. + :return: Deserialized int :rtype: long or int :raises: ValueError if string format invalid. """ @@ -1877,6 +1989,7 @@ def deserialize_duration(attr): """Deserialize ISO-8601 formatted string into TimeDelta object. :param str attr: response string to be deserialized. + :return: Deserialized duration :rtype: TimeDelta :raises: DeserializationError if string format invalid. """ @@ -1887,14 +2000,14 @@ def deserialize_duration(attr): except (ValueError, OverflowError, AttributeError) as err: msg = "Cannot deserialize duration object." raise DeserializationError(msg) from err - else: - return duration + return duration @staticmethod def deserialize_date(attr): """Deserialize ISO-8601 formatted string into Date object. :param str attr: response string to be deserialized. + :return: Deserialized date :rtype: Date :raises: DeserializationError if string format invalid. """ @@ -1910,6 +2023,7 @@ def deserialize_time(attr): """Deserialize ISO-8601 formatted string into time object. :param str attr: response string to be deserialized. + :return: Deserialized time :rtype: datetime.time :raises: DeserializationError if string format invalid. """ @@ -1924,6 +2038,7 @@ def deserialize_rfc(attr): """Deserialize RFC-1123 formatted string into Datetime object. :param str attr: response string to be deserialized. + :return: Deserialized RFC datetime :rtype: Datetime :raises: DeserializationError if string format invalid. """ @@ -1939,14 +2054,14 @@ def deserialize_rfc(attr): except ValueError as err: msg = "Cannot deserialize to rfc datetime object." raise DeserializationError(msg) from err - else: - return date_obj + return date_obj @staticmethod def deserialize_iso(attr): """Deserialize ISO-8601 formatted string into Datetime object. :param str attr: response string to be deserialized. + :return: Deserialized ISO datetime :rtype: Datetime :raises: DeserializationError if string format invalid. """ @@ -1976,8 +2091,7 @@ def deserialize_iso(attr): except (ValueError, OverflowError, AttributeError) as err: msg = "Cannot deserialize datetime object." raise DeserializationError(msg) from err - else: - return date_obj + return date_obj @staticmethod def deserialize_unix(attr): @@ -1985,6 +2099,7 @@ def deserialize_unix(attr): This is represented as seconds. :param int attr: Object to be serialized. + :return: Deserialized datetime :rtype: Datetime :raises: DeserializationError if format invalid """ @@ -1996,5 +2111,4 @@ def deserialize_unix(attr): except ValueError as err: msg = "Cannot deserialize to unix datetime object." raise DeserializationError(msg) from err - else: - return date_obj + return date_obj diff --git a/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/_vendor.py b/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/_vendor.py index 6dbcb5c20a91..5af45cbb9df5 100644 --- a/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/_vendor.py +++ b/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/_vendor.py @@ -11,7 +11,6 @@ from ._configuration import DeidentificationClientConfiguration if TYPE_CHECKING: - # pylint: disable=unused-import,ungrouped-imports from azure.core import PipelineClient from ._serialization import Deserializer, Serializer diff --git a/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/_version.py b/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/_version.py index bbcd28b4aa67..be71c81bd282 100644 --- a/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/_version.py +++ b/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/_version.py @@ -6,4 +6,4 @@ # Changes may cause incorrect behavior and will be lost if the code is regenerated. # -------------------------------------------------------------------------- -VERSION = "1.0.0b2" +VERSION = "1.0.0b1" diff --git a/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/aio/__init__.py b/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/aio/__init__.py index 245e207d364a..432fe8a82dba 100644 --- a/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/aio/__init__.py +++ b/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/aio/__init__.py @@ -5,12 +5,18 @@ # Code generated by Microsoft (R) Python Code Generator. # Changes may cause incorrect behavior and will be lost if the code is regenerated. # -------------------------------------------------------------------------- +# pylint: disable=wrong-import-position -from ._client import DeidentificationClient +from typing import TYPE_CHECKING + +if TYPE_CHECKING: + from ._patch import * # pylint: disable=unused-wildcard-import + +from ._client import DeidentificationClient # type: ignore try: from ._patch import __all__ as _patch_all - from ._patch import * # pylint: disable=unused-wildcard-import + from ._patch import * except ImportError: _patch_all = [] from ._patch import patch_sdk as _patch_sdk @@ -18,6 +24,6 @@ __all__ = [ "DeidentificationClient", ] -__all__.extend([p for p in _patch_all if p not in __all__]) +__all__.extend([p for p in _patch_all if p not in __all__]) # pyright: ignore _patch_sdk() diff --git a/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/aio/_client.py b/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/aio/_client.py index b257b9201e01..b1d2af86fbc9 100644 --- a/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/aio/_client.py +++ b/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/aio/_client.py @@ -19,29 +19,25 @@ from ._operations import DeidentificationClientOperationsMixin if TYPE_CHECKING: - # pylint: disable=unused-import,ungrouped-imports from azure.core.credentials_async import AsyncTokenCredential -class DeidentificationClient( - DeidentificationClientOperationsMixin -): # pylint: disable=client-accepts-api-version-keyword +class DeidentificationClient(DeidentificationClientOperationsMixin): """DeidentificationClient. :param endpoint: Url of your De-identification Service. Required. :type endpoint: str :param credential: Credential used to authenticate requests to the service. Required. :type credential: ~azure.core.credentials_async.AsyncTokenCredential - :keyword api_version: The API version to use for this operation. Default value is - "2024-07-12-preview". Note that overriding this default value may result in unsupported - behavior. + :keyword api_version: The API version to use for this operation. Default value is "2024-11-15". + Note that overriding this default value may result in unsupported behavior. :paramtype api_version: str :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. """ def __init__(self, endpoint: str, credential: "AsyncTokenCredential", **kwargs: Any) -> None: - _endpoint = "https://{endpoint}" + _endpoint = "{endpoint}" self._config = DeidentificationClientConfiguration(endpoint=endpoint, credential=credential, **kwargs) _policies = kwargs.pop("policies", None) if _policies is None: @@ -88,7 +84,7 @@ def send_request( request_copy = deepcopy(request) path_format_arguments = { - "endpoint": self._serialize.url("self._config.endpoint", self._config.endpoint, "str"), + "endpoint": self._serialize.url("self._config.endpoint", self._config.endpoint, "str", skip_quote=True), } request_copy.url = self._client.format_url(request_copy.url, **path_format_arguments) diff --git a/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/aio/_configuration.py b/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/aio/_configuration.py index 3799c4c1d7b2..199a4c5e3aee 100644 --- a/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/aio/_configuration.py +++ b/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/aio/_configuration.py @@ -13,11 +13,10 @@ from .._version import VERSION if TYPE_CHECKING: - # pylint: disable=unused-import,ungrouped-imports from azure.core.credentials_async import AsyncTokenCredential -class DeidentificationClientConfiguration: # pylint: disable=too-many-instance-attributes,name-too-long +class DeidentificationClientConfiguration: # pylint: disable=too-many-instance-attributes """Configuration for DeidentificationClient. Note that all parameters used to create this instance are saved as instance @@ -27,14 +26,13 @@ class DeidentificationClientConfiguration: # pylint: disable=too-many-instance- :type endpoint: str :param credential: Credential used to authenticate requests to the service. Required. :type credential: ~azure.core.credentials_async.AsyncTokenCredential - :keyword api_version: The API version to use for this operation. Default value is - "2024-07-12-preview". Note that overriding this default value may result in unsupported - behavior. + :keyword api_version: The API version to use for this operation. Default value is "2024-11-15". + Note that overriding this default value may result in unsupported behavior. :paramtype api_version: str """ def __init__(self, endpoint: str, credential: "AsyncTokenCredential", **kwargs: Any) -> None: - api_version: str = kwargs.pop("api_version", "2024-07-12-preview") + api_version: str = kwargs.pop("api_version", "2024-11-15") if endpoint is None: raise ValueError("Parameter 'endpoint' must not be None.") diff --git a/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/aio/_operations/__init__.py b/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/aio/_operations/__init__.py index f30b11092e89..8a3952cdf768 100644 --- a/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/aio/_operations/__init__.py +++ b/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/aio/_operations/__init__.py @@ -5,15 +5,21 @@ # Code generated by Microsoft (R) Python Code Generator. # Changes may cause incorrect behavior and will be lost if the code is regenerated. # -------------------------------------------------------------------------- +# pylint: disable=wrong-import-position -from ._operations import DeidentificationClientOperationsMixin +from typing import TYPE_CHECKING + +if TYPE_CHECKING: + from ._patch import * # pylint: disable=unused-wildcard-import + +from ._operations import DeidentificationClientOperationsMixin # type: ignore from ._patch import __all__ as _patch_all -from ._patch import * # pylint: disable=unused-wildcard-import +from ._patch import * from ._patch import patch_sdk as _patch_sdk __all__ = [ "DeidentificationClientOperationsMixin", ] -__all__.extend([p for p in _patch_all if p not in __all__]) +__all__.extend([p for p in _patch_all if p not in __all__]) # pyright: ignore _patch_sdk() diff --git a/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/aio/_operations/_operations.py b/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/aio/_operations/_operations.py index dd3a7c4dbdcb..da3f26d33411 100644 --- a/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/aio/_operations/_operations.py +++ b/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/aio/_operations/_operations.py @@ -1,4 +1,3 @@ -# pylint: disable=too-many-lines,too-many-statements # coding=utf-8 # -------------------------------------------------------------------------- # Copyright (c) Microsoft Corporation. All rights reserved. @@ -9,21 +8,7 @@ from io import IOBase import json import sys -from typing import ( - Any, - AsyncIterable, - AsyncIterator, - Callable, - Dict, - IO, - List, - Optional, - Type, - TypeVar, - Union, - cast, - overload, -) +from typing import Any, AsyncIterable, AsyncIterator, Callable, Dict, IO, List, Optional, TypeVar, Union, cast, overload import urllib.parse from azure.core.async_paging import AsyncItemPaged, AsyncList @@ -33,6 +18,8 @@ ResourceExistsError, ResourceNotFoundError, ResourceNotModifiedError, + StreamClosedError, + StreamConsumedError, map_error, ) from azure.core.pipeline import PipelineResponse @@ -47,19 +34,19 @@ from ..._model_base import SdkJSONEncoder, _deserialize from ..._operations._operations import ( build_deidentification_cancel_job_request, - build_deidentification_create_job_request, - build_deidentification_deidentify_request, + build_deidentification_deidentify_documents_request, + build_deidentification_deidentify_text_request, build_deidentification_delete_job_request, build_deidentification_get_job_request, - build_deidentification_list_job_documents_request, - build_deidentification_list_jobs_request, + build_deidentification_list_job_documents_internal_request, + build_deidentification_list_jobs_internal_request, ) from .._vendor import DeidentificationClientMixinABC if sys.version_info >= (3, 9): from collections.abc import MutableMapping else: - from typing import MutableMapping # type: ignore # pylint: disable=ungrouped-imports + from typing import MutableMapping # type: ignore JSON = MutableMapping[str, Any] # pylint: disable=unsubscriptable-object T = TypeVar("T") ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, Dict[str, Any]], Any]] @@ -78,53 +65,8 @@ async def get_job(self, name: str, **kwargs: Any) -> _models.DeidentificationJob :return: DeidentificationJob. The DeidentificationJob is compatible with MutableMapping :rtype: ~azure.health.deidentification.models.DeidentificationJob :raises ~azure.core.exceptions.HttpResponseError: - - Example: - .. code-block:: python - - # response body for status code(s): 200 - response == { - "createdAt": "2020-02-20 00:00:00", - "lastUpdatedAt": "2020-02-20 00:00:00", - "name": "str", - "sourceLocation": { - "location": "str", - "prefix": "str", - "extensions": [ - "str" - ] - }, - "status": "str", - "targetLocation": { - "location": "str", - "prefix": "str" - }, - "dataType": "str", - "error": { - "code": "str", - "message": "str", - "details": [ - ... - ], - "innererror": { - "code": "str", - "innererror": ... - }, - "target": "str" - }, - "operation": "str", - "redactionFormat": "str", - "startedAt": "2020-02-20 00:00:00", - "summary": { - "bytesProcessed": 0, - "canceled": 0, - "failed": 0, - "successful": 0, - "total": 0 - } - } """ - error_map: MutableMapping[int, Type[HttpResponseError]] = { + error_map: MutableMapping = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError, @@ -144,7 +86,7 @@ async def get_job(self, name: str, **kwargs: Any) -> _models.DeidentificationJob params=_params, ) path_format_arguments = { - "endpoint": self._serialize.url("self._config.endpoint", self._config.endpoint, "str"), + "endpoint": self._serialize.url("self._config.endpoint", self._config.endpoint, "str", skip_quote=True), } _request.url = self._client.format_url(_request.url, **path_format_arguments) @@ -157,7 +99,10 @@ async def get_job(self, name: str, **kwargs: Any) -> _models.DeidentificationJob if response.status_code not in [200]: if _stream: - await response.read() # Load the body in memory and close the socket + try: + await response.read() # Load the body in memory and close the socket + except (StreamConsumedError, StreamClosedError): + pass map_error(status_code=response.status_code, response=response, error_map=error_map) raise HttpResponseError(response=response) @@ -176,10 +121,10 @@ async def get_job(self, name: str, **kwargs: Any) -> _models.DeidentificationJob return deserialized # type: ignore - async def _create_job_initial( + async def _deidentify_documents_initial( self, name: str, resource: Union[_models.DeidentificationJob, JSON, IO[bytes]], **kwargs: Any ) -> AsyncIterator[bytes]: - error_map: MutableMapping[int, Type[HttpResponseError]] = { + error_map: MutableMapping = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError, @@ -200,7 +145,7 @@ async def _create_job_initial( else: _content = json.dumps(resource, cls=SdkJSONEncoder, exclude_readonly=True) # type: ignore - _request = build_deidentification_create_job_request( + _request = build_deidentification_deidentify_documents_request( name=name, content_type=content_type, api_version=self._config.api_version, @@ -209,7 +154,7 @@ async def _create_job_initial( params=_params, ) path_format_arguments = { - "endpoint": self._serialize.url("self._config.endpoint", self._config.endpoint, "str"), + "endpoint": self._serialize.url("self._config.endpoint", self._config.endpoint, "str", skip_quote=True), } _request.url = self._client.format_url(_request.url, **path_format_arguments) @@ -221,30 +166,20 @@ async def _create_job_initial( response = pipeline_response.http_response if response.status_code not in [200, 201]: - await response.read() # Load the body in memory and close the socket + try: + await response.read() # Load the body in memory and close the socket + except (StreamConsumedError, StreamClosedError): + pass map_error(status_code=response.status_code, response=response, error_map=error_map) raise HttpResponseError(response=response) response_headers = {} - if response.status_code == 200: - response_headers["x-ms-client-request-id"] = self._deserialize( - "str", response.headers.get("x-ms-client-request-id") - ) - response_headers["Operation-Location"] = self._deserialize( - "str", response.headers.get("Operation-Location") - ) - - deserialized = response.iter_bytes() - - if response.status_code == 201: - response_headers["x-ms-client-request-id"] = self._deserialize( - "str", response.headers.get("x-ms-client-request-id") - ) - response_headers["Operation-Location"] = self._deserialize( - "str", response.headers.get("Operation-Location") - ) + response_headers["x-ms-client-request-id"] = self._deserialize( + "str", response.headers.get("x-ms-client-request-id") + ) + response_headers["Operation-Location"] = self._deserialize("str", response.headers.get("Operation-Location")) - deserialized = response.iter_bytes() + deserialized = response.iter_bytes() if cls: return cls(pipeline_response, deserialized, response_headers) # type: ignore @@ -252,7 +187,7 @@ async def _create_job_initial( return deserialized # type: ignore @overload - async def begin_create_job( + async def begin_deidentify_documents( self, name: str, resource: _models.DeidentificationJob, *, content_type: str = "application/json", **kwargs: Any ) -> AsyncLROPoller[_models.DeidentificationJob]: """Create a de-identification job. @@ -271,97 +206,10 @@ async def begin_create_job( :rtype: ~azure.core.polling.AsyncLROPoller[~azure.health.deidentification.models.DeidentificationJob] :raises ~azure.core.exceptions.HttpResponseError: - - Example: - .. code-block:: python - - # JSON input template you can fill out and use as your body input. - resource = { - "createdAt": "2020-02-20 00:00:00", - "lastUpdatedAt": "2020-02-20 00:00:00", - "name": "str", - "sourceLocation": { - "location": "str", - "prefix": "str", - "extensions": [ - "str" - ] - }, - "status": "str", - "targetLocation": { - "location": "str", - "prefix": "str" - }, - "dataType": "str", - "error": { - "code": "str", - "message": "str", - "details": [ - ... - ], - "innererror": { - "code": "str", - "innererror": ... - }, - "target": "str" - }, - "operation": "str", - "redactionFormat": "str", - "startedAt": "2020-02-20 00:00:00", - "summary": { - "bytesProcessed": 0, - "canceled": 0, - "failed": 0, - "successful": 0, - "total": 0 - } - } - - # response body for status code(s): 201, 200 - response == { - "createdAt": "2020-02-20 00:00:00", - "lastUpdatedAt": "2020-02-20 00:00:00", - "name": "str", - "sourceLocation": { - "location": "str", - "prefix": "str", - "extensions": [ - "str" - ] - }, - "status": "str", - "targetLocation": { - "location": "str", - "prefix": "str" - }, - "dataType": "str", - "error": { - "code": "str", - "message": "str", - "details": [ - ... - ], - "innererror": { - "code": "str", - "innererror": ... - }, - "target": "str" - }, - "operation": "str", - "redactionFormat": "str", - "startedAt": "2020-02-20 00:00:00", - "summary": { - "bytesProcessed": 0, - "canceled": 0, - "failed": 0, - "successful": 0, - "total": 0 - } - } """ @overload - async def begin_create_job( + async def begin_deidentify_documents( self, name: str, resource: JSON, *, content_type: str = "application/json", **kwargs: Any ) -> AsyncLROPoller[_models.DeidentificationJob]: """Create a de-identification job. @@ -380,55 +228,10 @@ async def begin_create_job( :rtype: ~azure.core.polling.AsyncLROPoller[~azure.health.deidentification.models.DeidentificationJob] :raises ~azure.core.exceptions.HttpResponseError: - - Example: - .. code-block:: python - - # response body for status code(s): 201, 200 - response == { - "createdAt": "2020-02-20 00:00:00", - "lastUpdatedAt": "2020-02-20 00:00:00", - "name": "str", - "sourceLocation": { - "location": "str", - "prefix": "str", - "extensions": [ - "str" - ] - }, - "status": "str", - "targetLocation": { - "location": "str", - "prefix": "str" - }, - "dataType": "str", - "error": { - "code": "str", - "message": "str", - "details": [ - ... - ], - "innererror": { - "code": "str", - "innererror": ... - }, - "target": "str" - }, - "operation": "str", - "redactionFormat": "str", - "startedAt": "2020-02-20 00:00:00", - "summary": { - "bytesProcessed": 0, - "canceled": 0, - "failed": 0, - "successful": 0, - "total": 0 - } - } """ @overload - async def begin_create_job( + async def begin_deidentify_documents( self, name: str, resource: IO[bytes], *, content_type: str = "application/json", **kwargs: Any ) -> AsyncLROPoller[_models.DeidentificationJob]: """Create a de-identification job. @@ -447,55 +250,10 @@ async def begin_create_job( :rtype: ~azure.core.polling.AsyncLROPoller[~azure.health.deidentification.models.DeidentificationJob] :raises ~azure.core.exceptions.HttpResponseError: - - Example: - .. code-block:: python - - # response body for status code(s): 201, 200 - response == { - "createdAt": "2020-02-20 00:00:00", - "lastUpdatedAt": "2020-02-20 00:00:00", - "name": "str", - "sourceLocation": { - "location": "str", - "prefix": "str", - "extensions": [ - "str" - ] - }, - "status": "str", - "targetLocation": { - "location": "str", - "prefix": "str" - }, - "dataType": "str", - "error": { - "code": "str", - "message": "str", - "details": [ - ... - ], - "innererror": { - "code": "str", - "innererror": ... - }, - "target": "str" - }, - "operation": "str", - "redactionFormat": "str", - "startedAt": "2020-02-20 00:00:00", - "summary": { - "bytesProcessed": 0, - "canceled": 0, - "failed": 0, - "successful": 0, - "total": 0 - } - } """ @distributed_trace_async - async def begin_create_job( + async def begin_deidentify_documents( self, name: str, resource: Union[_models.DeidentificationJob, JSON, IO[bytes]], **kwargs: Any ) -> AsyncLROPoller[_models.DeidentificationJob]: """Create a de-identification job. @@ -512,93 +270,6 @@ async def begin_create_job( :rtype: ~azure.core.polling.AsyncLROPoller[~azure.health.deidentification.models.DeidentificationJob] :raises ~azure.core.exceptions.HttpResponseError: - - Example: - .. code-block:: python - - # JSON input template you can fill out and use as your body input. - resource = { - "createdAt": "2020-02-20 00:00:00", - "lastUpdatedAt": "2020-02-20 00:00:00", - "name": "str", - "sourceLocation": { - "location": "str", - "prefix": "str", - "extensions": [ - "str" - ] - }, - "status": "str", - "targetLocation": { - "location": "str", - "prefix": "str" - }, - "dataType": "str", - "error": { - "code": "str", - "message": "str", - "details": [ - ... - ], - "innererror": { - "code": "str", - "innererror": ... - }, - "target": "str" - }, - "operation": "str", - "redactionFormat": "str", - "startedAt": "2020-02-20 00:00:00", - "summary": { - "bytesProcessed": 0, - "canceled": 0, - "failed": 0, - "successful": 0, - "total": 0 - } - } - - # response body for status code(s): 201, 200 - response == { - "createdAt": "2020-02-20 00:00:00", - "lastUpdatedAt": "2020-02-20 00:00:00", - "name": "str", - "sourceLocation": { - "location": "str", - "prefix": "str", - "extensions": [ - "str" - ] - }, - "status": "str", - "targetLocation": { - "location": "str", - "prefix": "str" - }, - "dataType": "str", - "error": { - "code": "str", - "message": "str", - "details": [ - ... - ], - "innererror": { - "code": "str", - "innererror": ... - }, - "target": "str" - }, - "operation": "str", - "redactionFormat": "str", - "startedAt": "2020-02-20 00:00:00", - "summary": { - "bytesProcessed": 0, - "canceled": 0, - "failed": 0, - "successful": 0, - "total": 0 - } - } """ _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) _params = kwargs.pop("params", {}) or {} @@ -609,7 +280,7 @@ async def begin_create_job( lro_delay = kwargs.pop("polling_interval", self._config.polling_interval) cont_token: Optional[str] = kwargs.pop("continuation_token", None) if cont_token is None: - raw_result = await self._create_job_initial( + raw_result = await self._deidentify_documents_initial( name=name, resource=resource, content_type=content_type, @@ -637,7 +308,7 @@ def get_long_running_output(pipeline_response): return deserialized path_format_arguments = { - "endpoint": self._serialize.url("self._config.endpoint", self._config.endpoint, "str"), + "endpoint": self._serialize.url("self._config.endpoint", self._config.endpoint, "str", skip_quote=True), } if polling is True: @@ -661,7 +332,7 @@ def get_long_running_output(pipeline_response): ) @distributed_trace - def list_jobs( + def _list_jobs_internal( self, *, continuation_token_parameter: Optional[str] = None, **kwargs: Any ) -> AsyncIterable["_models.DeidentificationJob"]: """List de-identification jobs. @@ -675,51 +346,6 @@ def list_jobs( :rtype: ~azure.core.async_paging.AsyncItemPaged[~azure.health.deidentification.models.DeidentificationJob] :raises ~azure.core.exceptions.HttpResponseError: - - Example: - .. code-block:: python - - # response body for status code(s): 200 - response == { - "createdAt": "2020-02-20 00:00:00", - "lastUpdatedAt": "2020-02-20 00:00:00", - "name": "str", - "sourceLocation": { - "location": "str", - "prefix": "str", - "extensions": [ - "str" - ] - }, - "status": "str", - "targetLocation": { - "location": "str", - "prefix": "str" - }, - "dataType": "str", - "error": { - "code": "str", - "message": "str", - "details": [ - ... - ], - "innererror": { - "code": "str", - "innererror": ... - }, - "target": "str" - }, - "operation": "str", - "redactionFormat": "str", - "startedAt": "2020-02-20 00:00:00", - "summary": { - "bytesProcessed": 0, - "canceled": 0, - "failed": 0, - "successful": 0, - "total": 0 - } - } """ _headers = kwargs.pop("headers", {}) or {} _params = kwargs.pop("params", {}) or {} @@ -727,7 +353,7 @@ def list_jobs( maxpagesize = kwargs.pop("maxpagesize", None) cls: ClsType[List[_models.DeidentificationJob]] = kwargs.pop("cls", None) - error_map: MutableMapping[int, Type[HttpResponseError]] = { + error_map: MutableMapping = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError, @@ -738,7 +364,7 @@ def list_jobs( def prepare_request(next_link=None): if not next_link: - _request = build_deidentification_list_jobs_request( + _request = build_deidentification_list_jobs_internal_request( maxpagesize=maxpagesize, continuation_token_parameter=continuation_token_parameter, api_version=self._config.api_version, @@ -746,7 +372,9 @@ def prepare_request(next_link=None): params=_params, ) path_format_arguments = { - "endpoint": self._serialize.url("self._config.endpoint", self._config.endpoint, "str"), + "endpoint": self._serialize.url( + "self._config.endpoint", self._config.endpoint, "str", skip_quote=True + ), } _request.url = self._client.format_url(_request.url, **path_format_arguments) @@ -764,7 +392,9 @@ def prepare_request(next_link=None): "GET", urllib.parse.urljoin(next_link, _parsed_next_link.path), params=_next_request_params ) path_format_arguments = { - "endpoint": self._serialize.url("self._config.endpoint", self._config.endpoint, "str"), + "endpoint": self._serialize.url( + "self._config.endpoint", self._config.endpoint, "str", skip_quote=True + ), } _request.url = self._client.format_url(_request.url, **path_format_arguments) @@ -795,59 +425,30 @@ async def get_next(next_link=None): return AsyncItemPaged(get_next, extract_data) @distributed_trace - def list_job_documents( - self, name: str, *, continuation_token_parameter: Optional[str] = None, **kwargs: Any - ) -> AsyncIterable["_models.DocumentDetails"]: + def _list_job_documents_internal( + self, job_name: str, *, continuation_token_parameter: Optional[str] = None, **kwargs: Any + ) -> AsyncIterable["_models.DeidentificationDocumentDetails"]: """List processed documents within a job. - Resource list operation template. + The most basic operation. - :param name: The name of a job. Required. - :type name: str + :param job_name: The name of a job. Required. + :type job_name: str :keyword continuation_token_parameter: Token to continue a previous query. Default value is None. :paramtype continuation_token_parameter: str - :return: An iterator like instance of DocumentDetails + :return: An iterator like instance of DeidentificationDocumentDetails :rtype: - ~azure.core.async_paging.AsyncItemPaged[~azure.health.deidentification.models.DocumentDetails] + ~azure.core.async_paging.AsyncItemPaged[~azure.health.deidentification.models.DeidentificationDocumentDetails] :raises ~azure.core.exceptions.HttpResponseError: - - Example: - .. code-block:: python - - # response body for status code(s): 200 - response == { - "id": "str", - "input": { - "etag": "str", - "path": "str" - }, - "status": "str", - "error": { - "code": "str", - "message": "str", - "details": [ - ... - ], - "innererror": { - "code": "str", - "innererror": ... - }, - "target": "str" - }, - "output": { - "etag": "str", - "path": "str" - } - } """ _headers = kwargs.pop("headers", {}) or {} _params = kwargs.pop("params", {}) or {} maxpagesize = kwargs.pop("maxpagesize", None) - cls: ClsType[List[_models.DocumentDetails]] = kwargs.pop("cls", None) + cls: ClsType[List[_models.DeidentificationDocumentDetails]] = kwargs.pop("cls", None) - error_map: MutableMapping[int, Type[HttpResponseError]] = { + error_map: MutableMapping = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError, @@ -858,8 +459,8 @@ def list_job_documents( def prepare_request(next_link=None): if not next_link: - _request = build_deidentification_list_job_documents_request( - name=name, + _request = build_deidentification_list_job_documents_internal_request( + job_name=job_name, maxpagesize=maxpagesize, continuation_token_parameter=continuation_token_parameter, api_version=self._config.api_version, @@ -867,7 +468,9 @@ def prepare_request(next_link=None): params=_params, ) path_format_arguments = { - "endpoint": self._serialize.url("self._config.endpoint", self._config.endpoint, "str"), + "endpoint": self._serialize.url( + "self._config.endpoint", self._config.endpoint, "str", skip_quote=True + ), } _request.url = self._client.format_url(_request.url, **path_format_arguments) @@ -885,7 +488,9 @@ def prepare_request(next_link=None): "GET", urllib.parse.urljoin(next_link, _parsed_next_link.path), params=_next_request_params ) path_format_arguments = { - "endpoint": self._serialize.url("self._config.endpoint", self._config.endpoint, "str"), + "endpoint": self._serialize.url( + "self._config.endpoint", self._config.endpoint, "str", skip_quote=True + ), } _request.url = self._client.format_url(_request.url, **path_format_arguments) @@ -893,7 +498,7 @@ def prepare_request(next_link=None): async def extract_data(pipeline_response): deserialized = pipeline_response.http_response.json() - list_of_elem = _deserialize(List[_models.DocumentDetails], deserialized["value"]) + list_of_elem = _deserialize(List[_models.DeidentificationDocumentDetails], deserialized["value"]) if cls: list_of_elem = cls(list_of_elem) # type: ignore return deserialized.get("nextLink") or None, AsyncList(list_of_elem) @@ -931,53 +536,8 @@ async def cancel_job(self, name: str, **kwargs: Any) -> _models.Deidentification :return: DeidentificationJob. The DeidentificationJob is compatible with MutableMapping :rtype: ~azure.health.deidentification.models.DeidentificationJob :raises ~azure.core.exceptions.HttpResponseError: - - Example: - .. code-block:: python - - # response body for status code(s): 200 - response == { - "createdAt": "2020-02-20 00:00:00", - "lastUpdatedAt": "2020-02-20 00:00:00", - "name": "str", - "sourceLocation": { - "location": "str", - "prefix": "str", - "extensions": [ - "str" - ] - }, - "status": "str", - "targetLocation": { - "location": "str", - "prefix": "str" - }, - "dataType": "str", - "error": { - "code": "str", - "message": "str", - "details": [ - ... - ], - "innererror": { - "code": "str", - "innererror": ... - }, - "target": "str" - }, - "operation": "str", - "redactionFormat": "str", - "startedAt": "2020-02-20 00:00:00", - "summary": { - "bytesProcessed": 0, - "canceled": 0, - "failed": 0, - "successful": 0, - "total": 0 - } - } """ - error_map: MutableMapping[int, Type[HttpResponseError]] = { + error_map: MutableMapping = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError, @@ -997,7 +557,7 @@ async def cancel_job(self, name: str, **kwargs: Any) -> _models.Deidentification params=_params, ) path_format_arguments = { - "endpoint": self._serialize.url("self._config.endpoint", self._config.endpoint, "str"), + "endpoint": self._serialize.url("self._config.endpoint", self._config.endpoint, "str", skip_quote=True), } _request.url = self._client.format_url(_request.url, **path_format_arguments) @@ -1010,7 +570,10 @@ async def cancel_job(self, name: str, **kwargs: Any) -> _models.Deidentification if response.status_code not in [200]: if _stream: - await response.read() # Load the body in memory and close the socket + try: + await response.read() # Load the body in memory and close the socket + except (StreamConsumedError, StreamClosedError): + pass map_error(status_code=response.status_code, response=response, error_map=error_map) raise HttpResponseError(response=response) @@ -1030,7 +593,7 @@ async def cancel_job(self, name: str, **kwargs: Any) -> _models.Deidentification return deserialized # type: ignore @distributed_trace_async - async def delete_job(self, name: str, **kwargs: Any) -> None: # pylint: disable=inconsistent-return-statements + async def delete_job(self, name: str, **kwargs: Any) -> None: """Delete a de-identification job. Removes the record of the job from the service. Does not delete any documents. @@ -1041,7 +604,7 @@ async def delete_job(self, name: str, **kwargs: Any) -> None: # pylint: disable :rtype: None :raises ~azure.core.exceptions.HttpResponseError: """ - error_map: MutableMapping[int, Type[HttpResponseError]] = { + error_map: MutableMapping = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError, @@ -1061,7 +624,7 @@ async def delete_job(self, name: str, **kwargs: Any) -> None: # pylint: disable params=_params, ) path_format_arguments = { - "endpoint": self._serialize.url("self._config.endpoint", self._config.endpoint, "str"), + "endpoint": self._serialize.url("self._config.endpoint", self._config.endpoint, "str", skip_quote=True), } _request.url = self._client.format_url(_request.url, **path_format_arguments) @@ -1085,7 +648,7 @@ async def delete_job(self, name: str, **kwargs: Any) -> None: # pylint: disable return cls(pipeline_response, None, response_headers) # type: ignore @overload - async def deidentify( + async def deidentify_text( self, body: _models.DeidentificationContent, *, content_type: str = "application/json", **kwargs: Any ) -> _models.DeidentificationResult: """De-identify text. @@ -1100,47 +663,10 @@ async def deidentify( :return: DeidentificationResult. The DeidentificationResult is compatible with MutableMapping :rtype: ~azure.health.deidentification.models.DeidentificationResult :raises ~azure.core.exceptions.HttpResponseError: - - Example: - .. code-block:: python - - # JSON input template you can fill out and use as your body input. - body = { - "inputText": "str", - "dataType": "str", - "operation": "str", - "redactionFormat": "str" - } - - # response body for status code(s): 200 - response == { - "outputText": "str", - "taggerResult": { - "entities": [ - { - "category": "str", - "length": { - "codePoint": 0, - "utf16": 0, - "utf8": 0 - }, - "offset": { - "codePoint": 0, - "utf16": 0, - "utf8": 0 - }, - "confidenceScore": 0.0, - "text": "str" - } - ], - "etag": "str", - "path": "str" - } - } """ @overload - async def deidentify( + async def deidentify_text( self, body: JSON, *, content_type: str = "application/json", **kwargs: Any ) -> _models.DeidentificationResult: """De-identify text. @@ -1155,39 +681,10 @@ async def deidentify( :return: DeidentificationResult. The DeidentificationResult is compatible with MutableMapping :rtype: ~azure.health.deidentification.models.DeidentificationResult :raises ~azure.core.exceptions.HttpResponseError: - - Example: - .. code-block:: python - - # response body for status code(s): 200 - response == { - "outputText": "str", - "taggerResult": { - "entities": [ - { - "category": "str", - "length": { - "codePoint": 0, - "utf16": 0, - "utf8": 0 - }, - "offset": { - "codePoint": 0, - "utf16": 0, - "utf8": 0 - }, - "confidenceScore": 0.0, - "text": "str" - } - ], - "etag": "str", - "path": "str" - } - } """ @overload - async def deidentify( + async def deidentify_text( self, body: IO[bytes], *, content_type: str = "application/json", **kwargs: Any ) -> _models.DeidentificationResult: """De-identify text. @@ -1202,39 +699,10 @@ async def deidentify( :return: DeidentificationResult. The DeidentificationResult is compatible with MutableMapping :rtype: ~azure.health.deidentification.models.DeidentificationResult :raises ~azure.core.exceptions.HttpResponseError: - - Example: - .. code-block:: python - - # response body for status code(s): 200 - response == { - "outputText": "str", - "taggerResult": { - "entities": [ - { - "category": "str", - "length": { - "codePoint": 0, - "utf16": 0, - "utf8": 0 - }, - "offset": { - "codePoint": 0, - "utf16": 0, - "utf8": 0 - }, - "confidenceScore": 0.0, - "text": "str" - } - ], - "etag": "str", - "path": "str" - } - } """ @distributed_trace_async - async def deidentify( + async def deidentify_text( self, body: Union[_models.DeidentificationContent, JSON, IO[bytes]], **kwargs: Any ) -> _models.DeidentificationResult: """De-identify text. @@ -1247,45 +715,8 @@ async def deidentify( :return: DeidentificationResult. The DeidentificationResult is compatible with MutableMapping :rtype: ~azure.health.deidentification.models.DeidentificationResult :raises ~azure.core.exceptions.HttpResponseError: - - Example: - .. code-block:: python - - # JSON input template you can fill out and use as your body input. - body = { - "inputText": "str", - "dataType": "str", - "operation": "str", - "redactionFormat": "str" - } - - # response body for status code(s): 200 - response == { - "outputText": "str", - "taggerResult": { - "entities": [ - { - "category": "str", - "length": { - "codePoint": 0, - "utf16": 0, - "utf8": 0 - }, - "offset": { - "codePoint": 0, - "utf16": 0, - "utf8": 0 - }, - "confidenceScore": 0.0, - "text": "str" - } - ], - "etag": "str", - "path": "str" - } - } """ - error_map: MutableMapping[int, Type[HttpResponseError]] = { + error_map: MutableMapping = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError, @@ -1306,7 +737,7 @@ async def deidentify( else: _content = json.dumps(body, cls=SdkJSONEncoder, exclude_readonly=True) # type: ignore - _request = build_deidentification_deidentify_request( + _request = build_deidentification_deidentify_text_request( content_type=content_type, api_version=self._config.api_version, content=_content, @@ -1314,7 +745,7 @@ async def deidentify( params=_params, ) path_format_arguments = { - "endpoint": self._serialize.url("self._config.endpoint", self._config.endpoint, "str"), + "endpoint": self._serialize.url("self._config.endpoint", self._config.endpoint, "str", skip_quote=True), } _request.url = self._client.format_url(_request.url, **path_format_arguments) @@ -1327,16 +758,24 @@ async def deidentify( if response.status_code not in [200]: if _stream: - await response.read() # Load the body in memory and close the socket + try: + await response.read() # Load the body in memory and close the socket + except (StreamConsumedError, StreamClosedError): + pass map_error(status_code=response.status_code, response=response, error_map=error_map) raise HttpResponseError(response=response) + response_headers = {} + response_headers["x-ms-client-request-id"] = self._deserialize( + "str", response.headers.get("x-ms-client-request-id") + ) + if _stream: deserialized = response.iter_bytes() else: deserialized = _deserialize(_models.DeidentificationResult, response.json()) if cls: - return cls(pipeline_response, deserialized, {}) # type: ignore + return cls(pipeline_response, deserialized, response_headers) # type: ignore return deserialized # type: ignore diff --git a/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/aio/_patch.py b/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/aio/_patch.py index f7dd32510333..62b77226c9f1 100644 --- a/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/aio/_patch.py +++ b/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/aio/_patch.py @@ -6,9 +6,42 @@ Follow our quickstart for examples: https://aka.ms/azsdk/python/dpcodegen/python/customize """ -from typing import List +from typing import Any, AsyncIterable, List -__all__: List[str] = [] # Add all objects you want publicly available to users at this package level +from azure.core.tracing.decorator import distributed_trace +from azure.health.deidentification.models import DeidentificationDocumentDetails, DeidentificationJob +from ._client import DeidentificationClient as DeidentificationClientGenerated + +__all__: List[str] = [ + "DeidentificationClient", +] # Add all objects you want publicly available to users at this package level + + +class DeidentificationClient(DeidentificationClientGenerated): + + @distributed_trace + def list_jobs(self, **kwargs: Any) -> AsyncIterable[DeidentificationJob]: + """ + List de-identification jobs. + + :return: An iterator like instance of DeidentificationJob + :rtype: ~azure.core.paging.ItemPaged[~azure.health.deidentification.models.DeidentificationJob] + :raises ~azure.core.exceptions.HttpResponseError: + """ + return super()._list_jobs_internal(continuation_token_parameter=None, **kwargs) + + @distributed_trace + def list_job_documents(self, job_name: str, **kwargs: Any) -> AsyncIterable[DeidentificationDocumentDetails]: + """ + List processed documents within a job. + + :param job_name: The name of a job. Required. + :type job_name: str + :return: An iterator like instance of DocumentDetails + :rtype: ~azure.core.paging.ItemPaged[~azure.health.deidentification.models.DocumentDetails] + :raises ~azure.core.exceptions.HttpResponseError: + """ + return super()._list_job_documents_internal(job_name=job_name, continuation_token_parameter=None, **kwargs) def patch_sdk(): diff --git a/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/aio/_vendor.py b/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/aio/_vendor.py index 39bc7460b3a7..0afc83d417cc 100644 --- a/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/aio/_vendor.py +++ b/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/aio/_vendor.py @@ -11,7 +11,6 @@ from ._configuration import DeidentificationClientConfiguration if TYPE_CHECKING: - # pylint: disable=unused-import,ungrouped-imports from azure.core import AsyncPipelineClient from .._serialization import Deserializer, Serializer diff --git a/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/models/__init__.py b/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/models/__init__.py index 2bbbe6e08cab..5b5bfa8d86a0 100644 --- a/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/models/__init__.py +++ b/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/models/__init__.py @@ -5,49 +5,58 @@ # Code generated by Microsoft (R) Python Code Generator. # Changes may cause incorrect behavior and will be lost if the code is regenerated. # -------------------------------------------------------------------------- +# pylint: disable=wrong-import-position -from ._models import DeidentificationContent -from ._models import DeidentificationJob -from ._models import DeidentificationResult -from ._models import DocumentDetails -from ._models import DocumentLocation -from ._models import Error -from ._models import InnerError -from ._models import JobSummary -from ._models import PhiEntity -from ._models import PhiTaggerResult -from ._models import SourceStorageLocation -from ._models import StringIndex -from ._models import TargetStorageLocation +from typing import TYPE_CHECKING -from ._enums import DocumentDataType -from ._enums import JobStatus -from ._enums import OperationState -from ._enums import OperationType -from ._enums import PhiCategory +if TYPE_CHECKING: + from ._patch import * # pylint: disable=unused-wildcard-import + + +from ._models import ( # type: ignore + DeidentificationContent, + DeidentificationCustomizationOptions, + DeidentificationDocumentDetails, + DeidentificationDocumentLocation, + DeidentificationJob, + DeidentificationJobCustomizationOptions, + DeidentificationJobSummary, + DeidentificationResult, + PhiEntity, + PhiTaggerResult, + SourceStorageLocation, + StringIndex, + TargetStorageLocation, +) + +from ._enums import ( # type: ignore + DeidentificationJobStatus, + DeidentificationOperationType, + OperationState, + PhiCategory, +) from ._patch import __all__ as _patch_all -from ._patch import * # pylint: disable=unused-wildcard-import +from ._patch import * from ._patch import patch_sdk as _patch_sdk __all__ = [ "DeidentificationContent", + "DeidentificationCustomizationOptions", + "DeidentificationDocumentDetails", + "DeidentificationDocumentLocation", "DeidentificationJob", + "DeidentificationJobCustomizationOptions", + "DeidentificationJobSummary", "DeidentificationResult", - "DocumentDetails", - "DocumentLocation", - "Error", - "InnerError", - "JobSummary", "PhiEntity", "PhiTaggerResult", "SourceStorageLocation", "StringIndex", "TargetStorageLocation", - "DocumentDataType", - "JobStatus", + "DeidentificationJobStatus", + "DeidentificationOperationType", "OperationState", - "OperationType", "PhiCategory", ] -__all__.extend([p for p in _patch_all if p not in __all__]) +__all__.extend([p for p in _patch_all if p not in __all__]) # pyright: ignore _patch_sdk() diff --git a/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/models/_enums.py b/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/models/_enums.py index c05e0003b5b7..9584b79e1578 100644 --- a/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/models/_enums.py +++ b/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/models/_enums.py @@ -10,14 +10,7 @@ from azure.core import CaseInsensitiveEnumMeta -class DocumentDataType(str, Enum, metaclass=CaseInsensitiveEnumMeta): - """Enum of supported Data Types.""" - - PLAINTEXT = "Plaintext" - """Plain text data type.""" - - -class JobStatus(str, Enum, metaclass=CaseInsensitiveEnumMeta): +class DeidentificationJobStatus(str, Enum, metaclass=CaseInsensitiveEnumMeta): """List of statuses a job can have.""" NOT_STARTED = "NotStarted" @@ -34,6 +27,18 @@ class JobStatus(str, Enum, metaclass=CaseInsensitiveEnumMeta): """Job has been canceled after user request.""" +class DeidentificationOperationType(str, Enum, metaclass=CaseInsensitiveEnumMeta): + """Enum of supported Operation Types.""" + + REDACT = "Redact" + """Redact Operation will remove all entities of PHI and replace them with a placeholder value.""" + SURROGATE = "Surrogate" + """Surrogation Operation will replace all entities of PHI with a surrogate value.""" + TAG = "Tag" + """Tag Operation will detect all entities of PHI, their type, and return their locations in the + document.""" + + class OperationState(str, Enum, metaclass=CaseInsensitiveEnumMeta): """Enum describing allowed operation states.""" @@ -49,18 +54,6 @@ class OperationState(str, Enum, metaclass=CaseInsensitiveEnumMeta): """The operation has been canceled by the user.""" -class OperationType(str, Enum, metaclass=CaseInsensitiveEnumMeta): - """Enum of supported Operation Types.""" - - REDACT = "Redact" - """Redact Operation will remove all entities of PHI and replace them with a placeholder value.""" - SURROGATE = "Surrogate" - """Surrogation Operation will replace all entities of PHI with a surrogate value.""" - TAG = "Tag" - """Tag Operation will detect all entities of PHI, their type, and return their locations in the - document.""" - - class PhiCategory(str, Enum, metaclass=CaseInsensitiveEnumMeta): """List of PHI Entities.""" @@ -70,7 +63,7 @@ class PhiCategory(str, Enum, metaclass=CaseInsensitiveEnumMeta): """Account Number.""" AGE = "Age" """Age.""" - BIO_I_D = "BioID" + BIO_ID = "BioID" """Biological Identifier, such as a fingerprint or retinal scan.""" CITY = "City" """City.""" @@ -90,9 +83,9 @@ class PhiCategory(str, Enum, metaclass=CaseInsensitiveEnumMeta): """Health Plan ID Numbers.""" HOSPITAL = "Hospital" """Hospital Name.""" - I_D_NUM = "IDNum" + ID_NUM = "IDNum" """Id Number, eg. passport number.""" - I_P_ADDRESS = "IPAddress" + IP_ADDRESS = "IPAddress" """IP Address.""" LICENSE = "License" """License, eg. Driver's license or medical license.""" diff --git a/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/models/_models.py b/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/models/_models.py index 2929c7e9b5d3..9c220eb3bc66 100644 --- a/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/models/_models.py +++ b/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/models/_models.py @@ -1,20 +1,21 @@ # coding=utf-8 -# pylint: disable=too-many-lines # -------------------------------------------------------------------------- # Copyright (c) Microsoft Corporation. All rights reserved. # Licensed under the MIT License. See License.txt in the project root for license information. # Code generated by Microsoft (R) Python Code Generator. # Changes may cause incorrect behavior and will be lost if the code is regenerated. # -------------------------------------------------------------------------- +# pylint: disable=useless-super-delegation import datetime from typing import Any, List, Mapping, Optional, TYPE_CHECKING, Union, overload +from azure.core.exceptions import ODataV4Format + from .. import _model_base from .._model_base import rest_field if TYPE_CHECKING: - # pylint: disable=unused-import,ungrouped-imports from .. import models as _models @@ -25,172 +26,76 @@ class DeidentificationContent(_model_base.Model): :ivar input_text: Input text to de-identify. Required. :vartype input_text: str - :ivar operation: Operation to perform on the input. Known values are: "Redact", "Surrogate", - and "Tag". - :vartype operation: str or ~azure.health.deidentification.models.OperationType - :ivar data_type: Data type of the input. "Plaintext" - :vartype data_type: str or ~azure.health.deidentification.models.DocumentDataType - :ivar redaction_format: Format of the redacted output. Only valid when OperationType is - "Redact". - :vartype redaction_format: str + :ivar operation: Operation to perform on the input documents. Known values are: "Redact", + "Surrogate", and "Tag". + :vartype operation: str or ~azure.health.deidentification.models.DeidentificationOperationType + :ivar customizations: Customization parameters to override default service behaviors. + :vartype customizations: + ~azure.health.deidentification.models.DeidentificationCustomizationOptions """ input_text: str = rest_field(name="inputText") """Input text to de-identify. Required.""" - operation: Optional[Union[str, "_models.OperationType"]] = rest_field() - """Operation to perform on the input. Known values are: \"Redact\", \"Surrogate\", and \"Tag\".""" - data_type: Optional[Union[str, "_models.DocumentDataType"]] = rest_field(name="dataType") - """Data type of the input. \"Plaintext\"""" - redaction_format: Optional[str] = rest_field(name="redactionFormat") - """Format of the redacted output. Only valid when OperationType is \"Redact\".""" + operation: Optional[Union[str, "_models.DeidentificationOperationType"]] = rest_field() + """Operation to perform on the input documents. Known values are: \"Redact\", \"Surrogate\", and + \"Tag\".""" + customizations: Optional["_models.DeidentificationCustomizationOptions"] = rest_field() + """Customization parameters to override default service behaviors.""" @overload def __init__( self, *, input_text: str, - operation: Optional[Union[str, "_models.OperationType"]] = None, - data_type: Optional[Union[str, "_models.DocumentDataType"]] = None, - redaction_format: Optional[str] = None, - ): ... + operation: Optional[Union[str, "_models.DeidentificationOperationType"]] = None, + customizations: Optional["_models.DeidentificationCustomizationOptions"] = None, + ) -> None: ... @overload - def __init__(self, mapping: Mapping[str, Any]): + def __init__(self, mapping: Mapping[str, Any]) -> None: """ :param mapping: raw JSON to initialize the model. :type mapping: Mapping[str, Any] """ - def __init__(self, *args: Any, **kwargs: Any) -> None: # pylint: disable=useless-super-delegation + def __init__(self, *args: Any, **kwargs: Any) -> None: super().__init__(*args, **kwargs) -class DeidentificationJob(_model_base.Model): # pylint: disable=too-many-instance-attributes - """A job containing a batch of documents to de-identify. - - Readonly variables are only populated by the server, and will be ignored when sending a request. - - All required parameters must be populated in order to send to server. +class DeidentificationCustomizationOptions(_model_base.Model): + """Customizations options to override default service behaviors for synchronous usage. - :ivar name: The name of a job. Required. - :vartype name: str - :ivar source_location: Storage location to perform the operation on. Required. - :vartype source_location: ~azure.health.deidentification.models.SourceStorageLocation - :ivar target_location: Target location to store output of operation. Required. - :vartype target_location: ~azure.health.deidentification.models.TargetStorageLocation - :ivar operation: Operation to perform on the input documents. Known values are: "Redact", - "Surrogate", and "Tag". - :vartype operation: str or ~azure.health.deidentification.models.OperationType - :ivar data_type: Data type of the input documents. "Plaintext" - :vartype data_type: str or ~azure.health.deidentification.models.DocumentDataType :ivar redaction_format: Format of the redacted output. Only valid when Operation is Redact. :vartype redaction_format: str - :ivar status: Current status of a job. Required. Known values are: "NotStarted", "Running", - "Succeeded", "PartialFailed", "Failed", and "Canceled". - :vartype status: str or ~azure.health.deidentification.models.JobStatus - :ivar error: Error when job fails in it's entirety. - :vartype error: ~azure.health.deidentification.models.Error - :ivar last_updated_at: Date and time when the job was completed. - - If the job is canceled, this is the time when the job was canceled. - - If the job failed, this is the time when the job failed. Required. - :vartype last_updated_at: ~datetime.datetime - :ivar created_at: Date and time when the job was created. Required. - :vartype created_at: ~datetime.datetime - :ivar started_at: Date and time when the job was started. - :vartype started_at: ~datetime.datetime - :ivar summary: Summary of a job. Exists only when the job is completed. - :vartype summary: ~azure.health.deidentification.models.JobSummary + :ivar surrogate_locale: Locale in which the output surrogates are written. + :vartype surrogate_locale: str """ - name: str = rest_field(visibility=["read"]) - """The name of a job. Required.""" - source_location: "_models.SourceStorageLocation" = rest_field(name="sourceLocation") - """Storage location to perform the operation on. Required.""" - target_location: "_models.TargetStorageLocation" = rest_field(name="targetLocation") - """Target location to store output of operation. Required.""" - operation: Optional[Union[str, "_models.OperationType"]] = rest_field() - """Operation to perform on the input documents. Known values are: \"Redact\", \"Surrogate\", and - \"Tag\".""" - data_type: Optional[Union[str, "_models.DocumentDataType"]] = rest_field(name="dataType") - """Data type of the input documents. \"Plaintext\"""" redaction_format: Optional[str] = rest_field(name="redactionFormat") """Format of the redacted output. Only valid when Operation is Redact.""" - status: Union[str, "_models.JobStatus"] = rest_field(visibility=["read"]) - """Current status of a job. Required. Known values are: \"NotStarted\", \"Running\", - \"Succeeded\", \"PartialFailed\", \"Failed\", and \"Canceled\".""" - error: Optional["_models.Error"] = rest_field(visibility=["read"]) - """Error when job fails in it's entirety.""" - last_updated_at: datetime.datetime = rest_field(name="lastUpdatedAt", visibility=["read"], format="rfc3339") - """Date and time when the job was completed. - - If the job is canceled, this is the time when the job was canceled. - - If the job failed, this is the time when the job failed. Required.""" - created_at: datetime.datetime = rest_field(name="createdAt", visibility=["read"], format="rfc3339") - """Date and time when the job was created. Required.""" - started_at: Optional[datetime.datetime] = rest_field(name="startedAt", visibility=["read"], format="rfc3339") - """Date and time when the job was started.""" - summary: Optional["_models.JobSummary"] = rest_field(visibility=["read"]) - """Summary of a job. Exists only when the job is completed.""" + surrogate_locale: Optional[str] = rest_field(name="surrogateLocale") + """Locale in which the output surrogates are written.""" @overload def __init__( self, *, - source_location: "_models.SourceStorageLocation", - target_location: "_models.TargetStorageLocation", - operation: Optional[Union[str, "_models.OperationType"]] = None, - data_type: Optional[Union[str, "_models.DocumentDataType"]] = None, redaction_format: Optional[str] = None, - ): ... - - @overload - def __init__(self, mapping: Mapping[str, Any]): - """ - :param mapping: raw JSON to initialize the model. - :type mapping: Mapping[str, Any] - """ - - def __init__(self, *args: Any, **kwargs: Any) -> None: # pylint: disable=useless-super-delegation - super().__init__(*args, **kwargs) - - -class DeidentificationResult(_model_base.Model): - """Response body for de-identification operation. - - :ivar output_text: Output text after de-identification. Not available for "Tag" operation. - :vartype output_text: str - :ivar tagger_result: Result of the "Tag" operation. Only available for "Tag" Operation. - :vartype tagger_result: ~azure.health.deidentification.models.PhiTaggerResult - """ - - output_text: Optional[str] = rest_field(name="outputText") - """Output text after de-identification. Not available for \"Tag\" operation.""" - tagger_result: Optional["_models.PhiTaggerResult"] = rest_field(name="taggerResult") - """Result of the \"Tag\" operation. Only available for \"Tag\" Operation.""" + surrogate_locale: Optional[str] = None, + ) -> None: ... @overload - def __init__( - self, - *, - output_text: Optional[str] = None, - tagger_result: Optional["_models.PhiTaggerResult"] = None, - ): ... - - @overload - def __init__(self, mapping: Mapping[str, Any]): + def __init__(self, mapping: Mapping[str, Any]) -> None: """ :param mapping: raw JSON to initialize the model. :type mapping: Mapping[str, Any] """ - def __init__(self, *args: Any, **kwargs: Any) -> None: # pylint: disable=useless-super-delegation + def __init__(self, *args: Any, **kwargs: Any) -> None: super().__init__(*args, **kwargs) -class DocumentDetails(_model_base.Model): +class DeidentificationDocumentDetails(_model_base.Model): """Details of a single document in a job. Readonly variables are only populated by the server, and will be ignored when sending a request. @@ -199,63 +104,63 @@ class DocumentDetails(_model_base.Model): :ivar id: Id of the document details. Required. :vartype id: str :ivar input: Location for the input. Required. - :vartype input: ~azure.health.deidentification.models.DocumentLocation + :vartype input: ~azure.health.deidentification.models.DeidentificationDocumentLocation :ivar output: Location for the output. - :vartype output: ~azure.health.deidentification.models.DocumentLocation + :vartype output: ~azure.health.deidentification.models.DeidentificationDocumentLocation :ivar status: Status of the document. Required. Known values are: "NotStarted", "Running", "Succeeded", "Failed", and "Canceled". :vartype status: str or ~azure.health.deidentification.models.OperationState :ivar error: Error when document fails. - :vartype error: ~azure.health.deidentification.models.Error + :vartype error: ~azure.core.ODataV4Format """ id: str = rest_field(visibility=["read"]) """Id of the document details. Required.""" - input: "_models.DocumentLocation" = rest_field() + input: "_models.DeidentificationDocumentLocation" = rest_field() """Location for the input. Required.""" - output: Optional["_models.DocumentLocation"] = rest_field() + output: Optional["_models.DeidentificationDocumentLocation"] = rest_field() """Location for the output.""" status: Union[str, "_models.OperationState"] = rest_field() """Status of the document. Required. Known values are: \"NotStarted\", \"Running\", \"Succeeded\", \"Failed\", and \"Canceled\".""" - error: Optional["_models.Error"] = rest_field() + error: Optional[ODataV4Format] = rest_field() """Error when document fails.""" @overload def __init__( self, *, - input: "_models.DocumentLocation", + input: "_models.DeidentificationDocumentLocation", status: Union[str, "_models.OperationState"], - output: Optional["_models.DocumentLocation"] = None, - error: Optional["_models.Error"] = None, - ): ... + output: Optional["_models.DeidentificationDocumentLocation"] = None, + error: Optional[ODataV4Format] = None, + ) -> None: ... @overload - def __init__(self, mapping: Mapping[str, Any]): + def __init__(self, mapping: Mapping[str, Any]) -> None: """ :param mapping: raw JSON to initialize the model. :type mapping: Mapping[str, Any] """ - def __init__(self, *args: Any, **kwargs: Any) -> None: # pylint: disable=useless-super-delegation + def __init__(self, *args: Any, **kwargs: Any) -> None: super().__init__(*args, **kwargs) -class DocumentLocation(_model_base.Model): +class DeidentificationDocumentLocation(_model_base.Model): """Location of a document. Readonly variables are only populated by the server, and will be ignored when sending a request. - :ivar path: Path of document in storage. Required. - :vartype path: str + :ivar location: Location of document in storage. Required. + :vartype location: str :ivar etag: The entity tag for this resource. Required. :vartype etag: str """ - path: str = rest_field() - """Path of document in storage. Required.""" + location: str = rest_field() + """Location of document in storage. Required.""" etag: str = rest_field(visibility=["read"]) """The entity tag for this resource. Required.""" @@ -263,107 +168,141 @@ class DocumentLocation(_model_base.Model): def __init__( self, *, - path: str, - ): ... + location: str, + ) -> None: ... @overload - def __init__(self, mapping: Mapping[str, Any]): + def __init__(self, mapping: Mapping[str, Any]) -> None: """ :param mapping: raw JSON to initialize the model. :type mapping: Mapping[str, Any] """ - def __init__(self, *args: Any, **kwargs: Any) -> None: # pylint: disable=useless-super-delegation + def __init__(self, *args: Any, **kwargs: Any) -> None: super().__init__(*args, **kwargs) -class Error(_model_base.Model): - """The error object. +class DeidentificationJob(_model_base.Model): + """A job containing a batch of documents to de-identify. + + Readonly variables are only populated by the server, and will be ignored when sending a request. - All required parameters must be populated in order to send to server. - :ivar code: One of a server-defined set of error codes. Required. - :vartype code: str - :ivar message: A human-readable representation of the error. Required. - :vartype message: str - :ivar target: The target of the error. - :vartype target: str - :ivar details: An array of details about specific errors that led to this reported error. - :vartype details: list[~azure.health.deidentification.models.Error] - :ivar innererror: An object containing more specific information than the current object about - the error. - :vartype innererror: ~azure.health.deidentification.models.InnerError + :ivar name: The name of a job. Required. + :vartype name: str + :ivar operation: Operation to perform on the input documents. Known values are: "Redact", + "Surrogate", and "Tag". + :vartype operation: str or ~azure.health.deidentification.models.DeidentificationOperationType + :ivar source_location: Storage location to perform the operation on. Required. + :vartype source_location: ~azure.health.deidentification.models.SourceStorageLocation + :ivar target_location: Target location to store output of operation. Required. + :vartype target_location: ~azure.health.deidentification.models.TargetStorageLocation + :ivar customizations: Customization parameters to override default service behaviors. + :vartype customizations: + ~azure.health.deidentification.models.DeidentificationJobCustomizationOptions + :ivar status: Current status of a job. Required. Known values are: "NotStarted", "Running", + "Succeeded", "PartialFailed", "Failed", and "Canceled". + :vartype status: str or ~azure.health.deidentification.models.DeidentificationJobStatus + :ivar error: Error when job fails in it's entirety. + :vartype error: ~azure.core.ODataV4Format + :ivar last_updated_at: Date and time when the job was completed. + + If the job is canceled, this is the time when the job was canceled. + + If the job failed, this is the time when the job failed. Required. + :vartype last_updated_at: ~datetime.datetime + :ivar created_at: Date and time when the job was created. Required. + :vartype created_at: ~datetime.datetime + :ivar started_at: Date and time when the job was started. + :vartype started_at: ~datetime.datetime + :ivar summary: Summary of a job. Exists only when the job is completed. + :vartype summary: ~azure.health.deidentification.models.DeidentificationJobSummary """ - code: str = rest_field() - """One of a server-defined set of error codes. Required.""" - message: str = rest_field() - """A human-readable representation of the error. Required.""" - target: Optional[str] = rest_field() - """The target of the error.""" - details: Optional[List["_models.Error"]] = rest_field() - """An array of details about specific errors that led to this reported error.""" - innererror: Optional["_models.InnerError"] = rest_field() - """An object containing more specific information than the current object about the error.""" + name: str = rest_field(visibility=["read"]) + """The name of a job. Required.""" + operation: Optional[Union[str, "_models.DeidentificationOperationType"]] = rest_field() + """Operation to perform on the input documents. Known values are: \"Redact\", \"Surrogate\", and + \"Tag\".""" + source_location: "_models.SourceStorageLocation" = rest_field(name="sourceLocation") + """Storage location to perform the operation on. Required.""" + target_location: "_models.TargetStorageLocation" = rest_field(name="targetLocation") + """Target location to store output of operation. Required.""" + customizations: Optional["_models.DeidentificationJobCustomizationOptions"] = rest_field() + """Customization parameters to override default service behaviors.""" + status: Union[str, "_models.DeidentificationJobStatus"] = rest_field(visibility=["read"]) + """Current status of a job. Required. Known values are: \"NotStarted\", \"Running\", + \"Succeeded\", \"PartialFailed\", \"Failed\", and \"Canceled\".""" + error: Optional[ODataV4Format] = rest_field(visibility=["read"]) + """Error when job fails in it's entirety.""" + last_updated_at: datetime.datetime = rest_field(name="lastUpdatedAt", visibility=["read"], format="rfc3339") + """Date and time when the job was completed. + + If the job is canceled, this is the time when the job was canceled. + + If the job failed, this is the time when the job failed. Required.""" + created_at: datetime.datetime = rest_field(name="createdAt", visibility=["read"], format="rfc3339") + """Date and time when the job was created. Required.""" + started_at: Optional[datetime.datetime] = rest_field(name="startedAt", visibility=["read"], format="rfc3339") + """Date and time when the job was started.""" + summary: Optional["_models.DeidentificationJobSummary"] = rest_field(visibility=["read"]) + """Summary of a job. Exists only when the job is completed.""" @overload def __init__( self, *, - code: str, - message: str, - target: Optional[str] = None, - details: Optional[List["_models.Error"]] = None, - innererror: Optional["_models.InnerError"] = None, - ): ... + source_location: "_models.SourceStorageLocation", + target_location: "_models.TargetStorageLocation", + operation: Optional[Union[str, "_models.DeidentificationOperationType"]] = None, + customizations: Optional["_models.DeidentificationJobCustomizationOptions"] = None, + ) -> None: ... @overload - def __init__(self, mapping: Mapping[str, Any]): + def __init__(self, mapping: Mapping[str, Any]) -> None: """ :param mapping: raw JSON to initialize the model. :type mapping: Mapping[str, Any] """ - def __init__(self, *args: Any, **kwargs: Any) -> None: # pylint: disable=useless-super-delegation + def __init__(self, *args: Any, **kwargs: Any) -> None: super().__init__(*args, **kwargs) -class InnerError(_model_base.Model): - """An object containing more specific information about the error. As per Microsoft One API - guidelines - - https://github.com/Microsoft/api-guidelines/blob/vNext/Guidelines.md#7102-error-condition-responses. +class DeidentificationJobCustomizationOptions(_model_base.Model): + """Customizations options to override default service behaviors for job usage. - :ivar code: One of a server-defined set of error codes. - :vartype code: str - :ivar innererror: Inner error. - :vartype innererror: ~azure.health.deidentification.models.InnerError + :ivar redaction_format: Format of the redacted output. Only valid when Operation is Redact. + :vartype redaction_format: str + :ivar surrogate_locale: Locale in which the output surrogates are written. + :vartype surrogate_locale: str """ - code: Optional[str] = rest_field() - """One of a server-defined set of error codes.""" - innererror: Optional["_models.InnerError"] = rest_field() - """Inner error.""" + redaction_format: Optional[str] = rest_field(name="redactionFormat") + """Format of the redacted output. Only valid when Operation is Redact.""" + surrogate_locale: Optional[str] = rest_field(name="surrogateLocale") + """Locale in which the output surrogates are written.""" @overload def __init__( self, *, - code: Optional[str] = None, - innererror: Optional["_models.InnerError"] = None, - ): ... + redaction_format: Optional[str] = None, + surrogate_locale: Optional[str] = None, + ) -> None: ... @overload - def __init__(self, mapping: Mapping[str, Any]): + def __init__(self, mapping: Mapping[str, Any]) -> None: """ :param mapping: raw JSON to initialize the model. :type mapping: Mapping[str, Any] """ - def __init__(self, *args: Any, **kwargs: Any) -> None: # pylint: disable=useless-super-delegation + def __init__(self, *args: Any, **kwargs: Any) -> None: super().__init__(*args, **kwargs) -class JobSummary(_model_base.Model): +class DeidentificationJobSummary(_model_base.Model): """Summary metrics of a job. @@ -399,16 +338,49 @@ def __init__( canceled: int, total: int, bytes_processed: int, - ): ... + ) -> None: ... + + @overload + def __init__(self, mapping: Mapping[str, Any]) -> None: + """ + :param mapping: raw JSON to initialize the model. + :type mapping: Mapping[str, Any] + """ + + def __init__(self, *args: Any, **kwargs: Any) -> None: + super().__init__(*args, **kwargs) + + +class DeidentificationResult(_model_base.Model): + """Response body for de-identification operation. + + :ivar output_text: Output text after de-identification. Not available for "Tag" operation. + :vartype output_text: str + :ivar tagger_result: Result of the "Tag" operation. Only available for "Tag" Operation. + :vartype tagger_result: ~azure.health.deidentification.models.PhiTaggerResult + """ + + output_text: Optional[str] = rest_field(name="outputText") + """Output text after de-identification. Not available for \"Tag\" operation.""" + tagger_result: Optional["_models.PhiTaggerResult"] = rest_field(name="taggerResult") + """Result of the \"Tag\" operation. Only available for \"Tag\" Operation.""" @overload - def __init__(self, mapping: Mapping[str, Any]): + def __init__( + self, + *, + output_text: Optional[str] = None, + tagger_result: Optional["_models.PhiTaggerResult"] = None, + ) -> None: ... + + @overload + def __init__(self, mapping: Mapping[str, Any]) -> None: """ :param mapping: raw JSON to initialize the model. :type mapping: Mapping[str, Any] """ - def __init__(self, *args: Any, **kwargs: Any) -> None: # pylint: disable=useless-super-delegation + def __init__(self, *args: Any, **kwargs: Any) -> None: super().__init__(*args, **kwargs) @@ -456,16 +428,16 @@ def __init__( length: "_models.StringIndex", text: Optional[str] = None, confidence_score: Optional[float] = None, - ): ... + ) -> None: ... @overload - def __init__(self, mapping: Mapping[str, Any]): + def __init__(self, mapping: Mapping[str, Any]) -> None: """ :param mapping: raw JSON to initialize the model. :type mapping: Mapping[str, Any] """ - def __init__(self, *args: Any, **kwargs: Any) -> None: # pylint: disable=useless-super-delegation + def __init__(self, *args: Any, **kwargs: Any) -> None: super().__init__(*args, **kwargs) @@ -475,43 +447,32 @@ class PhiTaggerResult(_model_base.Model): :ivar entities: List of entities detected in the input. Required. :vartype entities: list[~azure.health.deidentification.models.PhiEntity] - :ivar path: Path to the document in storage. - :vartype path: str - :ivar etag: The entity tag for this resource. - :vartype etag: str """ entities: List["_models.PhiEntity"] = rest_field() """List of entities detected in the input. Required.""" - path: Optional[str] = rest_field() - """Path to the document in storage.""" - etag: Optional[str] = rest_field() - """The entity tag for this resource.""" @overload def __init__( self, *, entities: List["_models.PhiEntity"], - path: Optional[str] = None, - etag: Optional[str] = None, - ): ... + ) -> None: ... @overload - def __init__(self, mapping: Mapping[str, Any]): + def __init__(self, mapping: Mapping[str, Any]) -> None: """ :param mapping: raw JSON to initialize the model. :type mapping: Mapping[str, Any] """ - def __init__(self, *args: Any, **kwargs: Any) -> None: # pylint: disable=useless-super-delegation + def __init__(self, *args: Any, **kwargs: Any) -> None: super().__init__(*args, **kwargs) class SourceStorageLocation(_model_base.Model): """Storage location. - All required parameters must be populated in order to send to server. :ivar location: URL to storage location. Required. :vartype location: str @@ -535,16 +496,16 @@ def __init__( location: str, prefix: str, extensions: Optional[List[str]] = None, - ): ... + ) -> None: ... @overload - def __init__(self, mapping: Mapping[str, Any]): + def __init__(self, mapping: Mapping[str, Any]) -> None: """ :param mapping: raw JSON to initialize the model. :type mapping: Mapping[str, Any] """ - def __init__(self, *args: Any, **kwargs: Any) -> None: # pylint: disable=useless-super-delegation + def __init__(self, *args: Any, **kwargs: Any) -> None: super().__init__(*args, **kwargs) @@ -582,34 +543,55 @@ def __init__( utf8: int, utf16: int, code_point: int, - ): ... + ) -> None: ... @overload - def __init__(self, mapping: Mapping[str, Any]): + def __init__(self, mapping: Mapping[str, Any]) -> None: """ :param mapping: raw JSON to initialize the model. :type mapping: Mapping[str, Any] """ - def __init__(self, *args: Any, **kwargs: Any) -> None: # pylint: disable=useless-super-delegation + def __init__(self, *args: Any, **kwargs: Any) -> None: super().__init__(*args, **kwargs) class TargetStorageLocation(_model_base.Model): """Storage location. - All required parameters must be populated in order to send to server. :ivar location: URL to storage location. Required. :vartype location: str - :ivar prefix: Prefix to filter path by. Required. + :ivar prefix: Replaces the input prefix of a file path with the output prefix, preserving the + rest of the path structure. + + Example: + File full path: documents/user/note.txt + Input Prefix: "documents/user/" + Output Prefix: "output_docs/" + + Output file: "output_docs/note.txt". Required. :vartype prefix: str + :ivar overwrite: When set to true during a job, the service will overwrite the output location + if it already exists. + :vartype overwrite: bool """ location: str = rest_field() """URL to storage location. Required.""" prefix: str = rest_field() - """Prefix to filter path by. Required.""" + """Replaces the input prefix of a file path with the output prefix, preserving the rest of the + path structure. + + Example: + File full path: documents/user/note.txt + Input Prefix: \"documents/user/\" + Output Prefix: \"output_docs/\" + + Output file: \"output_docs/note.txt\". Required.""" + overwrite: Optional[bool] = rest_field() + """When set to true during a job, the service will overwrite the output location if it already + exists.""" @overload def __init__( @@ -617,14 +599,15 @@ def __init__( *, location: str, prefix: str, - ): ... + overwrite: Optional[bool] = None, + ) -> None: ... @overload - def __init__(self, mapping: Mapping[str, Any]): + def __init__(self, mapping: Mapping[str, Any]) -> None: """ :param mapping: raw JSON to initialize the model. :type mapping: Mapping[str, Any] """ - def __init__(self, *args: Any, **kwargs: Any) -> None: # pylint: disable=useless-super-delegation + def __init__(self, *args: Any, **kwargs: Any) -> None: super().__init__(*args, **kwargs) diff --git a/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/models/_patch.py b/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/models/_patch.py index 807491528291..36869c12d74a 100644 --- a/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/models/_patch.py +++ b/sdk/healthdataaiservices/azure-health-deidentification/azure/health/deidentification/models/_patch.py @@ -9,7 +9,6 @@ from typing import TYPE_CHECKING, List if TYPE_CHECKING: - # pylint: disable=unused-import,ungrouped-imports from .. import models as _models __all__: List[str] = [] # Add all objects you want publicly available to users at this package level diff --git a/sdk/healthdataaiservices/azure-health-deidentification/generated_samples/cancel_job.py b/sdk/healthdataaiservices/azure-health-deidentification/generated_samples/cancel_job.py new file mode 100644 index 000000000000..e99249686e66 --- /dev/null +++ b/sdk/healthdataaiservices/azure-health-deidentification/generated_samples/cancel_job.py @@ -0,0 +1,41 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for license information. +# Code generated by Microsoft (R) Python Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is regenerated. +# -------------------------------------------------------------------------- + +from azure.identity import DefaultAzureCredential + +from azure.health.deidentification import DeidentificationClient + +""" +# PREREQUISITES + pip install azure-identity + pip install azure-health-deidentification +# USAGE + python cancel_job.py + + Before run the sample, please set the values of the client ID, tenant ID and client secret + of the AAD application as environment variables: AZURE_CLIENT_ID, AZURE_TENANT_ID, + AZURE_CLIENT_SECRET. For more info about how to get the value, please see: + https://docs.microsoft.com/azure/active-directory/develop/howto-create-service-principal-portal +""" + + +def main(): + client = DeidentificationClient( + endpoint="ENDPOINT", + credential=DefaultAzureCredential(), + ) + + response = client.cancel_job( + name="job_smith_documents_1", + ) + print(response) + + +# x-ms-original-file: 2024-11-15/CancelJob.json +if __name__ == "__main__": + main() diff --git a/sdk/healthdataaiservices/azure-health-deidentification/generated_samples/deidentify_documents.py b/sdk/healthdataaiservices/azure-health-deidentification/generated_samples/deidentify_documents.py new file mode 100644 index 000000000000..e568f1e74441 --- /dev/null +++ b/sdk/healthdataaiservices/azure-health-deidentification/generated_samples/deidentify_documents.py @@ -0,0 +1,51 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for license information. +# Code generated by Microsoft (R) Python Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is regenerated. +# -------------------------------------------------------------------------- + +from azure.identity import DefaultAzureCredential + +from azure.health.deidentification import DeidentificationClient + +""" +# PREREQUISITES + pip install azure-identity + pip install azure-health-deidentification +# USAGE + python deidentify_documents.py + + Before run the sample, please set the values of the client ID, tenant ID and client secret + of the AAD application as environment variables: AZURE_CLIENT_ID, AZURE_TENANT_ID, + AZURE_CLIENT_SECRET. For more info about how to get the value, please see: + https://docs.microsoft.com/azure/active-directory/develop/howto-create-service-principal-portal +""" + + +def main(): + client = DeidentificationClient( + endpoint="ENDPOINT", + credential=DefaultAzureCredential(), + ) + + response = client.begin_deidentify_documents( + name="job_smith_documents_1", + resource={ + "customizations": {"redactionFormat": "[{type}]"}, + "operation": "Redact", + "sourceLocation": {"location": "https://blobtest.blob.core.windows.net/container", "prefix": "documents/"}, + "targetLocation": { + "location": "https://blobtest.blob.core.windows.net/container", + "overwrite": True, + "prefix": "_output/", + }, + }, + ).result() + print(response) + + +# x-ms-original-file: 2024-11-15/DeidentifyDocuments.json +if __name__ == "__main__": + main() diff --git a/sdk/healthdataaiservices/azure-health-deidentification/generated_samples/deidentify_text.py b/sdk/healthdataaiservices/azure-health-deidentification/generated_samples/deidentify_text.py new file mode 100644 index 000000000000..bd5faead45e8 --- /dev/null +++ b/sdk/healthdataaiservices/azure-health-deidentification/generated_samples/deidentify_text.py @@ -0,0 +1,45 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for license information. +# Code generated by Microsoft (R) Python Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is regenerated. +# -------------------------------------------------------------------------- + +from azure.identity import DefaultAzureCredential + +from azure.health.deidentification import DeidentificationClient + +""" +# PREREQUISITES + pip install azure-identity + pip install azure-health-deidentification +# USAGE + python deidentify_text.py + + Before run the sample, please set the values of the client ID, tenant ID and client secret + of the AAD application as environment variables: AZURE_CLIENT_ID, AZURE_TENANT_ID, + AZURE_CLIENT_SECRET. For more info about how to get the value, please see: + https://docs.microsoft.com/azure/active-directory/develop/howto-create-service-principal-portal +""" + + +def main(): + client = DeidentificationClient( + endpoint="ENDPOINT", + credential=DefaultAzureCredential(), + ) + + response = client.deidentify_text( + body={ + "customizations": {"redactionFormat": "[{type}]"}, + "inputText": "Hello my name is John Smith.", + "operation": "Redact", + }, + ) + print(response) + + +# x-ms-original-file: 2024-11-15/DeidentifyText.json +if __name__ == "__main__": + main() diff --git a/sdk/healthdataaiservices/azure-health-deidentification/generated_samples/delete_job.py b/sdk/healthdataaiservices/azure-health-deidentification/generated_samples/delete_job.py new file mode 100644 index 000000000000..0757df58c9f9 --- /dev/null +++ b/sdk/healthdataaiservices/azure-health-deidentification/generated_samples/delete_job.py @@ -0,0 +1,40 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for license information. +# Code generated by Microsoft (R) Python Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is regenerated. +# -------------------------------------------------------------------------- + +from azure.identity import DefaultAzureCredential + +from azure.health.deidentification import DeidentificationClient + +""" +# PREREQUISITES + pip install azure-identity + pip install azure-health-deidentification +# USAGE + python delete_job.py + + Before run the sample, please set the values of the client ID, tenant ID and client secret + of the AAD application as environment variables: AZURE_CLIENT_ID, AZURE_TENANT_ID, + AZURE_CLIENT_SECRET. For more info about how to get the value, please see: + https://docs.microsoft.com/azure/active-directory/develop/howto-create-service-principal-portal +""" + + +def main(): + client = DeidentificationClient( + endpoint="ENDPOINT", + credential=DefaultAzureCredential(), + ) + + client.delete_job( + name="job_smith_documents_1", + ) + + +# x-ms-original-file: 2024-11-15/DeleteJob.json +if __name__ == "__main__": + main() diff --git a/sdk/healthdataaiservices/azure-health-deidentification/generated_samples/get_job.py b/sdk/healthdataaiservices/azure-health-deidentification/generated_samples/get_job.py new file mode 100644 index 000000000000..efb79e44d4e1 --- /dev/null +++ b/sdk/healthdataaiservices/azure-health-deidentification/generated_samples/get_job.py @@ -0,0 +1,41 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for license information. +# Code generated by Microsoft (R) Python Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is regenerated. +# -------------------------------------------------------------------------- + +from azure.identity import DefaultAzureCredential + +from azure.health.deidentification import DeidentificationClient + +""" +# PREREQUISITES + pip install azure-identity + pip install azure-health-deidentification +# USAGE + python get_job.py + + Before run the sample, please set the values of the client ID, tenant ID and client secret + of the AAD application as environment variables: AZURE_CLIENT_ID, AZURE_TENANT_ID, + AZURE_CLIENT_SECRET. For more info about how to get the value, please see: + https://docs.microsoft.com/azure/active-directory/develop/howto-create-service-principal-portal +""" + + +def main(): + client = DeidentificationClient( + endpoint="ENDPOINT", + credential=DefaultAzureCredential(), + ) + + response = client.get_job( + name="job_smith_documents_1", + ) + print(response) + + +# x-ms-original-file: 2024-11-15/GetJob.json +if __name__ == "__main__": + main() diff --git a/sdk/healthdataaiservices/azure-health-deidentification/generated_samples/list_job_documents.py b/sdk/healthdataaiservices/azure-health-deidentification/generated_samples/list_job_documents.py new file mode 100644 index 000000000000..7336adfcb8cd --- /dev/null +++ b/sdk/healthdataaiservices/azure-health-deidentification/generated_samples/list_job_documents.py @@ -0,0 +1,42 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for license information. +# Code generated by Microsoft (R) Python Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is regenerated. +# -------------------------------------------------------------------------- + +from azure.identity import DefaultAzureCredential + +from azure.health.deidentification import DeidentificationClient + +""" +# PREREQUISITES + pip install azure-identity + pip install azure-health-deidentification +# USAGE + python list_job_documents.py + + Before run the sample, please set the values of the client ID, tenant ID and client secret + of the AAD application as environment variables: AZURE_CLIENT_ID, AZURE_TENANT_ID, + AZURE_CLIENT_SECRET. For more info about how to get the value, please see: + https://docs.microsoft.com/azure/active-directory/develop/howto-create-service-principal-portal +""" + + +def main(): + client = DeidentificationClient( + endpoint="ENDPOINT", + credential=DefaultAzureCredential(), + ) + + response = client.list_job_documents_internal( + job_name="Replace this value with a string matching RegExp ^[a-zA-Z0-9][a-zA-Z0-9-_]+[a-zA-Z0-9]$", + ) + for item in response: + print(item) + + +# x-ms-original-file: 2024-11-15/ListJobDocuments.json +if __name__ == "__main__": + main() diff --git a/sdk/healthdataaiservices/azure-health-deidentification/generated_samples/list_jobs.py b/sdk/healthdataaiservices/azure-health-deidentification/generated_samples/list_jobs.py new file mode 100644 index 000000000000..75fb30c1ebbf --- /dev/null +++ b/sdk/healthdataaiservices/azure-health-deidentification/generated_samples/list_jobs.py @@ -0,0 +1,40 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for license information. +# Code generated by Microsoft (R) Python Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is regenerated. +# -------------------------------------------------------------------------- + +from azure.identity import DefaultAzureCredential + +from azure.health.deidentification import DeidentificationClient + +""" +# PREREQUISITES + pip install azure-identity + pip install azure-health-deidentification +# USAGE + python list_jobs.py + + Before run the sample, please set the values of the client ID, tenant ID and client secret + of the AAD application as environment variables: AZURE_CLIENT_ID, AZURE_TENANT_ID, + AZURE_CLIENT_SECRET. For more info about how to get the value, please see: + https://docs.microsoft.com/azure/active-directory/develop/howto-create-service-principal-portal +""" + + +def main(): + client = DeidentificationClient( + endpoint="ENDPOINT", + credential=DefaultAzureCredential(), + ) + + response = client.list_jobs_internal() + for item in response: + print(item) + + +# x-ms-original-file: 2024-11-15/ListJobs.json +if __name__ == "__main__": + main() diff --git a/sdk/healthdataaiservices/azure-health-deidentification/generated_tests/conftest.py b/sdk/healthdataaiservices/azure-health-deidentification/generated_tests/conftest.py new file mode 100644 index 000000000000..23dfda522b5b --- /dev/null +++ b/sdk/healthdataaiservices/azure-health-deidentification/generated_tests/conftest.py @@ -0,0 +1,39 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for license information. +# Code generated by Microsoft (R) Python Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is regenerated. +# -------------------------------------------------------------------------- +import os +import pytest +from dotenv import load_dotenv +from devtools_testutils import ( + test_proxy, + add_general_regex_sanitizer, + add_body_key_sanitizer, + add_header_regex_sanitizer, +) + +load_dotenv() + + +# For security, please avoid record sensitive identity information in recordings +@pytest.fixture(scope="session", autouse=True) +def add_sanitizers(test_proxy): + deidentification_subscription_id = os.environ.get( + "DEIDENTIFICATION_SUBSCRIPTION_ID", "00000000-0000-0000-0000-000000000000" + ) + deidentification_tenant_id = os.environ.get("DEIDENTIFICATION_TENANT_ID", "00000000-0000-0000-0000-000000000000") + deidentification_client_id = os.environ.get("DEIDENTIFICATION_CLIENT_ID", "00000000-0000-0000-0000-000000000000") + deidentification_client_secret = os.environ.get( + "DEIDENTIFICATION_CLIENT_SECRET", "00000000-0000-0000-0000-000000000000" + ) + add_general_regex_sanitizer(regex=deidentification_subscription_id, value="00000000-0000-0000-0000-000000000000") + add_general_regex_sanitizer(regex=deidentification_tenant_id, value="00000000-0000-0000-0000-000000000000") + add_general_regex_sanitizer(regex=deidentification_client_id, value="00000000-0000-0000-0000-000000000000") + add_general_regex_sanitizer(regex=deidentification_client_secret, value="00000000-0000-0000-0000-000000000000") + + add_header_regex_sanitizer(key="Set-Cookie", value="[set-cookie;]") + add_header_regex_sanitizer(key="Cookie", value="cookie;") + add_body_key_sanitizer(json_path="$..access_token", value="access_token") diff --git a/sdk/healthdataaiservices/azure-health-deidentification/generated_tests/test_deidentification.py b/sdk/healthdataaiservices/azure-health-deidentification/generated_tests/test_deidentification.py new file mode 100644 index 000000000000..6ce428ac5cba --- /dev/null +++ b/sdk/healthdataaiservices/azure-health-deidentification/generated_tests/test_deidentification.py @@ -0,0 +1,85 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for license information. +# Code generated by Microsoft (R) Python Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is regenerated. +# -------------------------------------------------------------------------- +import pytest +from devtools_testutils import recorded_by_proxy +from testpreparer import DeidentificationClientTestBase, DeidentificationPreparer + + +@pytest.mark.skip("you may need to update the auto-generated test case before run it") +class TestDeidentification(DeidentificationClientTestBase): + @DeidentificationPreparer() + @recorded_by_proxy + def test_get_job(self, deidentification_endpoint): + client = self.create_client(endpoint=deidentification_endpoint) + response = client.get_job( + name="str", + ) + + # please add some check logic here by yourself + # ... + + @DeidentificationPreparer() + @recorded_by_proxy + def test_begin_deidentify_documents(self, deidentification_endpoint): + client = self.create_client(endpoint=deidentification_endpoint) + response = client.begin_deidentify_documents( + name="str", + resource={ + "createdAt": "2020-02-20 00:00:00", + "lastUpdatedAt": "2020-02-20 00:00:00", + "name": "str", + "sourceLocation": {"location": "str", "prefix": "str", "extensions": ["str"]}, + "status": "str", + "targetLocation": {"location": "str", "prefix": "str", "overwrite": bool}, + "customizations": {"redactionFormat": "str", "surrogateLocale": "str"}, + "error": ~azure.core.ODataV4Format, + "operation": "str", + "startedAt": "2020-02-20 00:00:00", + "summary": {"bytesProcessed": 0, "canceled": 0, "failed": 0, "successful": 0, "total": 0}, + }, + ).result() # call '.result()' to poll until service return final result + + # please add some check logic here by yourself + # ... + + @DeidentificationPreparer() + @recorded_by_proxy + def test_cancel_job(self, deidentification_endpoint): + client = self.create_client(endpoint=deidentification_endpoint) + response = client.cancel_job( + name="str", + ) + + # please add some check logic here by yourself + # ... + + @DeidentificationPreparer() + @recorded_by_proxy + def test_delete_job(self, deidentification_endpoint): + client = self.create_client(endpoint=deidentification_endpoint) + response = client.delete_job( + name="str", + ) + + # please add some check logic here by yourself + # ... + + @DeidentificationPreparer() + @recorded_by_proxy + def test_deidentify_text(self, deidentification_endpoint): + client = self.create_client(endpoint=deidentification_endpoint) + response = client.deidentify_text( + body={ + "inputText": "str", + "customizations": {"redactionFormat": "str", "surrogateLocale": "str"}, + "operation": "str", + }, + ) + + # please add some check logic here by yourself + # ... diff --git a/sdk/healthdataaiservices/azure-health-deidentification/generated_tests/test_deidentification_async.py b/sdk/healthdataaiservices/azure-health-deidentification/generated_tests/test_deidentification_async.py new file mode 100644 index 000000000000..32f4f9cf4c59 --- /dev/null +++ b/sdk/healthdataaiservices/azure-health-deidentification/generated_tests/test_deidentification_async.py @@ -0,0 +1,88 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for license information. +# Code generated by Microsoft (R) Python Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is regenerated. +# -------------------------------------------------------------------------- +import pytest +from devtools_testutils.aio import recorded_by_proxy_async +from testpreparer import DeidentificationPreparer +from testpreparer_async import DeidentificationClientTestBaseAsync + + +@pytest.mark.skip("you may need to update the auto-generated test case before run it") +class TestDeidentificationAsync(DeidentificationClientTestBaseAsync): + @DeidentificationPreparer() + @recorded_by_proxy_async + async def test_get_job(self, deidentification_endpoint): + client = self.create_async_client(endpoint=deidentification_endpoint) + response = await client.get_job( + name="str", + ) + + # please add some check logic here by yourself + # ... + + @DeidentificationPreparer() + @recorded_by_proxy_async + async def test_begin_deidentify_documents(self, deidentification_endpoint): + client = self.create_async_client(endpoint=deidentification_endpoint) + response = await ( + await client.begin_deidentify_documents( + name="str", + resource={ + "createdAt": "2020-02-20 00:00:00", + "lastUpdatedAt": "2020-02-20 00:00:00", + "name": "str", + "sourceLocation": {"location": "str", "prefix": "str", "extensions": ["str"]}, + "status": "str", + "targetLocation": {"location": "str", "prefix": "str", "overwrite": bool}, + "customizations": {"redactionFormat": "str", "surrogateLocale": "str"}, + "error": ~azure.core.ODataV4Format, + "operation": "str", + "startedAt": "2020-02-20 00:00:00", + "summary": {"bytesProcessed": 0, "canceled": 0, "failed": 0, "successful": 0, "total": 0}, + }, + ) + ).result() # call '.result()' to poll until service return final result + + # please add some check logic here by yourself + # ... + + @DeidentificationPreparer() + @recorded_by_proxy_async + async def test_cancel_job(self, deidentification_endpoint): + client = self.create_async_client(endpoint=deidentification_endpoint) + response = await client.cancel_job( + name="str", + ) + + # please add some check logic here by yourself + # ... + + @DeidentificationPreparer() + @recorded_by_proxy_async + async def test_delete_job(self, deidentification_endpoint): + client = self.create_async_client(endpoint=deidentification_endpoint) + response = await client.delete_job( + name="str", + ) + + # please add some check logic here by yourself + # ... + + @DeidentificationPreparer() + @recorded_by_proxy_async + async def test_deidentify_text(self, deidentification_endpoint): + client = self.create_async_client(endpoint=deidentification_endpoint) + response = await client.deidentify_text( + body={ + "inputText": "str", + "customizations": {"redactionFormat": "str", "surrogateLocale": "str"}, + "operation": "str", + }, + ) + + # please add some check logic here by yourself + # ... diff --git a/sdk/healthdataaiservices/azure-health-deidentification/generated_tests/testpreparer.py b/sdk/healthdataaiservices/azure-health-deidentification/generated_tests/testpreparer.py new file mode 100644 index 000000000000..015d30dbf7af --- /dev/null +++ b/sdk/healthdataaiservices/azure-health-deidentification/generated_tests/testpreparer.py @@ -0,0 +1,26 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for license information. +# Code generated by Microsoft (R) Python Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is regenerated. +# -------------------------------------------------------------------------- +from azure.health.deidentification import DeidentificationClient +from devtools_testutils import AzureRecordedTestCase, PowerShellPreparer +import functools + + +class DeidentificationClientTestBase(AzureRecordedTestCase): + + def create_client(self, endpoint): + credential = self.get_credential(DeidentificationClient) + return self.create_client_from_credential( + DeidentificationClient, + credential=credential, + endpoint=endpoint, + ) + + +DeidentificationPreparer = functools.partial( + PowerShellPreparer, "deidentification", deidentification_endpoint="https://fake_deidentification_endpoint.com" +) diff --git a/sdk/healthdataaiservices/azure-health-deidentification/generated_tests/testpreparer_async.py b/sdk/healthdataaiservices/azure-health-deidentification/generated_tests/testpreparer_async.py new file mode 100644 index 000000000000..5fa674a0c53d --- /dev/null +++ b/sdk/healthdataaiservices/azure-health-deidentification/generated_tests/testpreparer_async.py @@ -0,0 +1,20 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for license information. +# Code generated by Microsoft (R) Python Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is regenerated. +# -------------------------------------------------------------------------- +from azure.health.deidentification.aio import DeidentificationClient +from devtools_testutils import AzureRecordedTestCase + + +class DeidentificationClientTestBaseAsync(AzureRecordedTestCase): + + def create_async_client(self, endpoint): + credential = self.get_credential(DeidentificationClient, is_async=True) + return self.create_client_from_credential( + DeidentificationClient, + credential=credential, + endpoint=endpoint, + ) diff --git a/sdk/healthdataaiservices/azure-health-deidentification/samples/async_samples/sample_create_and_wait_job_async.py b/sdk/healthdataaiservices/azure-health-deidentification/samples/async_samples/sample_create_and_wait_job_async.py index e1816415c037..fffe5fc9a67f 100644 --- a/sdk/healthdataaiservices/azure-health-deidentification/samples/async_samples/sample_create_and_wait_job_async.py +++ b/sdk/healthdataaiservices/azure-health-deidentification/samples/async_samples/sample_create_and_wait_job_async.py @@ -37,7 +37,6 @@ async def sample_create_and_wait_job_async(): from azure.core.polling import AsyncLROPoller endpoint = os.environ["AZURE_HEALTH_DEIDENTIFICATION_ENDPOINT"] - endpoint = endpoint.replace("https://", "") storage_location = os.environ["AZURE_STORAGE_ACCOUNT_LOCATION"] inputPrefix = os.environ["INPUT_PREFIX"] @@ -54,22 +53,18 @@ async def sample_create_and_wait_job_async(): location=storage_location, prefix=inputPrefix, ), - target_location=TargetStorageLocation( - location=storage_location, prefix=outputPrefix - ), + target_location=TargetStorageLocation(location=storage_location, prefix=outputPrefix), ) async with client: - lro: AsyncLROPoller = await client.begin_create_job(jobname, job) + lro: AsyncLROPoller = await client.begin_deidentify_documents(jobname, job) finished_job: DeidentificationJob = await lro.result() await credential.close() print(f"Job Name: {finished_job.name}") print(f"Job Status: {finished_job.status}") # Succeeded - print( - f"File Count: {finished_job.summary.total if finished_job.summary is not None else 0}" - ) + print(f"File Count: {finished_job.summary.total if finished_job.summary is not None else 0}") # [END sample_create_and_wait_job_async] diff --git a/sdk/healthdataaiservices/azure-health-deidentification/samples/async_samples/sample_list_job_files_async.py b/sdk/healthdataaiservices/azure-health-deidentification/samples/async_samples/sample_list_job_files_async.py index 198153982a6c..d5dc42539c24 100644 --- a/sdk/healthdataaiservices/azure-health-deidentification/samples/async_samples/sample_list_job_files_async.py +++ b/sdk/healthdataaiservices/azure-health-deidentification/samples/async_samples/sample_list_job_files_async.py @@ -36,7 +36,6 @@ async def sample_list_job_documents_async(): from azure.core.polling import AsyncLROPoller endpoint = os.environ["AZURE_HEALTH_DEIDENTIFICATION_ENDPOINT"] - endpoint = endpoint.replace("https://", "") storage_location = os.environ["AZURE_STORAGE_ACCOUNT_LOCATION"] inputPrefix = os.environ["INPUT_PREFIX"] @@ -53,14 +52,12 @@ async def sample_list_job_documents_async(): location=storage_location, prefix=inputPrefix, ), - target_location=TargetStorageLocation( - location=storage_location, prefix=outputPrefix - ), + target_location=TargetStorageLocation(location=storage_location, prefix=outputPrefix), ) print(f"Creating job with name: {jobname}") async with client: - poller: AsyncLROPoller = await client.begin_create_job(jobname, job) + poller: AsyncLROPoller = await client.begin_deidentify_documents(jobname, job) job = await poller.result() print(f"Job Status: {job.status}") diff --git a/sdk/healthdataaiservices/azure-health-deidentification/samples/async_samples/sample_list_jobs_async.py b/sdk/healthdataaiservices/azure-health-deidentification/samples/async_samples/sample_list_jobs_async.py index 546cff83b471..d1fae77b309f 100644 --- a/sdk/healthdataaiservices/azure-health-deidentification/samples/async_samples/sample_list_jobs_async.py +++ b/sdk/healthdataaiservices/azure-health-deidentification/samples/async_samples/sample_list_jobs_async.py @@ -30,7 +30,6 @@ async def sample_list_jobs_async(): from azure.health.deidentification.aio import DeidentificationClient endpoint = os.environ["AZURE_HEALTH_DEIDENTIFICATION_ENDPOINT"] - endpoint = endpoint.replace("https://", "") credential = DefaultAzureCredential() diff --git a/sdk/healthdataaiservices/azure-health-deidentification/samples/async_samples/sample_realtime_deidentification_async.py b/sdk/healthdataaiservices/azure-health-deidentification/samples/async_samples/sample_realtime_deidentification_async.py index 02e6813c9199..b2edb961d66c 100644 --- a/sdk/healthdataaiservices/azure-health-deidentification/samples/async_samples/sample_realtime_deidentification_async.py +++ b/sdk/healthdataaiservices/azure-health-deidentification/samples/async_samples/sample_realtime_deidentification_async.py @@ -30,7 +30,6 @@ async def sample_realtime_deidentification_async(): ) endpoint = os.environ["AZURE_HEALTH_DEIDENTIFICATION_ENDPOINT"] - endpoint = endpoint.replace("https://", "") credential = DefaultAzureCredential() diff --git a/sdk/healthdataaiservices/azure-health-deidentification/samples/sample_create_and_wait_job.py b/sdk/healthdataaiservices/azure-health-deidentification/samples/sample_create_and_wait_job.py index 1c63c38cd86d..3ec5fdc84083 100644 --- a/sdk/healthdataaiservices/azure-health-deidentification/samples/sample_create_and_wait_job.py +++ b/sdk/healthdataaiservices/azure-health-deidentification/samples/sample_create_and_wait_job.py @@ -37,7 +37,6 @@ def sample_create_and_wait_job(): from azure.core.polling import LROPoller endpoint = os.environ["AZURE_HEALTH_DEIDENTIFICATION_ENDPOINT"] - endpoint = endpoint.replace("https://", "") storage_location = os.environ["AZURE_STORAGE_ACCOUNT_LOCATION"] inputPrefix = os.environ["INPUT_PREFIX"] @@ -54,20 +53,16 @@ def sample_create_and_wait_job(): location=storage_location, prefix=inputPrefix, ), - target_location=TargetStorageLocation( - location=storage_location, prefix=outputPrefix - ), + target_location=TargetStorageLocation(location=storage_location, prefix=outputPrefix), ) - lro: LROPoller = client.begin_create_job(jobname, job) + lro: LROPoller = client.begin_deidentify_documents(jobname, job) lro.wait(timeout=60) finished_job: DeidentificationJob = lro.result() print(f"Job Name: {finished_job.name}") print(f"Job Status: {finished_job.status}") - print( - f"File Count: {finished_job.summary.total if finished_job.summary is not None else 0}" - ) + print(f"File Count: {finished_job.summary.total if finished_job.summary is not None else 0}") # [END sample_create_and_wait_job] diff --git a/sdk/healthdataaiservices/azure-health-deidentification/samples/sample_list_job_files.py b/sdk/healthdataaiservices/azure-health-deidentification/samples/sample_list_job_files.py index 438ebfaa267a..dba1c6f17a55 100644 --- a/sdk/healthdataaiservices/azure-health-deidentification/samples/sample_list_job_files.py +++ b/sdk/healthdataaiservices/azure-health-deidentification/samples/sample_list_job_files.py @@ -36,7 +36,6 @@ def sample_list_job_documents(): from azure.core.polling import LROPoller endpoint = os.environ["AZURE_HEALTH_DEIDENTIFICATION_ENDPOINT"] - endpoint = endpoint.replace("https://", "") storage_location = os.environ["AZURE_STORAGE_ACCOUNT_LOCATION"] inputPrefix = os.environ["INPUT_PREFIX"] @@ -53,13 +52,11 @@ def sample_list_job_documents(): location=storage_location, prefix=inputPrefix, ), - target_location=TargetStorageLocation( - location=storage_location, prefix=outputPrefix - ), + target_location=TargetStorageLocation(location=storage_location, prefix=outputPrefix), ) print(f"Creating job with name: {jobname}") - poller: LROPoller = client.begin_create_job(jobname, job) + poller: LROPoller = client.begin_deidentify_documents(jobname, job) poller.wait(timeout=60) job = poller.result() diff --git a/sdk/healthdataaiservices/azure-health-deidentification/samples/sample_list_jobs.py b/sdk/healthdataaiservices/azure-health-deidentification/samples/sample_list_jobs.py index 4eb70c5c7af0..19e6be7e9806 100644 --- a/sdk/healthdataaiservices/azure-health-deidentification/samples/sample_list_jobs.py +++ b/sdk/healthdataaiservices/azure-health-deidentification/samples/sample_list_jobs.py @@ -31,7 +31,6 @@ def sample_list_jobs(): from azure.health.deidentification import DeidentificationClient endpoint = os.environ["AZURE_HEALTH_DEIDENTIFICATION_ENDPOINT"] - endpoint = endpoint.replace("https://", "") credential = DefaultAzureCredential() diff --git a/sdk/healthdataaiservices/azure-health-deidentification/samples/sample_realtime_deidentification.py b/sdk/healthdataaiservices/azure-health-deidentification/samples/sample_realtime_deidentification.py index 8ebef8c1af61..c24ed66dfbb7 100644 --- a/sdk/healthdataaiservices/azure-health-deidentification/samples/sample_realtime_deidentification.py +++ b/sdk/healthdataaiservices/azure-health-deidentification/samples/sample_realtime_deidentification.py @@ -31,7 +31,6 @@ def sample_realtime_deidentification(): ) endpoint = os.environ["AZURE_HEALTH_DEIDENTIFICATION_ENDPOINT"] - endpoint = endpoint.replace("https://", "") credential = DefaultAzureCredential() diff --git a/sdk/healthdataaiservices/azure-health-deidentification/sdk_packaging.toml b/sdk/healthdataaiservices/azure-health-deidentification/sdk_packaging.toml new file mode 100644 index 000000000000..e7687fdae93b --- /dev/null +++ b/sdk/healthdataaiservices/azure-health-deidentification/sdk_packaging.toml @@ -0,0 +1,2 @@ +[packaging] +auto_update = false \ No newline at end of file diff --git a/sdk/healthdataaiservices/azure-health-deidentification/tests/conftest.py b/sdk/healthdataaiservices/azure-health-deidentification/tests/conftest.py index e7f366daf096..92b008ef4c0a 100644 --- a/sdk/healthdataaiservices/azure-health-deidentification/tests/conftest.py +++ b/sdk/healthdataaiservices/azure-health-deidentification/tests/conftest.py @@ -3,7 +3,8 @@ import os from devtools_testutils import ( add_body_key_sanitizer, - add_general_string_sanitizer, + add_continuation_sanitizer, + add_general_regex_sanitizer, remove_batch_sanitizers, test_proxy, ) @@ -23,15 +24,16 @@ def start_proxy(test_proxy, patch_sleep, patch_async_sleep): @pytest.fixture(scope="session", autouse=True) def create_session_uniquifier(): if ( - os.environ.get("AZURE_TEST_RUN_LIVE", "false").lower() - == "true" # Don't override uniquifier by default + os.environ.get("AZURE_TEST_RUN_LIVE", "false").lower() == "true" # Don't override uniquifier by default and os.environ.get("AZURE_SKIP_LIVE_RECORDING", "false").lower() != "true" ): + print("Creating new uniquifier for live test run.") uniquifier = uuid.uuid4().hex[:6] os.environ["HEALTHDATAAISERVICES_UNIQUIFIER"] = uniquifier with open(uniquifier_file, "w") as file: file.write(uniquifier) else: + print("Using existing") with open(uniquifier_file, "r") as file: uniquifier = file.read() os.environ["HEALTHDATAAISERVICES_UNIQUIFIER"] = uniquifier @@ -43,13 +45,7 @@ def add_sanitizers(test_proxy): # $..id # uri sanitization in favor of substitution remove_batch_sanitizers(["AZSDK3493", "AZSDK3430", "AZSDK4001"]) - account_name = os.environ.get( - "HEALTHDATAAISERVICES_STORAGE_ACCOUNT_NAME", "Not Found." - ) - container_name = os.environ.get( - "HEALTHDATAAISERVICES_STORAGE_CONTAINER_NAME", "Not Found." - ) - add_body_key_sanitizer( - json_path="..location", - value=f"https://{account_name}.blob.core.windows.net:443/{container_name}", + add_general_regex_sanitizer( + regex="continuationToken=[^&]*", + value="continuationToken=Sanitized" ) diff --git a/sdk/healthdataaiservices/azure-health-deidentification/tests/data/example_patient_1/row-2-data.txt b/sdk/healthdataaiservices/azure-health-deidentification/tests/data/example_patient_1/row-2-data.txt new file mode 100644 index 000000000000..6fc315d75c92 --- /dev/null +++ b/sdk/healthdataaiservices/azure-health-deidentification/tests/data/example_patient_1/row-2-data.txt @@ -0,0 +1,54 @@ +"CHIEF COMPLAINT + +Follow-up of chronic problems. + +HISTORY OF PRESENT ILLNESS + +Brian White is a 58-year-old male with a past medical history significant for congestive heart failure and hypertension, who presents today for follow-up of his chronic problems. + +The patient states he has been feeling out of sorts lately. He is not sure if it is due to the change in the seasons or due to performing lots of projects and some construction on his home. He reports fatigue and lightheadedness. This has been going on for about 5 weeks. While exerting energy, he has experienced some shortness of breath and chest cramps. The patient also notes a slight cough, but he is not sure if it is just the change in seasons. + +He feels bloated every once in a while. His diet has been a little bit of a struggle. They had construction on their kitchen begin over Labor Day weekend, and have been eating less healthy food as a result. + +Regarding his heart failure, he has been pretty good with his salt intake. He has been pretty good about his diet since the last year and is staying on top of that as much as possible. The patient has continued to utilize Lasix daily. + +For his hypertension, this has been well controlled with lisinopril 20 mg a day. He has continued to monitor his blood pressure regularly. + +The patient did the review of systems sheet when he checked in. He denies weight gain, swelling in the lower extremities, fevers, chills, dizziness, nausea, vomiting, and diarrhea. + +REVIEW OF SYSTEMS + + Constitutional: Endorses fatigue. Denies fevers, chills, or weight loss. + Cardiovascular: Endorses chest pain or dyspnea on exertion. + Respiratory: Endorses cough and shortness of breath. + Gastrointestinal: Endorses bloating. + +PHYSICAL EXAMINATION + + Neck: JVD 8 cm. + Respiratory: Rales bilateral bases. + Cardiovascular: 3/6 systolic ejection murmur. + Musculoskeletal: 1+ pitting edema bilateral lower extremities. + +RESULTS + +X-ray of the chest demonstrates a mild amount of fluid in the lungs. + +Echocardiogram demonstrates decreased ejection fraction of 45% and mild mitral regurgitation. + +ASSESSMENT AND PLAN + +Brian White is a 58-year-old male with a past medical history significant for congestive heart failure and hypertension, who presents today for follow up of his chronic problems. + +Congestive heart failure. + Medical Reasoning: The patient reports increased fatigue, dizziness, and chest discomfort on exertion. He also exhibits some jugular venous distention, lung base crackles, and lower extremity edema on exam today. He has been compliant with his current medications but admits to dietary indiscretion lately. His recent echocardiogram demonstrated a reduced ejection fraction of 45%, as well as mitral regurgitation. + Additional Testing: We will order a repeat echocardiogram. + Medical Treatment: Increase Lasix to 80 mg daily. + Patient Education and Counseling: I advised the patient to monitor and record his daily weight and report those to me via the patient portal. He will contact me should he continue to experience any dyspnea. + +Hypertension. + Medical Reasoning: This is well controlled based on home monitoring. + Medical Treatment: Continue lisinopril 20 mg daily. + Patient Education and Counseling: I advised him to monitor and record his blood pressures at home and report these to me via the patient portal. + +Patient Agreements: The patient understands and agrees with the recommended medical treatment plan." \ No newline at end of file diff --git a/sdk/healthdataaiservices/azure-health-deidentification/tests/deid_base_test_case.py b/sdk/healthdataaiservices/azure-health-deidentification/tests/deid_base_test_case.py index 383e72676898..5d2b3a502afe 100644 --- a/sdk/healthdataaiservices/azure-health-deidentification/tests/deid_base_test_case.py +++ b/sdk/healthdataaiservices/azure-health-deidentification/tests/deid_base_test_case.py @@ -36,24 +36,14 @@ class DeidBaseTestCase(AzureRecordedTestCase): def make_client(self, endpoint) -> DeidentificationClient: credential = self.get_credential(DeidentificationClient) client = self.create_client_from_credential( - DeidentificationClient, - credential=credential, - # Client library expects just hostname - endpoint=endpoint.replace("https://", ""), - # TODO: test-proxy not playing well with SSL verification - # connection_verify=False, + DeidentificationClient, credential=credential, endpoint=endpoint, connection_verify=False ) return client def make_client_async(self, endpoint) -> DeidentificationClientAsync: credential = self.get_credential(DeidentificationClientAsync) client = self.create_client_from_credential( - DeidentificationClientAsync, - credential=credential, - # Client library expects just hostname - endpoint=endpoint.replace("https://", ""), - # TODO: test-proxy not playing well with SSL verification - connection_verify=False, + DeidentificationClientAsync, credential=credential, endpoint=endpoint, connection_verify=False ) return client @@ -66,7 +56,15 @@ def generate_job_name(self) -> str: def get_storage_location(self, kwargs): storage_name: str = kwargs.pop("healthdataaiservices_storage_account_name") container_name: str = kwargs.pop("healthdataaiservices_storage_container_name") - storage_location = ( - f"https://{storage_name}.blob.core.windows.net/{container_name}" - ) + storage_location = f"https://{storage_name}.blob.core.windows.net/{container_name}" + sas_uri = os.environ.get("HEALTHDATAAISERVICES_SAS_URI", "") + if ( + os.environ.get("AZURE_TEST_RUN_LIVE", "false").lower() == "true" # Don't override uniquifier by default + and os.environ.get("AZURE_SKIP_LIVE_RECORDING", "false").lower() != "true" + ): + if sas_uri != "" and os.environ: + print(f"Using SAS URI: {sas_uri}") + return sas_uri + + print(f"Using storage location: {storage_location}") return storage_location diff --git a/sdk/healthdataaiservices/azure-health-deidentification/tests/test_create_delete.py b/sdk/healthdataaiservices/azure-health-deidentification/tests/test_create_delete.py index cf2516fb7a8c..cadc8a6498e6 100644 --- a/sdk/healthdataaiservices/azure-health-deidentification/tests/test_create_delete.py +++ b/sdk/healthdataaiservices/azure-health-deidentification/tests/test_create_delete.py @@ -26,27 +26,24 @@ def test_create_cancel_delete(self, **kwargs): location=storage_location, prefix="example_patient_1", ), - target_location=TargetStorageLocation( - location=storage_location, prefix=self.OUTPUT_PATH - ), - operation=OperationType.SURROGATE, - data_type=DocumentDataType.PLAINTEXT, + target_location=TargetStorageLocation(location=storage_location, prefix=self.OUTPUT_PATH, overwrite=True), + operation=DeidentificationOperationType.SURROGATE, ) - client.begin_create_job(jobname, job) + client.begin_deidentify_documents(jobname, job) job = client.get_job(jobname) - while job.status == JobStatus.NOT_STARTED: + while job.status == DeidentificationJobStatus.NOT_STARTED: self.sleep(2) job = client.get_job(jobname) assert job.error is None, "Job should not have an error" - assert job.status == JobStatus.RUNNING, "Job should be running" + assert job.status == DeidentificationJobStatus.RUNNING, "Job should be running" job = client.cancel_job(jobname) assert job.error is None, "Job should not have an error after cancelling" - assert job.status == JobStatus.CANCELED, "Job should be cancelled" + assert job.status == DeidentificationJobStatus.CANCELED, "Job should be cancelled" client.delete_job(jobname) diff --git a/sdk/healthdataaiservices/azure-health-deidentification/tests/test_create_delete_async.py b/sdk/healthdataaiservices/azure-health-deidentification/tests/test_create_delete_async.py index ccf32214ba37..3350725e0629 100644 --- a/sdk/healthdataaiservices/azure-health-deidentification/tests/test_create_delete_async.py +++ b/sdk/healthdataaiservices/azure-health-deidentification/tests/test_create_delete_async.py @@ -27,25 +27,23 @@ async def test_create_cancel_delete_async(self, **kwargs): location=storage_location, prefix="example_patient_1", ), - target_location=TargetStorageLocation( - location=storage_location, prefix=self.OUTPUT_PATH - ), + target_location=TargetStorageLocation(location=storage_location, prefix=self.OUTPUT_PATH, overwrite=True), ) - await client.begin_create_job(jobname, job) + await client.begin_deidentify_documents(jobname, job) job = await client.get_job(jobname) - while job.status == JobStatus.NOT_STARTED: + while job.status == DeidentificationJobStatus.NOT_STARTED: self.sleep(2) job = await client.get_job(jobname) assert job.error is None, "Job should not have an error" - assert job.status == JobStatus.RUNNING, "Job should be running" + assert job.status == DeidentificationJobStatus.RUNNING, "Job should be running" job = await client.cancel_job(jobname) assert job.error is None, "Job should not have an error after cancelling" - assert job.status == JobStatus.CANCELED, "Job should be cancelled" + assert job.status == DeidentificationJobStatus.CANCELED, "Job should be cancelled" await client.delete_job(jobname) diff --git a/sdk/healthdataaiservices/azure-health-deidentification/tests/test_create_list.py b/sdk/healthdataaiservices/azure-health-deidentification/tests/test_create_list.py index 9cb3fedf0d5c..2bdd04a6429c 100644 --- a/sdk/healthdataaiservices/azure-health-deidentification/tests/test_create_list.py +++ b/sdk/healthdataaiservices/azure-health-deidentification/tests/test_create_list.py @@ -24,12 +24,12 @@ def test_create_list(self, **kwargs): location=storage_location, prefix=inputPrefix, ), - target_location=TargetStorageLocation(location=storage_location, prefix=self.OUTPUT_PATH), - operation=OperationType.TAG, - data_type=DocumentDataType.PLAINTEXT, + target_location=TargetStorageLocation(location=storage_location, prefix=self.OUTPUT_PATH, overwrite=True), + operation=DeidentificationOperationType.REDACT, + customizations=DeidentificationJobCustomizationOptions(redaction_format="[{type}]"), ) - client.begin_create_job(jobname, job) + client.begin_deidentify_documents(jobname, job) jobs = client.list_jobs() job = None @@ -42,11 +42,12 @@ def test_create_list(self, **kwargs): elif jobsToLookThrough <= 0: raise Exception("Job not found in list_jobs") + assert job is not None assert job.name == jobname - assert job.status == JobStatus.NOT_STARTED or job.status == JobStatus.RUNNING - assert job.operation == OperationType.TAG + assert job.status == DeidentificationJobStatus.NOT_STARTED or job.status == DeidentificationJobStatus.RUNNING + assert job.operation == DeidentificationOperationType.REDACT assert job.error is None - assert job.summary is None assert job.created_at is not None assert job.last_updated_at is not None - assert job.redaction_format is None + assert job.customizations is not None + assert job.customizations.redaction_format == "[{type}]" diff --git a/sdk/healthdataaiservices/azure-health-deidentification/tests/test_create_list_async.py b/sdk/healthdataaiservices/azure-health-deidentification/tests/test_create_list_async.py index 51383e5313a7..518f7b5f5eee 100644 --- a/sdk/healthdataaiservices/azure-health-deidentification/tests/test_create_list_async.py +++ b/sdk/healthdataaiservices/azure-health-deidentification/tests/test_create_list_async.py @@ -25,15 +25,12 @@ async def test_create_list_async(self, **kwargs): location=storage_location, prefix=inputPrefix, ), - target_location=TargetStorageLocation( - location=storage_location, prefix=self.OUTPUT_PATH - ), - operation=OperationType.TAG, - data_type=DocumentDataType.PLAINTEXT, + target_location=TargetStorageLocation(location=storage_location, prefix=self.OUTPUT_PATH, overwrite=True), + operation=DeidentificationOperationType.TAG, ) - await client.begin_create_job(jobname, job) - jobs = client.list_jobs() + await client.begin_deidentify_documents(jobname, job) + jobs = client.list_jobs(maxpagesize=1) job = None jobsToLookThrough = 10 @@ -45,11 +42,11 @@ async def test_create_list_async(self, **kwargs): elif jobsToLookThrough <= 0: raise Exception("Job not found in list_jobs") + assert job is not None assert job.name == jobname - assert job.status == JobStatus.NOT_STARTED or job.status == JobStatus.RUNNING - assert job.operation == OperationType.TAG + assert job.status == DeidentificationJobStatus.NOT_STARTED or job.status == DeidentificationJobStatus.RUNNING + assert job.operation == DeidentificationOperationType.TAG assert job.error is None - assert job.summary is None assert job.created_at is not None assert job.last_updated_at is not None - assert job.redaction_format is None + assert job.customizations is not None diff --git a/sdk/healthdataaiservices/azure-health-deidentification/tests/test_create_wait_finish.py b/sdk/healthdataaiservices/azure-health-deidentification/tests/test_create_wait_finish.py index 79954337cf3b..00349e1808f2 100644 --- a/sdk/healthdataaiservices/azure-health-deidentification/tests/test_create_wait_finish.py +++ b/sdk/healthdataaiservices/azure-health-deidentification/tests/test_create_wait_finish.py @@ -12,7 +12,7 @@ class TestHealthDeidentificationCreateJobWaitUntil(DeidBaseTestCase): @recorded_by_proxy def test_create_wait_finish(self, **kwargs): endpoint: str = kwargs.pop("healthdataaiservices_deid_service_endpoint") - inputPrefix = "example_patient_1" + input_prefix = "example_patient_1" storage_location: str = self.get_storage_location(kwargs) client = self.make_client(endpoint) assert client is not None @@ -22,39 +22,38 @@ def test_create_wait_finish(self, **kwargs): job = DeidentificationJob( source_location=SourceStorageLocation( location=storage_location, - prefix=inputPrefix, + prefix=input_prefix, ), - target_location=TargetStorageLocation( - location=storage_location, prefix=self.OUTPUT_PATH - ), - operation=OperationType.SURROGATE, - data_type=DocumentDataType.PLAINTEXT, + target_location=TargetStorageLocation(location=storage_location, prefix=self.OUTPUT_PATH, overwrite=True), + operation=DeidentificationOperationType.SURROGATE, ) - lro: LROPoller = client.begin_create_job(jobname, job) + lro: LROPoller = client.begin_deidentify_documents(jobname, job) lro.wait(timeout=60) finished_job: DeidentificationJob = lro.result() - assert finished_job.status == JobStatus.SUCCEEDED + assert finished_job.status == DeidentificationJobStatus.SUCCEEDED assert finished_job.name == jobname - assert finished_job.operation == OperationType.SURROGATE - assert finished_job.data_type == DocumentDataType.PLAINTEXT - assert finished_job.summary.total == 2 - assert finished_job.summary.successful == 2 + assert finished_job.operation == DeidentificationOperationType.SURROGATE + assert finished_job.summary is not None + assert finished_job.summary.total == 3 + assert finished_job.summary.successful == 3 assert finished_job.summary.failed == 0 - assert finished_job.started_at > finished_job.created_at + assert finished_job.started_at is not None and finished_job.started_at > finished_job.created_at assert finished_job.last_updated_at > finished_job.started_at - assert finished_job.redaction_format is None + assert finished_job.customizations is not None + assert finished_job.customizations.surrogate_locale == "en-US" assert finished_job.error is None - assert finished_job.source_location.prefix == inputPrefix + assert finished_job.source_location.prefix == input_prefix files = client.list_job_documents(jobname) count = 0 for my_file in files: assert len(my_file.id) == 36 # GUID - assert my_file.input.path.startswith(inputPrefix) + assert input_prefix in my_file.input.location assert my_file.status == OperationState.SUCCEEDED - assert my_file.output.path.startswith(self.OUTPUT_PATH) + assert my_file.output is not None + assert self.OUTPUT_PATH in my_file.output.location count += 1 - assert count == 2, f"Expected 2 files, found {count}" + assert count == 3, f"Expected 3 files, found {count}" diff --git a/sdk/healthdataaiservices/azure-health-deidentification/tests/test_create_wait_finish_async.py b/sdk/healthdataaiservices/azure-health-deidentification/tests/test_create_wait_finish_async.py index 97af65c1d224..f13a2e00779f 100644 --- a/sdk/healthdataaiservices/azure-health-deidentification/tests/test_create_wait_finish_async.py +++ b/sdk/healthdataaiservices/azure-health-deidentification/tests/test_create_wait_finish_async.py @@ -14,7 +14,7 @@ class TestHealthDeidentificationCreateJobWaitUntil(DeidBaseTestCase): @recorded_by_proxy_async async def test_create_wait_finish_async(self, **kwargs): endpoint: str = kwargs.pop("healthdataaiservices_deid_service_endpoint") - inputPrefix = "example_patient_1" + input_prefix = "example_patient_1" storage_location: str = self.get_storage_location(kwargs) client = self.make_client_async(endpoint) assert client is not None @@ -24,39 +24,39 @@ async def test_create_wait_finish_async(self, **kwargs): job = DeidentificationJob( source_location=SourceStorageLocation( location=storage_location, - prefix=inputPrefix, + prefix=input_prefix, ), - target_location=TargetStorageLocation( - location=storage_location, prefix=self.OUTPUT_PATH - ), - operation=OperationType.SURROGATE, - data_type=DocumentDataType.PLAINTEXT, + target_location=TargetStorageLocation(location=storage_location, prefix=self.OUTPUT_PATH, overwrite=True), + operation=DeidentificationOperationType.SURROGATE, ) - lro: AsyncLROPoller = await client.begin_create_job(jobname, job) - lro.wait() + lro: AsyncLROPoller = await client.begin_deidentify_documents(jobname, job) + await lro.wait() finished_job: DeidentificationJob = await lro.result() - assert finished_job.status == JobStatus.SUCCEEDED + assert finished_job.status == DeidentificationJobStatus.SUCCEEDED assert finished_job.name == jobname - assert finished_job.operation == OperationType.SURROGATE - assert finished_job.data_type == DocumentDataType.PLAINTEXT - assert finished_job.summary.total == 2 - assert finished_job.summary.successful == 2 + assert finished_job.operation == DeidentificationOperationType.SURROGATE + assert finished_job.summary is not None + assert finished_job.summary.total == 3 + assert finished_job.summary.successful == 3 assert finished_job.summary.failed == 0 + assert finished_job.started_at is not None assert finished_job.started_at > finished_job.created_at assert finished_job.last_updated_at > finished_job.started_at - assert finished_job.redaction_format is None + assert finished_job.customizations is not None + assert finished_job.customizations.surrogate_locale == "en-US" assert finished_job.error is None - assert finished_job.source_location.prefix == inputPrefix + assert finished_job.source_location.prefix == input_prefix files = client.list_job_documents(jobname) count = 0 async for my_file in files: assert len(my_file.id) == 36 # GUID - assert my_file.input.path.startswith(inputPrefix) + assert input_prefix in my_file.input.location assert my_file.status == OperationState.SUCCEEDED - assert my_file.output.path.startswith(self.OUTPUT_PATH) + assert my_file.output is not None + assert self.OUTPUT_PATH in my_file.output.location count += 1 - assert count == 2, f"Expected 2 files, found {count}" + assert count == 3, f"Expected 3 files, found {count}" diff --git a/sdk/healthdataaiservices/azure-health-deidentification/tests/test_exception_throws.py b/sdk/healthdataaiservices/azure-health-deidentification/tests/test_exception_throws.py index ff0daaad26d6..965590adab5b 100644 --- a/sdk/healthdataaiservices/azure-health-deidentification/tests/test_exception_throws.py +++ b/sdk/healthdataaiservices/azure-health-deidentification/tests/test_exception_throws.py @@ -28,18 +28,18 @@ def test_exception_throws(self, **kwargs): location=storage_location, prefix="no_files_in_this_folder", ), - target_location=TargetStorageLocation(location=storage_location, prefix=self.OUTPUT_PATH), - operation=OperationType.SURROGATE, - data_type=DocumentDataType.PLAINTEXT, + target_location=TargetStorageLocation(location=storage_location, prefix=self.OUTPUT_PATH, overwrite=True), + operation=DeidentificationOperationType.SURROGATE, ) - lro: LROPoller = client.begin_create_job(jobname, job) + lro: LROPoller = client.begin_deidentify_documents(jobname, job) with pytest.raises(HttpResponseError): lro.wait(timeout=60) job = client.get_job(jobname) - assert job.status == JobStatus.FAILED + assert job.status == DeidentificationJobStatus.FAILED assert job.error is not None - assert job.error.code == "JobValidationError" + assert job.error.code == "EmptyJob" + assert job.error.message is not None assert len(job.error.message) > 10 diff --git a/sdk/healthdataaiservices/azure-health-deidentification/tests/test_exception_throws_async.py b/sdk/healthdataaiservices/azure-health-deidentification/tests/test_exception_throws_async.py index 2f1858a2ce9e..356154e716bd 100644 --- a/sdk/healthdataaiservices/azure-health-deidentification/tests/test_exception_throws_async.py +++ b/sdk/healthdataaiservices/azure-health-deidentification/tests/test_exception_throws_async.py @@ -29,20 +29,18 @@ async def test_exception_throws_async(self, **kwargs): location=storage_location, prefix="no_files_in_this_folder", ), - target_location=TargetStorageLocation( - location=storage_location, prefix=self.OUTPUT_PATH - ), - operation=OperationType.SURROGATE, - data_type=DocumentDataType.PLAINTEXT, + target_location=TargetStorageLocation(location=storage_location, prefix=self.OUTPUT_PATH, overwrite=True), + operation=DeidentificationOperationType.SURROGATE, ) - lro: AsyncLROPoller = await client.begin_create_job(jobname, job) + lro: AsyncLROPoller = await client.begin_deidentify_documents(jobname, job) with pytest.raises(HttpResponseError): await lro.wait() job = await client.get_job(jobname) - assert job.status == JobStatus.FAILED + assert job.status == DeidentificationJobStatus.FAILED assert job.error is not None - assert job.error.code == "JobValidationError" + assert job.error.code == "EmptyJob" + assert job.error.message is not None assert len(job.error.message) > 10 diff --git a/sdk/healthdataaiservices/azure-health-deidentification/tests/test_hello_world.py b/sdk/healthdataaiservices/azure-health-deidentification/tests/test_hello_world.py index 32e700227510..358fb70a7aad 100644 --- a/sdk/healthdataaiservices/azure-health-deidentification/tests/test_hello_world.py +++ b/sdk/healthdataaiservices/azure-health-deidentification/tests/test_hello_world.py @@ -15,12 +15,10 @@ def test_hello_world(self, healthdataaiservices_deid_service_endpoint): assert client is not None content = DeidentificationContent( - input_text="Hello, my name is John Smith.", - operation=OperationType.SURROGATE, - data_type=DocumentDataType.PLAINTEXT, + input_text="Hello, my name is John Smith.", operation=DeidentificationOperationType.SURROGATE ) - result: DeidentificationResult = client.deidentify(content) + result: DeidentificationResult = client.deidentify_text(content) assert result is not None assert result.output_text is not None diff --git a/sdk/healthdataaiservices/azure-health-deidentification/tests/test_hello_world_async.py b/sdk/healthdataaiservices/azure-health-deidentification/tests/test_hello_world_async.py index 6506850de0f8..1a75f55dc524 100644 --- a/sdk/healthdataaiservices/azure-health-deidentification/tests/test_hello_world_async.py +++ b/sdk/healthdataaiservices/azure-health-deidentification/tests/test_hello_world_async.py @@ -16,12 +16,10 @@ async def test_hello_world_async(self, healthdataaiservices_deid_service_endpoin assert client is not None content = DeidentificationContent( - input_text="Hello, my name is John Smith.", - operation=OperationType.SURROGATE, - data_type=DocumentDataType.PLAINTEXT, + input_text="Hello, my name is John Smith.", operation=DeidentificationOperationType.SURROGATE ) - result: DeidentificationResult = await client.deidentify(content) + result: DeidentificationResult = await client.deidentify_text(content) assert result is not None assert result.output_text is not None diff --git a/sdk/healthdataaiservices/azure-health-deidentification/tests/test_list_job_docs_pagination.py b/sdk/healthdataaiservices/azure-health-deidentification/tests/test_list_job_docs_pagination.py new file mode 100644 index 000000000000..2c6ab0ec7df3 --- /dev/null +++ b/sdk/healthdataaiservices/azure-health-deidentification/tests/test_list_job_docs_pagination.py @@ -0,0 +1,68 @@ +from deid_base_test_case import * +from devtools_testutils import ( + recorded_by_proxy, +) + +from azure.health.deidentification.models import * +from azure.core.polling import LROPoller +from azure.core.paging import ItemPaged + + +class TestHealthDeidentificationCreateAndListJob(DeidBaseTestCase): + @BatchEnv() + @recorded_by_proxy + def test_list_job_docs_pagination(self, **kwargs): + """ + This test is verifying that pagination is working as expected. + The nextLink is being obfuscated by the SDK so we want to verify it is being used and not ignored + """ + endpoint: str = kwargs.pop("healthdataaiservices_deid_service_endpoint") + inputPrefix = "example_patient_1" + storage_location: str = self.get_storage_location(kwargs) + client = self.make_client(endpoint) + assert client is not None + + jobname = self.generate_job_name() + + job = DeidentificationJob( + source_location=SourceStorageLocation( + location=storage_location, + prefix=inputPrefix, + ), + target_location=TargetStorageLocation(location=storage_location, prefix=self.OUTPUT_PATH, overwrite=True), + operation=DeidentificationOperationType.REDACT, + customizations=DeidentificationJobCustomizationOptions(redaction_format="[{type}]"), + ) + + lro: LROPoller = client.begin_deidentify_documents(jobname, job) + lro.wait(timeout=60) + job_documents = client.list_job_documents(job_name=jobname, maxpagesize=2) + + _get_next = job_documents._args[0] + _extract_data = job_documents._args[1] + + job_documents_paged = ItemPaged( + get_next=_get_next, + extract_data=_extract_data, + ) + + job_ids = [] + + # Verify the first page contains our maxpagesize of 2 + page_iterator = job_documents_paged.by_page() + first_page = next(page_iterator) + first_page_items = list(first_page) + assert len(first_page_items) == 2, f"Expected 2 items in the first page, found {len(first_page_items)}" + job_ids.extend(item.id for item in first_page_items) + + # Verify there are no duplicates + assert len(set(job_ids)) == 2 + + # Verify the second page has the remaining 1 item + second_page = next(page_iterator) + second_page_items = list(second_page) + assert len(second_page_items) == 1, f"Expected 1 item in the second page, found {len(second_page_items)}" + job_ids.extend(item.id for item in second_page_items) + + # Verify the total count and uniqueness of job IDs + assert len(set(job_ids)) == 3, "Each job ID should be unique" diff --git a/sdk/healthdataaiservices/azure-health-deidentification/tests/uniquifier.conf b/sdk/healthdataaiservices/azure-health-deidentification/tests/uniquifier.conf index 5ceb18d86ac6..4417e9ddbb78 100644 --- a/sdk/healthdataaiservices/azure-health-deidentification/tests/uniquifier.conf +++ b/sdk/healthdataaiservices/azure-health-deidentification/tests/uniquifier.conf @@ -1 +1 @@ -10ea88 \ No newline at end of file +a0c9d4 \ No newline at end of file diff --git a/sdk/healthdataaiservices/azure-health-deidentification/tsp-location.yaml b/sdk/healthdataaiservices/azure-health-deidentification/tsp-location.yaml index 090caaf79c28..e2768130503c 100644 --- a/sdk/healthdataaiservices/azure-health-deidentification/tsp-location.yaml +++ b/sdk/healthdataaiservices/azure-health-deidentification/tsp-location.yaml @@ -1,5 +1,4 @@ directory: specification/healthdataaiservices/HealthDataAIServices.DeidServices -commit: 2771da5baeee73dfd70b2a5f2813a55549c2aa73 -additionalDirectories: [] +commit: e5ad7a2048e1b6778fbf11b024bee152a282853a repo: Azure/azure-rest-api-specs - +additionalDirectories: