diff --git a/azure-mgmt-media/MANIFEST.in b/azure-mgmt-media/MANIFEST.in index 6ceb27f7a96e..e4884efef41b 100644 --- a/azure-mgmt-media/MANIFEST.in +++ b/azure-mgmt-media/MANIFEST.in @@ -1,3 +1,4 @@ +recursive-include tests *.py *.yaml include *.rst include azure/__init__.py include azure/mgmt/__init__.py diff --git a/azure-mgmt-media/README.rst b/azure-mgmt-media/README.rst index d130e9d26dae..ad17d99b5e5a 100644 --- a/azure-mgmt-media/README.rst +++ b/azure-mgmt-media/README.rst @@ -14,25 +14,6 @@ For the older Azure Service Management (ASM) libraries, see For a more complete set of Azure libraries, see the `azure `__ bundle package. -Compatibility -============= - -**IMPORTANT**: If you have an earlier version of the azure package -(version < 1.0), you should uninstall it before installing this package. - -You can check the version using pip: - -.. code:: shell - - pip freeze - -If you see azure==0.11.0 (or any version below 1.0), uninstall it first: - -.. code:: shell - - pip uninstall azure - - Usage ===== diff --git a/azure-mgmt-media/azure/mgmt/media/models/__init__.py b/azure-mgmt-media/azure/mgmt/media/models/__init__.py index 313de9af0224..b288dbfab357 100644 --- a/azure-mgmt-media/azure/mgmt/media/models/__init__.py +++ b/azure-mgmt-media/azure/mgmt/media/models/__init__.py @@ -70,6 +70,7 @@ from .codec_py3 import Codec from .audio_py3 import Audio from .aac_audio_py3 import AacAudio + from .face_detector_preset_py3 import FaceDetectorPreset from .audio_analyzer_preset_py3 import AudioAnalyzerPreset from .overlay_py3 import Overlay from .audio_overlay_py3 import AudioOverlay @@ -212,6 +213,7 @@ from .codec import Codec from .audio import Audio from .aac_audio import AacAudio + from .face_detector_preset import FaceDetectorPreset from .audio_analyzer_preset import AudioAnalyzerPreset from .overlay import Overlay from .audio_overlay import AudioOverlay @@ -321,6 +323,7 @@ ContentKeyPolicyRestrictionTokenType, ContentKeyPolicyFairPlayRentalAndLeaseKeyType, AacAudioProfile, + AnalysisResolution, StretchMode, DeinterlaceParity, DeinterlaceMode, @@ -410,6 +413,7 @@ 'Codec', 'Audio', 'AacAudio', + 'FaceDetectorPreset', 'AudioAnalyzerPreset', 'Overlay', 'AudioOverlay', @@ -518,6 +522,7 @@ 'ContentKeyPolicyRestrictionTokenType', 'ContentKeyPolicyFairPlayRentalAndLeaseKeyType', 'AacAudioProfile', + 'AnalysisResolution', 'StretchMode', 'DeinterlaceParity', 'DeinterlaceMode', diff --git a/azure-mgmt-media/azure/mgmt/media/models/audio_analyzer_preset.py b/azure-mgmt-media/azure/mgmt/media/models/audio_analyzer_preset.py index 9f5f5f61defd..f821fdf9cf7d 100644 --- a/azure-mgmt-media/azure/mgmt/media/models/audio_analyzer_preset.py +++ b/azure-mgmt-media/azure/mgmt/media/models/audio_analyzer_preset.py @@ -25,15 +25,21 @@ class AudioAnalyzerPreset(Preset): :param odatatype: Required. Constant filled by server. :type odatatype: str :param audio_language: The language for the audio payload in the input - using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). The list - of supported languages are, 'en-US', 'en-GB', 'es-ES', 'es-MX', 'fr-FR', - 'it-IT', 'ja-JP', 'pt-BR', 'zh-CN', 'de-DE', 'ar-EG', 'ru-RU', 'hi-IN'. If - not specified, automatic language detection would be employed. This + using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). The list + of supported languages are English ('en-US' and 'en-GB'), Spanish ('es-ES' + and 'es-MX'), French ('fr-FR'), Italian ('it-IT'), Japanese ('ja-JP'), + Portuguese ('pt-BR'), Chinese ('zh-CN'), German ('de-DE'), Arabic + ('ar-EG'), Russian ('ru-RU'), Hindi ('hi-IN'), and Korean ('ko-KR'). If + you know the language of your content, it is recommended that you specify + it. If the language isn't specified or set to null, automatic language + detection will choose the first language detected and process with the + selected language for the duration of the file. This language detection feature currently supports English, Chinese, French, German, Italian, - Japanese, Spanish, Russian, and Portuguese. The automatic detection works - best with audio recordings with clearly discernable speech. If automatic - detection fails to find the language, transcription would fallback to - English. + Japanese, Spanish, Russian, and Portuguese. It does not currently support + dynamically switching between languages after the first language is + detected. The automatic detection works best with audio recordings with + clearly discernable speech. If automatic detection fails to find the + language, transcription would fallback to 'en-US'." :type audio_language: str """ diff --git a/azure-mgmt-media/azure/mgmt/media/models/audio_analyzer_preset_py3.py b/azure-mgmt-media/azure/mgmt/media/models/audio_analyzer_preset_py3.py index fb0c4f426094..9e0a0b31a3ae 100644 --- a/azure-mgmt-media/azure/mgmt/media/models/audio_analyzer_preset_py3.py +++ b/azure-mgmt-media/azure/mgmt/media/models/audio_analyzer_preset_py3.py @@ -25,15 +25,21 @@ class AudioAnalyzerPreset(Preset): :param odatatype: Required. Constant filled by server. :type odatatype: str :param audio_language: The language for the audio payload in the input - using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). The list - of supported languages are, 'en-US', 'en-GB', 'es-ES', 'es-MX', 'fr-FR', - 'it-IT', 'ja-JP', 'pt-BR', 'zh-CN', 'de-DE', 'ar-EG', 'ru-RU', 'hi-IN'. If - not specified, automatic language detection would be employed. This + using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). The list + of supported languages are English ('en-US' and 'en-GB'), Spanish ('es-ES' + and 'es-MX'), French ('fr-FR'), Italian ('it-IT'), Japanese ('ja-JP'), + Portuguese ('pt-BR'), Chinese ('zh-CN'), German ('de-DE'), Arabic + ('ar-EG'), Russian ('ru-RU'), Hindi ('hi-IN'), and Korean ('ko-KR'). If + you know the language of your content, it is recommended that you specify + it. If the language isn't specified or set to null, automatic language + detection will choose the first language detected and process with the + selected language for the duration of the file. This language detection feature currently supports English, Chinese, French, German, Italian, - Japanese, Spanish, Russian, and Portuguese. The automatic detection works - best with audio recordings with clearly discernable speech. If automatic - detection fails to find the language, transcription would fallback to - English. + Japanese, Spanish, Russian, and Portuguese. It does not currently support + dynamically switching between languages after the first language is + detected. The automatic detection works best with audio recordings with + clearly discernable speech. If automatic detection fails to find the + language, transcription would fallback to 'en-US'." :type audio_language: str """ diff --git a/azure-mgmt-media/azure/mgmt/media/models/audio_overlay.py b/azure-mgmt-media/azure/mgmt/media/models/audio_overlay.py index a2129cfcedec..9650f363101a 100644 --- a/azure-mgmt-media/azure/mgmt/media/models/audio_overlay.py +++ b/azure-mgmt-media/azure/mgmt/media/models/audio_overlay.py @@ -17,11 +17,11 @@ class AudioOverlay(Overlay): All required parameters must be populated in order to send to Azure. - :param input_label: The label of the job input which is to be used as an - overlay. The Input must specify exactly one file. You can specify an image - file in JPG or PNG formats, or an audio file (such as a WAV, MP3, WMA or - M4A file), or a video file. See https://aka.ms/mesformats for the complete - list of supported audio and video file formats. + :param input_label: Required. The label of the job input which is to be + used as an overlay. The Input must specify exactly one file. You can + specify an image file in JPG or PNG formats, or an audio file (such as a + WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats + for the complete list of supported audio and video file formats. :type input_label: str :param start: The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For @@ -51,6 +51,7 @@ class AudioOverlay(Overlay): """ _validation = { + 'input_label': {'required': True}, 'odatatype': {'required': True}, } diff --git a/azure-mgmt-media/azure/mgmt/media/models/audio_overlay_py3.py b/azure-mgmt-media/azure/mgmt/media/models/audio_overlay_py3.py index b8bf59476420..f00a28492e71 100644 --- a/azure-mgmt-media/azure/mgmt/media/models/audio_overlay_py3.py +++ b/azure-mgmt-media/azure/mgmt/media/models/audio_overlay_py3.py @@ -17,11 +17,11 @@ class AudioOverlay(Overlay): All required parameters must be populated in order to send to Azure. - :param input_label: The label of the job input which is to be used as an - overlay. The Input must specify exactly one file. You can specify an image - file in JPG or PNG formats, or an audio file (such as a WAV, MP3, WMA or - M4A file), or a video file. See https://aka.ms/mesformats for the complete - list of supported audio and video file formats. + :param input_label: Required. The label of the job input which is to be + used as an overlay. The Input must specify exactly one file. You can + specify an image file in JPG or PNG formats, or an audio file (such as a + WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats + for the complete list of supported audio and video file formats. :type input_label: str :param start: The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For @@ -51,6 +51,7 @@ class AudioOverlay(Overlay): """ _validation = { + 'input_label': {'required': True}, 'odatatype': {'required': True}, } @@ -64,6 +65,6 @@ class AudioOverlay(Overlay): 'odatatype': {'key': '@odata\\.type', 'type': 'str'}, } - def __init__(self, *, input_label: str=None, start=None, end=None, fade_in_duration=None, fade_out_duration=None, audio_gain_level: float=None, **kwargs) -> None: + def __init__(self, *, input_label: str, start=None, end=None, fade_in_duration=None, fade_out_duration=None, audio_gain_level: float=None, **kwargs) -> None: super(AudioOverlay, self).__init__(input_label=input_label, start=start, end=end, fade_in_duration=fade_in_duration, fade_out_duration=fade_out_duration, audio_gain_level=audio_gain_level, **kwargs) self.odatatype = '#Microsoft.Media.AudioOverlay' diff --git a/azure-mgmt-media/azure/mgmt/media/models/azure_media_services_enums.py b/azure-mgmt-media/azure/mgmt/media/models/azure_media_services_enums.py index 9a9f9b048077..45f3ef8f0531 100644 --- a/azure-mgmt-media/azure/mgmt/media/models/azure_media_services_enums.py +++ b/azure-mgmt-media/azure/mgmt/media/models/azure_media_services_enums.py @@ -106,6 +106,12 @@ class AacAudioProfile(str, Enum): he_aac_v2 = "HeAacV2" #: Specifies that the output audio is to be encoded into HE-AAC v2 profile. +class AnalysisResolution(str, Enum): + + source_resolution = "SourceResolution" + standard_definition = "StandardDefinition" + + class StretchMode(str, Enum): none = "None" #: Strictly respect the output resolution without considering the pixel aspect ratio or display aspect ratio of the input video. diff --git a/azure-mgmt-media/azure/mgmt/media/models/envelope_encryption.py b/azure-mgmt-media/azure/mgmt/media/models/envelope_encryption.py index 2c7ee01a26a0..be1db0b91753 100644 --- a/azure-mgmt-media/azure/mgmt/media/models/envelope_encryption.py +++ b/azure-mgmt-media/azure/mgmt/media/models/envelope_encryption.py @@ -22,8 +22,14 @@ class EnvelopeEncryption(Model): :param content_keys: Representing default content key for each encryption scheme and separate content keys for specific tracks :type content_keys: ~azure.mgmt.media.models.StreamingPolicyContentKeys - :param custom_key_acquisition_url_template: KeyAcquisitionUrlTemplate is - used to point to user specified service to delivery content keys + :param custom_key_acquisition_url_template: Template for the URL of the + custom service delivering keys to end user players. Not required when + using Azure Media Services for issuing keys. The template supports + replaceable tokens that the service will update at runtime with the value + specific to the request. The currently supported token values are + {AlternativeMediaId}, which is replaced with the value of + StreamingLocatorId.AlternativeMediaId, and {ContentKeyId}, which is + replaced with the value of identifier of the key being requested. :type custom_key_acquisition_url_template: str """ diff --git a/azure-mgmt-media/azure/mgmt/media/models/envelope_encryption_py3.py b/azure-mgmt-media/azure/mgmt/media/models/envelope_encryption_py3.py index 5b073a574fb4..250c50f00915 100644 --- a/azure-mgmt-media/azure/mgmt/media/models/envelope_encryption_py3.py +++ b/azure-mgmt-media/azure/mgmt/media/models/envelope_encryption_py3.py @@ -22,8 +22,14 @@ class EnvelopeEncryption(Model): :param content_keys: Representing default content key for each encryption scheme and separate content keys for specific tracks :type content_keys: ~azure.mgmt.media.models.StreamingPolicyContentKeys - :param custom_key_acquisition_url_template: KeyAcquisitionUrlTemplate is - used to point to user specified service to delivery content keys + :param custom_key_acquisition_url_template: Template for the URL of the + custom service delivering keys to end user players. Not required when + using Azure Media Services for issuing keys. The template supports + replaceable tokens that the service will update at runtime with the value + specific to the request. The currently supported token values are + {AlternativeMediaId}, which is replaced with the value of + StreamingLocatorId.AlternativeMediaId, and {ContentKeyId}, which is + replaced with the value of identifier of the key being requested. :type custom_key_acquisition_url_template: str """ diff --git a/azure-mgmt-media/azure/mgmt/media/models/face_detector_preset.py b/azure-mgmt-media/azure/mgmt/media/models/face_detector_preset.py new file mode 100644 index 000000000000..b834740de5a2 --- /dev/null +++ b/azure-mgmt-media/azure/mgmt/media/models/face_detector_preset.py @@ -0,0 +1,51 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .preset import Preset + + +class FaceDetectorPreset(Preset): + """Describes all the settings to be used when analyzing a video in order to + detect all the faces present. + + All required parameters must be populated in order to send to Azure. + + :param odatatype: Required. Constant filled by server. + :type odatatype: str + :param resolution: Specifies the maximum resolution at which your video is + analyzed. The default behavior is "SourceResolution," which will keep the + input video at its original resolution when analyzed. Using + "StandardDefinition" will resize input videos to standard definition while + preserving the appropriate aspect ratio. It will only resize if the video + is of higher resolution. For example, a 1920x1080 input would be scaled to + 640x360 before processing. Switching to "StandardDefinition" will reduce + the time it takes to process high resolution video. It may also reduce the + cost of using this component (see + https://azure.microsoft.com/en-us/pricing/details/media-services/#analytics + for details). However, faces that end up being too small in the resized + video may not be detected. Possible values include: 'SourceResolution', + 'StandardDefinition' + :type resolution: str or ~azure.mgmt.media.models.AnalysisResolution + """ + + _validation = { + 'odatatype': {'required': True}, + } + + _attribute_map = { + 'odatatype': {'key': '@odata\\.type', 'type': 'str'}, + 'resolution': {'key': 'resolution', 'type': 'str'}, + } + + def __init__(self, **kwargs): + super(FaceDetectorPreset, self).__init__(**kwargs) + self.resolution = kwargs.get('resolution', None) + self.odatatype = '#Microsoft.Media.FaceDetectorPreset' diff --git a/azure-mgmt-media/azure/mgmt/media/models/face_detector_preset_py3.py b/azure-mgmt-media/azure/mgmt/media/models/face_detector_preset_py3.py new file mode 100644 index 000000000000..58f33a47dbac --- /dev/null +++ b/azure-mgmt-media/azure/mgmt/media/models/face_detector_preset_py3.py @@ -0,0 +1,51 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .preset_py3 import Preset + + +class FaceDetectorPreset(Preset): + """Describes all the settings to be used when analyzing a video in order to + detect all the faces present. + + All required parameters must be populated in order to send to Azure. + + :param odatatype: Required. Constant filled by server. + :type odatatype: str + :param resolution: Specifies the maximum resolution at which your video is + analyzed. The default behavior is "SourceResolution," which will keep the + input video at its original resolution when analyzed. Using + "StandardDefinition" will resize input videos to standard definition while + preserving the appropriate aspect ratio. It will only resize if the video + is of higher resolution. For example, a 1920x1080 input would be scaled to + 640x360 before processing. Switching to "StandardDefinition" will reduce + the time it takes to process high resolution video. It may also reduce the + cost of using this component (see + https://azure.microsoft.com/en-us/pricing/details/media-services/#analytics + for details). However, faces that end up being too small in the resized + video may not be detected. Possible values include: 'SourceResolution', + 'StandardDefinition' + :type resolution: str or ~azure.mgmt.media.models.AnalysisResolution + """ + + _validation = { + 'odatatype': {'required': True}, + } + + _attribute_map = { + 'odatatype': {'key': '@odata\\.type', 'type': 'str'}, + 'resolution': {'key': 'resolution', 'type': 'str'}, + } + + def __init__(self, *, resolution=None, **kwargs) -> None: + super(FaceDetectorPreset, self).__init__(**kwargs) + self.resolution = resolution + self.odatatype = '#Microsoft.Media.FaceDetectorPreset' diff --git a/azure-mgmt-media/azure/mgmt/media/models/format.py b/azure-mgmt-media/azure/mgmt/media/models/format.py index 28fc658c74e9..334c7dd3103f 100644 --- a/azure-mgmt-media/azure/mgmt/media/models/format.py +++ b/azure-mgmt-media/azure/mgmt/media/models/format.py @@ -20,9 +20,9 @@ class Format(Model): All required parameters must be populated in order to send to Azure. - :param filename_pattern: The pattern of the file names for the generated - output files. The following macros are supported in the file name: - {Basename} - The base name of the input video {Extension} - The + :param filename_pattern: Required. The pattern of the file names for the + generated output files. The following macros are supported in the file + name: {Basename} - The base name of the input video {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to @@ -34,6 +34,7 @@ class Format(Model): """ _validation = { + 'filename_pattern': {'required': True}, 'odatatype': {'required': True}, } diff --git a/azure-mgmt-media/azure/mgmt/media/models/format_py3.py b/azure-mgmt-media/azure/mgmt/media/models/format_py3.py index 6a56ef9d29f2..62b347df9b8c 100644 --- a/azure-mgmt-media/azure/mgmt/media/models/format_py3.py +++ b/azure-mgmt-media/azure/mgmt/media/models/format_py3.py @@ -20,9 +20,9 @@ class Format(Model): All required parameters must be populated in order to send to Azure. - :param filename_pattern: The pattern of the file names for the generated - output files. The following macros are supported in the file name: - {Basename} - The base name of the input video {Extension} - The + :param filename_pattern: Required. The pattern of the file names for the + generated output files. The following macros are supported in the file + name: {Basename} - The base name of the input video {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to @@ -34,6 +34,7 @@ class Format(Model): """ _validation = { + 'filename_pattern': {'required': True}, 'odatatype': {'required': True}, } @@ -46,7 +47,7 @@ class Format(Model): 'odatatype': {'#Microsoft.Media.ImageFormat': 'ImageFormat', '#Microsoft.Media.MultiBitrateFormat': 'MultiBitrateFormat'} } - def __init__(self, *, filename_pattern: str=None, **kwargs) -> None: + def __init__(self, *, filename_pattern: str, **kwargs) -> None: super(Format, self).__init__(**kwargs) self.filename_pattern = filename_pattern self.odatatype = None diff --git a/azure-mgmt-media/azure/mgmt/media/models/h264_layer.py b/azure-mgmt-media/azure/mgmt/media/models/h264_layer.py index 0afea03c378f..1ad3b4ea1c73 100644 --- a/azure-mgmt-media/azure/mgmt/media/models/h264_layer.py +++ b/azure-mgmt-media/azure/mgmt/media/models/h264_layer.py @@ -32,8 +32,9 @@ class H264Layer(VideoLayer): :type label: str :param odatatype: Required. Constant filled by server. :type odatatype: str - :param bitrate: The average bitrate in bits per second at which to encode - the input video when generating this layer. This is a required field. + :param bitrate: Required. The average bitrate in bits per second at which + to encode the input video when generating this layer. This is a required + field. :type bitrate: int :param max_bitrate: The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the @@ -83,6 +84,7 @@ class H264Layer(VideoLayer): _validation = { 'odatatype': {'required': True}, + 'bitrate': {'required': True}, } _attribute_map = { diff --git a/azure-mgmt-media/azure/mgmt/media/models/h264_layer_py3.py b/azure-mgmt-media/azure/mgmt/media/models/h264_layer_py3.py index 2e32b8746148..d03148a3352d 100644 --- a/azure-mgmt-media/azure/mgmt/media/models/h264_layer_py3.py +++ b/azure-mgmt-media/azure/mgmt/media/models/h264_layer_py3.py @@ -32,8 +32,9 @@ class H264Layer(VideoLayer): :type label: str :param odatatype: Required. Constant filled by server. :type odatatype: str - :param bitrate: The average bitrate in bits per second at which to encode - the input video when generating this layer. This is a required field. + :param bitrate: Required. The average bitrate in bits per second at which + to encode the input video when generating this layer. This is a required + field. :type bitrate: int :param max_bitrate: The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the @@ -83,6 +84,7 @@ class H264Layer(VideoLayer): _validation = { 'odatatype': {'required': True}, + 'bitrate': {'required': True}, } _attribute_map = { @@ -103,7 +105,7 @@ class H264Layer(VideoLayer): 'entropy_mode': {'key': 'entropyMode', 'type': 'str'}, } - def __init__(self, *, width: str=None, height: str=None, label: str=None, bitrate: int=None, max_bitrate: int=None, b_frames: int=None, frame_rate: str=None, slices: int=None, adaptive_bframe: bool=None, profile=None, level: str=None, buffer_window=None, reference_frames: int=None, entropy_mode=None, **kwargs) -> None: + def __init__(self, *, bitrate: int, width: str=None, height: str=None, label: str=None, max_bitrate: int=None, b_frames: int=None, frame_rate: str=None, slices: int=None, adaptive_bframe: bool=None, profile=None, level: str=None, buffer_window=None, reference_frames: int=None, entropy_mode=None, **kwargs) -> None: super(H264Layer, self).__init__(width=width, height=height, label=label, bitrate=bitrate, max_bitrate=max_bitrate, b_frames=b_frames, frame_rate=frame_rate, slices=slices, adaptive_bframe=adaptive_bframe, **kwargs) self.profile = profile self.level = level diff --git a/azure-mgmt-media/azure/mgmt/media/models/image.py b/azure-mgmt-media/azure/mgmt/media/models/image.py index afb0fb501a5c..494ef1711066 100644 --- a/azure-mgmt-media/azure/mgmt/media/models/image.py +++ b/azure-mgmt-media/azure/mgmt/media/models/image.py @@ -35,10 +35,10 @@ class Image(Video): resized to fit the desired output resolution(s). Default is AutoSize. Possible values include: 'None', 'AutoSize', 'AutoFit' :type stretch_mode: str or ~azure.mgmt.media.models.StretchMode - :param start: The position in the input video from where to start - generating thumbnails. The value can be in absolute timestamp (ISO 8601, - e.g: PT05S), or a frame count (For example, 10 for the 10th frame), or a - relative value (For example, 1%). Also supports a macro {Best}, which + :param start: Required. The position in the input video from where to + start generating thumbnails. The value can be in absolute timestamp (ISO + 8601, e.g: PT05S), or a frame count (For example, 10 for the 10th frame), + or a relative value (For example, 1%). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video. :type start: str @@ -56,6 +56,7 @@ class Image(Video): _validation = { 'odatatype': {'required': True}, + 'start': {'required': True}, } _attribute_map = { diff --git a/azure-mgmt-media/azure/mgmt/media/models/image_format.py b/azure-mgmt-media/azure/mgmt/media/models/image_format.py index 4c705f65825a..9114e08860ff 100644 --- a/azure-mgmt-media/azure/mgmt/media/models/image_format.py +++ b/azure-mgmt-media/azure/mgmt/media/models/image_format.py @@ -20,9 +20,9 @@ class ImageFormat(Format): All required parameters must be populated in order to send to Azure. - :param filename_pattern: The pattern of the file names for the generated - output files. The following macros are supported in the file name: - {Basename} - The base name of the input video {Extension} - The + :param filename_pattern: Required. The pattern of the file names for the + generated output files. The following macros are supported in the file + name: {Basename} - The base name of the input video {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to @@ -34,6 +34,7 @@ class ImageFormat(Format): """ _validation = { + 'filename_pattern': {'required': True}, 'odatatype': {'required': True}, } diff --git a/azure-mgmt-media/azure/mgmt/media/models/image_format_py3.py b/azure-mgmt-media/azure/mgmt/media/models/image_format_py3.py index 7095adc370dd..b9cf38ed833c 100644 --- a/azure-mgmt-media/azure/mgmt/media/models/image_format_py3.py +++ b/azure-mgmt-media/azure/mgmt/media/models/image_format_py3.py @@ -20,9 +20,9 @@ class ImageFormat(Format): All required parameters must be populated in order to send to Azure. - :param filename_pattern: The pattern of the file names for the generated - output files. The following macros are supported in the file name: - {Basename} - The base name of the input video {Extension} - The + :param filename_pattern: Required. The pattern of the file names for the + generated output files. The following macros are supported in the file + name: {Basename} - The base name of the input video {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to @@ -34,6 +34,7 @@ class ImageFormat(Format): """ _validation = { + 'filename_pattern': {'required': True}, 'odatatype': {'required': True}, } @@ -46,6 +47,6 @@ class ImageFormat(Format): 'odatatype': {'#Microsoft.Media.JpgFormat': 'JpgFormat', '#Microsoft.Media.PngFormat': 'PngFormat'} } - def __init__(self, *, filename_pattern: str=None, **kwargs) -> None: + def __init__(self, *, filename_pattern: str, **kwargs) -> None: super(ImageFormat, self).__init__(filename_pattern=filename_pattern, **kwargs) self.odatatype = '#Microsoft.Media.ImageFormat' diff --git a/azure-mgmt-media/azure/mgmt/media/models/image_py3.py b/azure-mgmt-media/azure/mgmt/media/models/image_py3.py index 974dd6c2d213..86467ce40ac3 100644 --- a/azure-mgmt-media/azure/mgmt/media/models/image_py3.py +++ b/azure-mgmt-media/azure/mgmt/media/models/image_py3.py @@ -35,10 +35,10 @@ class Image(Video): resized to fit the desired output resolution(s). Default is AutoSize. Possible values include: 'None', 'AutoSize', 'AutoFit' :type stretch_mode: str or ~azure.mgmt.media.models.StretchMode - :param start: The position in the input video from where to start - generating thumbnails. The value can be in absolute timestamp (ISO 8601, - e.g: PT05S), or a frame count (For example, 10 for the 10th frame), or a - relative value (For example, 1%). Also supports a macro {Best}, which + :param start: Required. The position in the input video from where to + start generating thumbnails. The value can be in absolute timestamp (ISO + 8601, e.g: PT05S), or a frame count (For example, 10 for the 10th frame), + or a relative value (For example, 1%). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video. :type start: str @@ -56,6 +56,7 @@ class Image(Video): _validation = { 'odatatype': {'required': True}, + 'start': {'required': True}, } _attribute_map = { @@ -72,7 +73,7 @@ class Image(Video): 'odatatype': {'#Microsoft.Media.JpgImage': 'JpgImage', '#Microsoft.Media.PngImage': 'PngImage'} } - def __init__(self, *, label: str=None, key_frame_interval=None, stretch_mode=None, start: str=None, step: str=None, range: str=None, **kwargs) -> None: + def __init__(self, *, start: str, label: str=None, key_frame_interval=None, stretch_mode=None, step: str=None, range: str=None, **kwargs) -> None: super(Image, self).__init__(label=label, key_frame_interval=key_frame_interval, stretch_mode=stretch_mode, **kwargs) self.start = start self.step = step diff --git a/azure-mgmt-media/azure/mgmt/media/models/job.py b/azure-mgmt-media/azure/mgmt/media/models/job.py index 47495b54fe70..9a761401192a 100644 --- a/azure-mgmt-media/azure/mgmt/media/models/job.py +++ b/azure-mgmt-media/azure/mgmt/media/models/job.py @@ -47,7 +47,7 @@ class Job(ProxyResource): priority jobs are processed before lower priority jobs. If not set, the default is normal. Possible values include: 'Low', 'Normal', 'High' :type priority: str or ~azure.mgmt.media.models.Priority - :param correlation_data: Customer provided correlation data that will be + :param correlation_data: Customer provided key, value pairs that will be returned in Job and JobOutput state events. :type correlation_data: dict[str, str] """ diff --git a/azure-mgmt-media/azure/mgmt/media/models/job_input_asset.py b/azure-mgmt-media/azure/mgmt/media/models/job_input_asset.py index db205ff0064e..687100fc2c90 100644 --- a/azure-mgmt-media/azure/mgmt/media/models/job_input_asset.py +++ b/azure-mgmt-media/azure/mgmt/media/models/job_input_asset.py @@ -19,7 +19,8 @@ class JobInputAsset(JobInputClip): :param odatatype: Required. Constant filled by server. :type odatatype: str - :param files: List of files. Required for JobInputHttp. + :param files: List of files. Required for JobInputHttp. Maximum of 4000 + characters each. :type files: list[str] :param label: A label that is assigned to a JobInputClip, that is used to satisfy a reference used in the Transform. For example, a Transform can be diff --git a/azure-mgmt-media/azure/mgmt/media/models/job_input_asset_py3.py b/azure-mgmt-media/azure/mgmt/media/models/job_input_asset_py3.py index ea95bb8d8ced..263d4ac593c8 100644 --- a/azure-mgmt-media/azure/mgmt/media/models/job_input_asset_py3.py +++ b/azure-mgmt-media/azure/mgmt/media/models/job_input_asset_py3.py @@ -19,7 +19,8 @@ class JobInputAsset(JobInputClip): :param odatatype: Required. Constant filled by server. :type odatatype: str - :param files: List of files. Required for JobInputHttp. + :param files: List of files. Required for JobInputHttp. Maximum of 4000 + characters each. :type files: list[str] :param label: A label that is assigned to a JobInputClip, that is used to satisfy a reference used in the Transform. For example, a Transform can be diff --git a/azure-mgmt-media/azure/mgmt/media/models/job_input_clip.py b/azure-mgmt-media/azure/mgmt/media/models/job_input_clip.py index 5c36848da19b..5187d14e6234 100644 --- a/azure-mgmt-media/azure/mgmt/media/models/job_input_clip.py +++ b/azure-mgmt-media/azure/mgmt/media/models/job_input_clip.py @@ -22,7 +22,8 @@ class JobInputClip(JobInput): :param odatatype: Required. Constant filled by server. :type odatatype: str - :param files: List of files. Required for JobInputHttp. + :param files: List of files. Required for JobInputHttp. Maximum of 4000 + characters each. :type files: list[str] :param label: A label that is assigned to a JobInputClip, that is used to satisfy a reference used in the Transform. For example, a Transform can be diff --git a/azure-mgmt-media/azure/mgmt/media/models/job_input_clip_py3.py b/azure-mgmt-media/azure/mgmt/media/models/job_input_clip_py3.py index 18af26239f74..b6bbdb3e4e1e 100644 --- a/azure-mgmt-media/azure/mgmt/media/models/job_input_clip_py3.py +++ b/azure-mgmt-media/azure/mgmt/media/models/job_input_clip_py3.py @@ -22,7 +22,8 @@ class JobInputClip(JobInput): :param odatatype: Required. Constant filled by server. :type odatatype: str - :param files: List of files. Required for JobInputHttp. + :param files: List of files. Required for JobInputHttp. Maximum of 4000 + characters each. :type files: list[str] :param label: A label that is assigned to a JobInputClip, that is used to satisfy a reference used in the Transform. For example, a Transform can be diff --git a/azure-mgmt-media/azure/mgmt/media/models/job_input_http.py b/azure-mgmt-media/azure/mgmt/media/models/job_input_http.py index 3c739e4ee848..a12e24dec5a1 100644 --- a/azure-mgmt-media/azure/mgmt/media/models/job_input_http.py +++ b/azure-mgmt-media/azure/mgmt/media/models/job_input_http.py @@ -19,7 +19,8 @@ class JobInputHttp(JobInputClip): :param odatatype: Required. Constant filled by server. :type odatatype: str - :param files: List of files. Required for JobInputHttp. + :param files: List of files. Required for JobInputHttp. Maximum of 4000 + characters each. :type files: list[str] :param label: A label that is assigned to a JobInputClip, that is used to satisfy a reference used in the Transform. For example, a Transform can be @@ -29,8 +30,9 @@ class JobInputHttp(JobInputClip): have the label 'xyz'. :type label: str :param base_uri: Base URI for HTTPS job input. It will be concatenated - with provided file names. If no base uri is given, then the provided - file list is assumed to be fully qualified uris. + with provided file names. If no base uri is given, then the provided file + list is assumed to be fully qualified uris. Maximum length of 4000 + characters. :type base_uri: str """ diff --git a/azure-mgmt-media/azure/mgmt/media/models/job_input_http_py3.py b/azure-mgmt-media/azure/mgmt/media/models/job_input_http_py3.py index 55436af89760..0cc04900ba5e 100644 --- a/azure-mgmt-media/azure/mgmt/media/models/job_input_http_py3.py +++ b/azure-mgmt-media/azure/mgmt/media/models/job_input_http_py3.py @@ -19,7 +19,8 @@ class JobInputHttp(JobInputClip): :param odatatype: Required. Constant filled by server. :type odatatype: str - :param files: List of files. Required for JobInputHttp. + :param files: List of files. Required for JobInputHttp. Maximum of 4000 + characters each. :type files: list[str] :param label: A label that is assigned to a JobInputClip, that is used to satisfy a reference used in the Transform. For example, a Transform can be @@ -29,8 +30,9 @@ class JobInputHttp(JobInputClip): have the label 'xyz'. :type label: str :param base_uri: Base URI for HTTPS job input. It will be concatenated - with provided file names. If no base uri is given, then the provided - file list is assumed to be fully qualified uris. + with provided file names. If no base uri is given, then the provided file + list is assumed to be fully qualified uris. Maximum length of 4000 + characters. :type base_uri: str """ diff --git a/azure-mgmt-media/azure/mgmt/media/models/job_py3.py b/azure-mgmt-media/azure/mgmt/media/models/job_py3.py index 9890c618e1c0..6f6955010d60 100644 --- a/azure-mgmt-media/azure/mgmt/media/models/job_py3.py +++ b/azure-mgmt-media/azure/mgmt/media/models/job_py3.py @@ -47,7 +47,7 @@ class Job(ProxyResource): priority jobs are processed before lower priority jobs. If not set, the default is normal. Possible values include: 'Low', 'Normal', 'High' :type priority: str or ~azure.mgmt.media.models.Priority - :param correlation_data: Customer provided correlation data that will be + :param correlation_data: Customer provided key, value pairs that will be returned in Job and JobOutput state events. :type correlation_data: dict[str, str] """ diff --git a/azure-mgmt-media/azure/mgmt/media/models/jpg_format.py b/azure-mgmt-media/azure/mgmt/media/models/jpg_format.py index a8bd269ae80f..5ac2f5dd8275 100644 --- a/azure-mgmt-media/azure/mgmt/media/models/jpg_format.py +++ b/azure-mgmt-media/azure/mgmt/media/models/jpg_format.py @@ -17,9 +17,9 @@ class JpgFormat(ImageFormat): All required parameters must be populated in order to send to Azure. - :param filename_pattern: The pattern of the file names for the generated - output files. The following macros are supported in the file name: - {Basename} - The base name of the input video {Extension} - The + :param filename_pattern: Required. The pattern of the file names for the + generated output files. The following macros are supported in the file + name: {Basename} - The base name of the input video {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to @@ -31,6 +31,7 @@ class JpgFormat(ImageFormat): """ _validation = { + 'filename_pattern': {'required': True}, 'odatatype': {'required': True}, } diff --git a/azure-mgmt-media/azure/mgmt/media/models/jpg_format_py3.py b/azure-mgmt-media/azure/mgmt/media/models/jpg_format_py3.py index a8384445c006..e15163904bfd 100644 --- a/azure-mgmt-media/azure/mgmt/media/models/jpg_format_py3.py +++ b/azure-mgmt-media/azure/mgmt/media/models/jpg_format_py3.py @@ -17,9 +17,9 @@ class JpgFormat(ImageFormat): All required parameters must be populated in order to send to Azure. - :param filename_pattern: The pattern of the file names for the generated - output files. The following macros are supported in the file name: - {Basename} - The base name of the input video {Extension} - The + :param filename_pattern: Required. The pattern of the file names for the + generated output files. The following macros are supported in the file + name: {Basename} - The base name of the input video {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to @@ -31,6 +31,7 @@ class JpgFormat(ImageFormat): """ _validation = { + 'filename_pattern': {'required': True}, 'odatatype': {'required': True}, } @@ -39,6 +40,6 @@ class JpgFormat(ImageFormat): 'odatatype': {'key': '@odata\\.type', 'type': 'str'}, } - def __init__(self, *, filename_pattern: str=None, **kwargs) -> None: + def __init__(self, *, filename_pattern: str, **kwargs) -> None: super(JpgFormat, self).__init__(filename_pattern=filename_pattern, **kwargs) self.odatatype = '#Microsoft.Media.JpgFormat' diff --git a/azure-mgmt-media/azure/mgmt/media/models/jpg_image.py b/azure-mgmt-media/azure/mgmt/media/models/jpg_image.py index a062e8bc02aa..37cc8ad558f7 100644 --- a/azure-mgmt-media/azure/mgmt/media/models/jpg_image.py +++ b/azure-mgmt-media/azure/mgmt/media/models/jpg_image.py @@ -32,10 +32,10 @@ class JpgImage(Image): resized to fit the desired output resolution(s). Default is AutoSize. Possible values include: 'None', 'AutoSize', 'AutoFit' :type stretch_mode: str or ~azure.mgmt.media.models.StretchMode - :param start: The position in the input video from where to start - generating thumbnails. The value can be in absolute timestamp (ISO 8601, - e.g: PT05S), or a frame count (For example, 10 for the 10th frame), or a - relative value (For example, 1%). Also supports a macro {Best}, which + :param start: Required. The position in the input video from where to + start generating thumbnails. The value can be in absolute timestamp (ISO + 8601, e.g: PT05S), or a frame count (For example, 10 for the 10th frame), + or a relative value (For example, 1%). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video. :type start: str @@ -56,6 +56,7 @@ class JpgImage(Image): _validation = { 'odatatype': {'required': True}, + 'start': {'required': True}, } _attribute_map = { diff --git a/azure-mgmt-media/azure/mgmt/media/models/jpg_image_py3.py b/azure-mgmt-media/azure/mgmt/media/models/jpg_image_py3.py index 4edb4bcf516f..9f1ca28a41f7 100644 --- a/azure-mgmt-media/azure/mgmt/media/models/jpg_image_py3.py +++ b/azure-mgmt-media/azure/mgmt/media/models/jpg_image_py3.py @@ -32,10 +32,10 @@ class JpgImage(Image): resized to fit the desired output resolution(s). Default is AutoSize. Possible values include: 'None', 'AutoSize', 'AutoFit' :type stretch_mode: str or ~azure.mgmt.media.models.StretchMode - :param start: The position in the input video from where to start - generating thumbnails. The value can be in absolute timestamp (ISO 8601, - e.g: PT05S), or a frame count (For example, 10 for the 10th frame), or a - relative value (For example, 1%). Also supports a macro {Best}, which + :param start: Required. The position in the input video from where to + start generating thumbnails. The value can be in absolute timestamp (ISO + 8601, e.g: PT05S), or a frame count (For example, 10 for the 10th frame), + or a relative value (For example, 1%). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video. :type start: str @@ -56,6 +56,7 @@ class JpgImage(Image): _validation = { 'odatatype': {'required': True}, + 'start': {'required': True}, } _attribute_map = { @@ -69,7 +70,7 @@ class JpgImage(Image): 'layers': {'key': 'layers', 'type': '[JpgLayer]'}, } - def __init__(self, *, label: str=None, key_frame_interval=None, stretch_mode=None, start: str=None, step: str=None, range: str=None, layers=None, **kwargs) -> None: + def __init__(self, *, start: str, label: str=None, key_frame_interval=None, stretch_mode=None, step: str=None, range: str=None, layers=None, **kwargs) -> None: super(JpgImage, self).__init__(label=label, key_frame_interval=key_frame_interval, stretch_mode=stretch_mode, start=start, step=step, range=range, **kwargs) self.layers = layers self.odatatype = '#Microsoft.Media.JpgImage' diff --git a/azure-mgmt-media/azure/mgmt/media/models/mp4_format.py b/azure-mgmt-media/azure/mgmt/media/models/mp4_format.py index 70e74e9ba348..36cd267a07ac 100644 --- a/azure-mgmt-media/azure/mgmt/media/models/mp4_format.py +++ b/azure-mgmt-media/azure/mgmt/media/models/mp4_format.py @@ -17,9 +17,9 @@ class Mp4Format(MultiBitrateFormat): All required parameters must be populated in order to send to Azure. - :param filename_pattern: The pattern of the file names for the generated - output files. The following macros are supported in the file name: - {Basename} - The base name of the input video {Extension} - The + :param filename_pattern: Required. The pattern of the file names for the + generated output files. The following macros are supported in the file + name: {Basename} - The base name of the input video {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to @@ -34,6 +34,7 @@ class Mp4Format(MultiBitrateFormat): """ _validation = { + 'filename_pattern': {'required': True}, 'odatatype': {'required': True}, } diff --git a/azure-mgmt-media/azure/mgmt/media/models/mp4_format_py3.py b/azure-mgmt-media/azure/mgmt/media/models/mp4_format_py3.py index bd731424b19f..3e2397bce3a7 100644 --- a/azure-mgmt-media/azure/mgmt/media/models/mp4_format_py3.py +++ b/azure-mgmt-media/azure/mgmt/media/models/mp4_format_py3.py @@ -17,9 +17,9 @@ class Mp4Format(MultiBitrateFormat): All required parameters must be populated in order to send to Azure. - :param filename_pattern: The pattern of the file names for the generated - output files. The following macros are supported in the file name: - {Basename} - The base name of the input video {Extension} - The + :param filename_pattern: Required. The pattern of the file names for the + generated output files. The following macros are supported in the file + name: {Basename} - The base name of the input video {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to @@ -34,6 +34,7 @@ class Mp4Format(MultiBitrateFormat): """ _validation = { + 'filename_pattern': {'required': True}, 'odatatype': {'required': True}, } @@ -43,6 +44,6 @@ class Mp4Format(MultiBitrateFormat): 'output_files': {'key': 'outputFiles', 'type': '[OutputFile]'}, } - def __init__(self, *, filename_pattern: str=None, output_files=None, **kwargs) -> None: + def __init__(self, *, filename_pattern: str, output_files=None, **kwargs) -> None: super(Mp4Format, self).__init__(filename_pattern=filename_pattern, output_files=output_files, **kwargs) self.odatatype = '#Microsoft.Media.Mp4Format' diff --git a/azure-mgmt-media/azure/mgmt/media/models/multi_bitrate_format.py b/azure-mgmt-media/azure/mgmt/media/models/multi_bitrate_format.py index 78a7381316ec..1ca1dc7d1bc2 100644 --- a/azure-mgmt-media/azure/mgmt/media/models/multi_bitrate_format.py +++ b/azure-mgmt-media/azure/mgmt/media/models/multi_bitrate_format.py @@ -24,9 +24,9 @@ class MultiBitrateFormat(Format): All required parameters must be populated in order to send to Azure. - :param filename_pattern: The pattern of the file names for the generated - output files. The following macros are supported in the file name: - {Basename} - The base name of the input video {Extension} - The + :param filename_pattern: Required. The pattern of the file names for the + generated output files. The following macros are supported in the file + name: {Basename} - The base name of the input video {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to @@ -41,6 +41,7 @@ class MultiBitrateFormat(Format): """ _validation = { + 'filename_pattern': {'required': True}, 'odatatype': {'required': True}, } diff --git a/azure-mgmt-media/azure/mgmt/media/models/multi_bitrate_format_py3.py b/azure-mgmt-media/azure/mgmt/media/models/multi_bitrate_format_py3.py index 9e3e7534afaa..d6657f2b213d 100644 --- a/azure-mgmt-media/azure/mgmt/media/models/multi_bitrate_format_py3.py +++ b/azure-mgmt-media/azure/mgmt/media/models/multi_bitrate_format_py3.py @@ -24,9 +24,9 @@ class MultiBitrateFormat(Format): All required parameters must be populated in order to send to Azure. - :param filename_pattern: The pattern of the file names for the generated - output files. The following macros are supported in the file name: - {Basename} - The base name of the input video {Extension} - The + :param filename_pattern: Required. The pattern of the file names for the + generated output files. The following macros are supported in the file + name: {Basename} - The base name of the input video {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to @@ -41,6 +41,7 @@ class MultiBitrateFormat(Format): """ _validation = { + 'filename_pattern': {'required': True}, 'odatatype': {'required': True}, } @@ -54,7 +55,7 @@ class MultiBitrateFormat(Format): 'odatatype': {'#Microsoft.Media.Mp4Format': 'Mp4Format', '#Microsoft.Media.TransportStreamFormat': 'TransportStreamFormat'} } - def __init__(self, *, filename_pattern: str=None, output_files=None, **kwargs) -> None: + def __init__(self, *, filename_pattern: str, output_files=None, **kwargs) -> None: super(MultiBitrateFormat, self).__init__(filename_pattern=filename_pattern, **kwargs) self.output_files = output_files self.odatatype = '#Microsoft.Media.MultiBitrateFormat' diff --git a/azure-mgmt-media/azure/mgmt/media/models/output_file.py b/azure-mgmt-media/azure/mgmt/media/models/output_file.py index 291b7864da58..6d6ec390b536 100644 --- a/azure-mgmt-media/azure/mgmt/media/models/output_file.py +++ b/azure-mgmt-media/azure/mgmt/media/models/output_file.py @@ -15,15 +15,21 @@ class OutputFile(Model): """Represents an output file produced. - :param labels: The list of labels that describe how the encoder should - multiplex video and audio into an output file. For example, if the encoder - is producing two video layers with labels v1 and v2, and one audio layer - with label a1, then an array like '[v1, a1]' tells the encoder to produce - an output file with the video track represented by v1 and the audio track - represented by a1. + All required parameters must be populated in order to send to Azure. + + :param labels: Required. The list of labels that describe how the encoder + should multiplex video and audio into an output file. For example, if the + encoder is producing two video layers with labels v1 and v2, and one audio + layer with label a1, then an array like '[v1, a1]' tells the encoder to + produce an output file with the video track represented by v1 and the + audio track represented by a1. :type labels: list[str] """ + _validation = { + 'labels': {'required': True}, + } + _attribute_map = { 'labels': {'key': 'labels', 'type': '[str]'}, } diff --git a/azure-mgmt-media/azure/mgmt/media/models/output_file_py3.py b/azure-mgmt-media/azure/mgmt/media/models/output_file_py3.py index 465e0f7df536..7986361806ba 100644 --- a/azure-mgmt-media/azure/mgmt/media/models/output_file_py3.py +++ b/azure-mgmt-media/azure/mgmt/media/models/output_file_py3.py @@ -15,19 +15,25 @@ class OutputFile(Model): """Represents an output file produced. - :param labels: The list of labels that describe how the encoder should - multiplex video and audio into an output file. For example, if the encoder - is producing two video layers with labels v1 and v2, and one audio layer - with label a1, then an array like '[v1, a1]' tells the encoder to produce - an output file with the video track represented by v1 and the audio track - represented by a1. + All required parameters must be populated in order to send to Azure. + + :param labels: Required. The list of labels that describe how the encoder + should multiplex video and audio into an output file. For example, if the + encoder is producing two video layers with labels v1 and v2, and one audio + layer with label a1, then an array like '[v1, a1]' tells the encoder to + produce an output file with the video track represented by v1 and the + audio track represented by a1. :type labels: list[str] """ + _validation = { + 'labels': {'required': True}, + } + _attribute_map = { 'labels': {'key': 'labels', 'type': '[str]'}, } - def __init__(self, *, labels=None, **kwargs) -> None: + def __init__(self, *, labels, **kwargs) -> None: super(OutputFile, self).__init__(**kwargs) self.labels = labels diff --git a/azure-mgmt-media/azure/mgmt/media/models/overlay.py b/azure-mgmt-media/azure/mgmt/media/models/overlay.py index 5e97fcd6d971..734fc0fef46d 100644 --- a/azure-mgmt-media/azure/mgmt/media/models/overlay.py +++ b/azure-mgmt-media/azure/mgmt/media/models/overlay.py @@ -20,11 +20,11 @@ class Overlay(Model): All required parameters must be populated in order to send to Azure. - :param input_label: The label of the job input which is to be used as an - overlay. The Input must specify exactly one file. You can specify an image - file in JPG or PNG formats, or an audio file (such as a WAV, MP3, WMA or - M4A file), or a video file. See https://aka.ms/mesformats for the complete - list of supported audio and video file formats. + :param input_label: Required. The label of the job input which is to be + used as an overlay. The Input must specify exactly one file. You can + specify an image file in JPG or PNG formats, or an audio file (such as a + WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats + for the complete list of supported audio and video file formats. :type input_label: str :param start: The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For @@ -54,6 +54,7 @@ class Overlay(Model): """ _validation = { + 'input_label': {'required': True}, 'odatatype': {'required': True}, } diff --git a/azure-mgmt-media/azure/mgmt/media/models/overlay_py3.py b/azure-mgmt-media/azure/mgmt/media/models/overlay_py3.py index b8150cda97cb..adce5be8b52a 100644 --- a/azure-mgmt-media/azure/mgmt/media/models/overlay_py3.py +++ b/azure-mgmt-media/azure/mgmt/media/models/overlay_py3.py @@ -20,11 +20,11 @@ class Overlay(Model): All required parameters must be populated in order to send to Azure. - :param input_label: The label of the job input which is to be used as an - overlay. The Input must specify exactly one file. You can specify an image - file in JPG or PNG formats, or an audio file (such as a WAV, MP3, WMA or - M4A file), or a video file. See https://aka.ms/mesformats for the complete - list of supported audio and video file formats. + :param input_label: Required. The label of the job input which is to be + used as an overlay. The Input must specify exactly one file. You can + specify an image file in JPG or PNG formats, or an audio file (such as a + WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats + for the complete list of supported audio and video file formats. :type input_label: str :param start: The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For @@ -54,6 +54,7 @@ class Overlay(Model): """ _validation = { + 'input_label': {'required': True}, 'odatatype': {'required': True}, } @@ -71,7 +72,7 @@ class Overlay(Model): 'odatatype': {'#Microsoft.Media.AudioOverlay': 'AudioOverlay', '#Microsoft.Media.VideoOverlay': 'VideoOverlay'} } - def __init__(self, *, input_label: str=None, start=None, end=None, fade_in_duration=None, fade_out_duration=None, audio_gain_level: float=None, **kwargs) -> None: + def __init__(self, *, input_label: str, start=None, end=None, fade_in_duration=None, fade_out_duration=None, audio_gain_level: float=None, **kwargs) -> None: super(Overlay, self).__init__(**kwargs) self.input_label = input_label self.start = start diff --git a/azure-mgmt-media/azure/mgmt/media/models/png_format.py b/azure-mgmt-media/azure/mgmt/media/models/png_format.py index 1b1ab9096fae..8e242e416dd2 100644 --- a/azure-mgmt-media/azure/mgmt/media/models/png_format.py +++ b/azure-mgmt-media/azure/mgmt/media/models/png_format.py @@ -17,9 +17,9 @@ class PngFormat(ImageFormat): All required parameters must be populated in order to send to Azure. - :param filename_pattern: The pattern of the file names for the generated - output files. The following macros are supported in the file name: - {Basename} - The base name of the input video {Extension} - The + :param filename_pattern: Required. The pattern of the file names for the + generated output files. The following macros are supported in the file + name: {Basename} - The base name of the input video {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to @@ -31,6 +31,7 @@ class PngFormat(ImageFormat): """ _validation = { + 'filename_pattern': {'required': True}, 'odatatype': {'required': True}, } diff --git a/azure-mgmt-media/azure/mgmt/media/models/png_format_py3.py b/azure-mgmt-media/azure/mgmt/media/models/png_format_py3.py index ad06a3511b03..1cbe8b042b41 100644 --- a/azure-mgmt-media/azure/mgmt/media/models/png_format_py3.py +++ b/azure-mgmt-media/azure/mgmt/media/models/png_format_py3.py @@ -17,9 +17,9 @@ class PngFormat(ImageFormat): All required parameters must be populated in order to send to Azure. - :param filename_pattern: The pattern of the file names for the generated - output files. The following macros are supported in the file name: - {Basename} - The base name of the input video {Extension} - The + :param filename_pattern: Required. The pattern of the file names for the + generated output files. The following macros are supported in the file + name: {Basename} - The base name of the input video {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to @@ -31,6 +31,7 @@ class PngFormat(ImageFormat): """ _validation = { + 'filename_pattern': {'required': True}, 'odatatype': {'required': True}, } @@ -39,6 +40,6 @@ class PngFormat(ImageFormat): 'odatatype': {'key': '@odata\\.type', 'type': 'str'}, } - def __init__(self, *, filename_pattern: str=None, **kwargs) -> None: + def __init__(self, *, filename_pattern: str, **kwargs) -> None: super(PngFormat, self).__init__(filename_pattern=filename_pattern, **kwargs) self.odatatype = '#Microsoft.Media.PngFormat' diff --git a/azure-mgmt-media/azure/mgmt/media/models/png_image.py b/azure-mgmt-media/azure/mgmt/media/models/png_image.py index 6219d6bd3858..3f1c90f83fe5 100644 --- a/azure-mgmt-media/azure/mgmt/media/models/png_image.py +++ b/azure-mgmt-media/azure/mgmt/media/models/png_image.py @@ -32,10 +32,10 @@ class PngImage(Image): resized to fit the desired output resolution(s). Default is AutoSize. Possible values include: 'None', 'AutoSize', 'AutoFit' :type stretch_mode: str or ~azure.mgmt.media.models.StretchMode - :param start: The position in the input video from where to start - generating thumbnails. The value can be in absolute timestamp (ISO 8601, - e.g: PT05S), or a frame count (For example, 10 for the 10th frame), or a - relative value (For example, 1%). Also supports a macro {Best}, which + :param start: Required. The position in the input video from where to + start generating thumbnails. The value can be in absolute timestamp (ISO + 8601, e.g: PT05S), or a frame count (For example, 10 for the 10th frame), + or a relative value (For example, 1%). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video. :type start: str @@ -56,6 +56,7 @@ class PngImage(Image): _validation = { 'odatatype': {'required': True}, + 'start': {'required': True}, } _attribute_map = { diff --git a/azure-mgmt-media/azure/mgmt/media/models/png_image_py3.py b/azure-mgmt-media/azure/mgmt/media/models/png_image_py3.py index 4b7635032bd7..4dee64d04d6e 100644 --- a/azure-mgmt-media/azure/mgmt/media/models/png_image_py3.py +++ b/azure-mgmt-media/azure/mgmt/media/models/png_image_py3.py @@ -32,10 +32,10 @@ class PngImage(Image): resized to fit the desired output resolution(s). Default is AutoSize. Possible values include: 'None', 'AutoSize', 'AutoFit' :type stretch_mode: str or ~azure.mgmt.media.models.StretchMode - :param start: The position in the input video from where to start - generating thumbnails. The value can be in absolute timestamp (ISO 8601, - e.g: PT05S), or a frame count (For example, 10 for the 10th frame), or a - relative value (For example, 1%). Also supports a macro {Best}, which + :param start: Required. The position in the input video from where to + start generating thumbnails. The value can be in absolute timestamp (ISO + 8601, e.g: PT05S), or a frame count (For example, 10 for the 10th frame), + or a relative value (For example, 1%). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video. :type start: str @@ -56,6 +56,7 @@ class PngImage(Image): _validation = { 'odatatype': {'required': True}, + 'start': {'required': True}, } _attribute_map = { @@ -69,7 +70,7 @@ class PngImage(Image): 'layers': {'key': 'layers', 'type': '[PngLayer]'}, } - def __init__(self, *, label: str=None, key_frame_interval=None, stretch_mode=None, start: str=None, step: str=None, range: str=None, layers=None, **kwargs) -> None: + def __init__(self, *, start: str, label: str=None, key_frame_interval=None, stretch_mode=None, step: str=None, range: str=None, layers=None, **kwargs) -> None: super(PngImage, self).__init__(label=label, key_frame_interval=key_frame_interval, stretch_mode=stretch_mode, start=start, step=step, range=range, **kwargs) self.layers = layers self.odatatype = '#Microsoft.Media.PngImage' diff --git a/azure-mgmt-media/azure/mgmt/media/models/preset.py b/azure-mgmt-media/azure/mgmt/media/models/preset.py index 40819aede81f..9f6cba9f8e6c 100644 --- a/azure-mgmt-media/azure/mgmt/media/models/preset.py +++ b/azure-mgmt-media/azure/mgmt/media/models/preset.py @@ -17,8 +17,8 @@ class Preset(Model): the input media files should be processed. You probably want to use the sub-classes and not this class directly. Known - sub-classes are: AudioAnalyzerPreset, BuiltInStandardEncoderPreset, - StandardEncoderPreset + sub-classes are: FaceDetectorPreset, AudioAnalyzerPreset, + BuiltInStandardEncoderPreset, StandardEncoderPreset All required parameters must be populated in order to send to Azure. @@ -35,7 +35,7 @@ class Preset(Model): } _subtype_map = { - 'odatatype': {'#Microsoft.Media.AudioAnalyzerPreset': 'AudioAnalyzerPreset', '#Microsoft.Media.BuiltInStandardEncoderPreset': 'BuiltInStandardEncoderPreset', '#Microsoft.Media.StandardEncoderPreset': 'StandardEncoderPreset'} + 'odatatype': {'#Microsoft.Media.FaceDetectorPreset': 'FaceDetectorPreset', '#Microsoft.Media.AudioAnalyzerPreset': 'AudioAnalyzerPreset', '#Microsoft.Media.BuiltInStandardEncoderPreset': 'BuiltInStandardEncoderPreset', '#Microsoft.Media.StandardEncoderPreset': 'StandardEncoderPreset'} } def __init__(self, **kwargs): diff --git a/azure-mgmt-media/azure/mgmt/media/models/preset_py3.py b/azure-mgmt-media/azure/mgmt/media/models/preset_py3.py index c2a0dd8865e7..4cbc48fd9a82 100644 --- a/azure-mgmt-media/azure/mgmt/media/models/preset_py3.py +++ b/azure-mgmt-media/azure/mgmt/media/models/preset_py3.py @@ -17,8 +17,8 @@ class Preset(Model): the input media files should be processed. You probably want to use the sub-classes and not this class directly. Known - sub-classes are: AudioAnalyzerPreset, BuiltInStandardEncoderPreset, - StandardEncoderPreset + sub-classes are: FaceDetectorPreset, AudioAnalyzerPreset, + BuiltInStandardEncoderPreset, StandardEncoderPreset All required parameters must be populated in order to send to Azure. @@ -35,7 +35,7 @@ class Preset(Model): } _subtype_map = { - 'odatatype': {'#Microsoft.Media.AudioAnalyzerPreset': 'AudioAnalyzerPreset', '#Microsoft.Media.BuiltInStandardEncoderPreset': 'BuiltInStandardEncoderPreset', '#Microsoft.Media.StandardEncoderPreset': 'StandardEncoderPreset'} + 'odatatype': {'#Microsoft.Media.FaceDetectorPreset': 'FaceDetectorPreset', '#Microsoft.Media.AudioAnalyzerPreset': 'AudioAnalyzerPreset', '#Microsoft.Media.BuiltInStandardEncoderPreset': 'BuiltInStandardEncoderPreset', '#Microsoft.Media.StandardEncoderPreset': 'StandardEncoderPreset'} } def __init__(self, **kwargs) -> None: diff --git a/azure-mgmt-media/azure/mgmt/media/models/standard_encoder_preset.py b/azure-mgmt-media/azure/mgmt/media/models/standard_encoder_preset.py index 887900c3dc56..875e096e7707 100644 --- a/azure-mgmt-media/azure/mgmt/media/models/standard_encoder_preset.py +++ b/azure-mgmt-media/azure/mgmt/media/models/standard_encoder_preset.py @@ -23,15 +23,18 @@ class StandardEncoderPreset(Preset): :param filters: One or more filtering operations that are applied to the input media before encoding. :type filters: ~azure.mgmt.media.models.Filters - :param codecs: The list of codecs to be used when encoding the input - video. + :param codecs: Required. The list of codecs to be used when encoding the + input video. :type codecs: list[~azure.mgmt.media.models.Codec] - :param formats: The list of outputs to be produced by the encoder. + :param formats: Required. The list of outputs to be produced by the + encoder. :type formats: list[~azure.mgmt.media.models.Format] """ _validation = { 'odatatype': {'required': True}, + 'codecs': {'required': True}, + 'formats': {'required': True}, } _attribute_map = { diff --git a/azure-mgmt-media/azure/mgmt/media/models/standard_encoder_preset_py3.py b/azure-mgmt-media/azure/mgmt/media/models/standard_encoder_preset_py3.py index 9a53ddbe400a..23f9592716bc 100644 --- a/azure-mgmt-media/azure/mgmt/media/models/standard_encoder_preset_py3.py +++ b/azure-mgmt-media/azure/mgmt/media/models/standard_encoder_preset_py3.py @@ -23,15 +23,18 @@ class StandardEncoderPreset(Preset): :param filters: One or more filtering operations that are applied to the input media before encoding. :type filters: ~azure.mgmt.media.models.Filters - :param codecs: The list of codecs to be used when encoding the input - video. + :param codecs: Required. The list of codecs to be used when encoding the + input video. :type codecs: list[~azure.mgmt.media.models.Codec] - :param formats: The list of outputs to be produced by the encoder. + :param formats: Required. The list of outputs to be produced by the + encoder. :type formats: list[~azure.mgmt.media.models.Format] """ _validation = { 'odatatype': {'required': True}, + 'codecs': {'required': True}, + 'formats': {'required': True}, } _attribute_map = { @@ -41,7 +44,7 @@ class StandardEncoderPreset(Preset): 'formats': {'key': 'formats', 'type': '[Format]'}, } - def __init__(self, *, filters=None, codecs=None, formats=None, **kwargs) -> None: + def __init__(self, *, codecs, formats, filters=None, **kwargs) -> None: super(StandardEncoderPreset, self).__init__(**kwargs) self.filters = filters self.codecs = codecs diff --git a/azure-mgmt-media/azure/mgmt/media/models/streaming_locator.py b/azure-mgmt-media/azure/mgmt/media/models/streaming_locator.py index 45e12258448b..d307dc184978 100644 --- a/azure-mgmt-media/azure/mgmt/media/models/streaming_locator.py +++ b/azure-mgmt-media/azure/mgmt/media/models/streaming_locator.py @@ -54,6 +54,9 @@ class StreamingLocator(ProxyResource): :param alternative_media_id: Alternative Media ID of this Streaming Locator :type alternative_media_id: str + :param filters: A list of asset or account filters which apply to this + streaming locator + :type filters: list[str] """ _validation = { @@ -78,6 +81,7 @@ class StreamingLocator(ProxyResource): 'default_content_key_policy_name': {'key': 'properties.defaultContentKeyPolicyName', 'type': 'str'}, 'content_keys': {'key': 'properties.contentKeys', 'type': '[StreamingLocatorContentKey]'}, 'alternative_media_id': {'key': 'properties.alternativeMediaId', 'type': 'str'}, + 'filters': {'key': 'properties.filters', 'type': '[str]'}, } def __init__(self, **kwargs): @@ -91,3 +95,4 @@ def __init__(self, **kwargs): self.default_content_key_policy_name = kwargs.get('default_content_key_policy_name', None) self.content_keys = kwargs.get('content_keys', None) self.alternative_media_id = kwargs.get('alternative_media_id', None) + self.filters = kwargs.get('filters', None) diff --git a/azure-mgmt-media/azure/mgmt/media/models/streaming_locator_py3.py b/azure-mgmt-media/azure/mgmt/media/models/streaming_locator_py3.py index fb6670dabaf5..25254abf608c 100644 --- a/azure-mgmt-media/azure/mgmt/media/models/streaming_locator_py3.py +++ b/azure-mgmt-media/azure/mgmt/media/models/streaming_locator_py3.py @@ -54,6 +54,9 @@ class StreamingLocator(ProxyResource): :param alternative_media_id: Alternative Media ID of this Streaming Locator :type alternative_media_id: str + :param filters: A list of asset or account filters which apply to this + streaming locator + :type filters: list[str] """ _validation = { @@ -78,9 +81,10 @@ class StreamingLocator(ProxyResource): 'default_content_key_policy_name': {'key': 'properties.defaultContentKeyPolicyName', 'type': 'str'}, 'content_keys': {'key': 'properties.contentKeys', 'type': '[StreamingLocatorContentKey]'}, 'alternative_media_id': {'key': 'properties.alternativeMediaId', 'type': 'str'}, + 'filters': {'key': 'properties.filters', 'type': '[str]'}, } - def __init__(self, *, asset_name: str, streaming_policy_name: str, start_time=None, end_time=None, streaming_locator_id: str=None, default_content_key_policy_name: str=None, content_keys=None, alternative_media_id: str=None, **kwargs) -> None: + def __init__(self, *, asset_name: str, streaming_policy_name: str, start_time=None, end_time=None, streaming_locator_id: str=None, default_content_key_policy_name: str=None, content_keys=None, alternative_media_id: str=None, filters=None, **kwargs) -> None: super(StreamingLocator, self).__init__(**kwargs) self.asset_name = asset_name self.created = None @@ -91,3 +95,4 @@ def __init__(self, *, asset_name: str, streaming_policy_name: str, start_time=No self.default_content_key_policy_name = default_content_key_policy_name self.content_keys = content_keys self.alternative_media_id = alternative_media_id + self.filters = filters diff --git a/azure-mgmt-media/azure/mgmt/media/models/streaming_policy_fair_play_configuration.py b/azure-mgmt-media/azure/mgmt/media/models/streaming_policy_fair_play_configuration.py index 0d81e9c9f570..a8288b311b66 100644 --- a/azure-mgmt-media/azure/mgmt/media/models/streaming_policy_fair_play_configuration.py +++ b/azure-mgmt-media/azure/mgmt/media/models/streaming_policy_fair_play_configuration.py @@ -17,9 +17,14 @@ class StreamingPolicyFairPlayConfiguration(Model): All required parameters must be populated in order to send to Azure. - :param custom_license_acquisition_url_template: The template for a - customer service to deliver keys to end users. Not needed when using - Azure Media Services for issuing keys. + :param custom_license_acquisition_url_template: Template for the URL of + the custom service delivering licenses to end user players. Not required + when using Azure Media Services for issuing licenses. The template + supports replaceable tokens that the service will update at runtime with + the value specific to the request. The currently supported token values + are {AlternativeMediaId}, which is replaced with the value of + StreamingLocatorId.AlternativeMediaId, and {ContentKeyId}, which is + replaced with the value of identifier of the key being requested. :type custom_license_acquisition_url_template: str :param allow_persistent_license: Required. All license to be persistent or not diff --git a/azure-mgmt-media/azure/mgmt/media/models/streaming_policy_fair_play_configuration_py3.py b/azure-mgmt-media/azure/mgmt/media/models/streaming_policy_fair_play_configuration_py3.py index 1e61e334ba68..90dbf01a080a 100644 --- a/azure-mgmt-media/azure/mgmt/media/models/streaming_policy_fair_play_configuration_py3.py +++ b/azure-mgmt-media/azure/mgmt/media/models/streaming_policy_fair_play_configuration_py3.py @@ -17,9 +17,14 @@ class StreamingPolicyFairPlayConfiguration(Model): All required parameters must be populated in order to send to Azure. - :param custom_license_acquisition_url_template: The template for a - customer service to deliver keys to end users. Not needed when using - Azure Media Services for issuing keys. + :param custom_license_acquisition_url_template: Template for the URL of + the custom service delivering licenses to end user players. Not required + when using Azure Media Services for issuing licenses. The template + supports replaceable tokens that the service will update at runtime with + the value specific to the request. The currently supported token values + are {AlternativeMediaId}, which is replaced with the value of + StreamingLocatorId.AlternativeMediaId, and {ContentKeyId}, which is + replaced with the value of identifier of the key being requested. :type custom_license_acquisition_url_template: str :param allow_persistent_license: Required. All license to be persistent or not diff --git a/azure-mgmt-media/azure/mgmt/media/models/streaming_policy_play_ready_configuration.py b/azure-mgmt-media/azure/mgmt/media/models/streaming_policy_play_ready_configuration.py index 932644d2b26d..9227f67acdad 100644 --- a/azure-mgmt-media/azure/mgmt/media/models/streaming_policy_play_ready_configuration.py +++ b/azure-mgmt-media/azure/mgmt/media/models/streaming_policy_play_ready_configuration.py @@ -15,9 +15,14 @@ class StreamingPolicyPlayReadyConfiguration(Model): """Class to specify configurations of PlayReady in Streaming Policy. - :param custom_license_acquisition_url_template: The template for a - customer service to deliver keys to end users. Not needed when using - Azure Media Services for issuing keys. + :param custom_license_acquisition_url_template: Template for the URL of + the custom service delivering licenses to end user players. Not required + when using Azure Media Services for issuing licenses. The template + supports replaceable tokens that the service will update at runtime with + the value specific to the request. The currently supported token values + are {AlternativeMediaId}, which is replaced with the value of + StreamingLocatorId.AlternativeMediaId, and {ContentKeyId}, which is + replaced with the value of identifier of the key being requested. :type custom_license_acquisition_url_template: str :param play_ready_custom_attributes: Custom attributes for PlayReady :type play_ready_custom_attributes: str diff --git a/azure-mgmt-media/azure/mgmt/media/models/streaming_policy_play_ready_configuration_py3.py b/azure-mgmt-media/azure/mgmt/media/models/streaming_policy_play_ready_configuration_py3.py index 8bb4c18162d4..4e4105d17d0f 100644 --- a/azure-mgmt-media/azure/mgmt/media/models/streaming_policy_play_ready_configuration_py3.py +++ b/azure-mgmt-media/azure/mgmt/media/models/streaming_policy_play_ready_configuration_py3.py @@ -15,9 +15,14 @@ class StreamingPolicyPlayReadyConfiguration(Model): """Class to specify configurations of PlayReady in Streaming Policy. - :param custom_license_acquisition_url_template: The template for a - customer service to deliver keys to end users. Not needed when using - Azure Media Services for issuing keys. + :param custom_license_acquisition_url_template: Template for the URL of + the custom service delivering licenses to end user players. Not required + when using Azure Media Services for issuing licenses. The template + supports replaceable tokens that the service will update at runtime with + the value specific to the request. The currently supported token values + are {AlternativeMediaId}, which is replaced with the value of + StreamingLocatorId.AlternativeMediaId, and {ContentKeyId}, which is + replaced with the value of identifier of the key being requested. :type custom_license_acquisition_url_template: str :param play_ready_custom_attributes: Custom attributes for PlayReady :type play_ready_custom_attributes: str diff --git a/azure-mgmt-media/azure/mgmt/media/models/streaming_policy_widevine_configuration.py b/azure-mgmt-media/azure/mgmt/media/models/streaming_policy_widevine_configuration.py index fbbff1d1d91f..83441f90a009 100644 --- a/azure-mgmt-media/azure/mgmt/media/models/streaming_policy_widevine_configuration.py +++ b/azure-mgmt-media/azure/mgmt/media/models/streaming_policy_widevine_configuration.py @@ -15,9 +15,14 @@ class StreamingPolicyWidevineConfiguration(Model): """Class to specify configurations of Widevine in Streaming Policy. - :param custom_license_acquisition_url_template: The template for a - customer service to deliver keys to end users. Not needed when using - Azure Media Services for issuing keys. + :param custom_license_acquisition_url_template: Template for the URL of + the custom service delivering licenses to end user players. Not required + when using Azure Media Services for issuing licenses. The template + supports replaceable tokens that the service will update at runtime with + the value specific to the request. The currently supported token values + are {AlternativeMediaId}, which is replaced with the value of + StreamingLocatorId.AlternativeMediaId, and {ContentKeyId}, which is + replaced with the value of identifier of the key being requested. :type custom_license_acquisition_url_template: str """ diff --git a/azure-mgmt-media/azure/mgmt/media/models/streaming_policy_widevine_configuration_py3.py b/azure-mgmt-media/azure/mgmt/media/models/streaming_policy_widevine_configuration_py3.py index ea59e06bcb36..ab3078e8384d 100644 --- a/azure-mgmt-media/azure/mgmt/media/models/streaming_policy_widevine_configuration_py3.py +++ b/azure-mgmt-media/azure/mgmt/media/models/streaming_policy_widevine_configuration_py3.py @@ -15,9 +15,14 @@ class StreamingPolicyWidevineConfiguration(Model): """Class to specify configurations of Widevine in Streaming Policy. - :param custom_license_acquisition_url_template: The template for a - customer service to deliver keys to end users. Not needed when using - Azure Media Services for issuing keys. + :param custom_license_acquisition_url_template: Template for the URL of + the custom service delivering licenses to end user players. Not required + when using Azure Media Services for issuing licenses. The template + supports replaceable tokens that the service will update at runtime with + the value specific to the request. The currently supported token values + are {AlternativeMediaId}, which is replaced with the value of + StreamingLocatorId.AlternativeMediaId, and {ContentKeyId}, which is + replaced with the value of identifier of the key being requested. :type custom_license_acquisition_url_template: str """ diff --git a/azure-mgmt-media/azure/mgmt/media/models/transport_stream_format.py b/azure-mgmt-media/azure/mgmt/media/models/transport_stream_format.py index 876041ca9770..0e1f155d27fb 100644 --- a/azure-mgmt-media/azure/mgmt/media/models/transport_stream_format.py +++ b/azure-mgmt-media/azure/mgmt/media/models/transport_stream_format.py @@ -18,9 +18,9 @@ class TransportStreamFormat(MultiBitrateFormat): All required parameters must be populated in order to send to Azure. - :param filename_pattern: The pattern of the file names for the generated - output files. The following macros are supported in the file name: - {Basename} - The base name of the input video {Extension} - The + :param filename_pattern: Required. The pattern of the file names for the + generated output files. The following macros are supported in the file + name: {Basename} - The base name of the input video {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to @@ -35,6 +35,7 @@ class TransportStreamFormat(MultiBitrateFormat): """ _validation = { + 'filename_pattern': {'required': True}, 'odatatype': {'required': True}, } diff --git a/azure-mgmt-media/azure/mgmt/media/models/transport_stream_format_py3.py b/azure-mgmt-media/azure/mgmt/media/models/transport_stream_format_py3.py index 779febd2957c..b71dd8c3052f 100644 --- a/azure-mgmt-media/azure/mgmt/media/models/transport_stream_format_py3.py +++ b/azure-mgmt-media/azure/mgmt/media/models/transport_stream_format_py3.py @@ -18,9 +18,9 @@ class TransportStreamFormat(MultiBitrateFormat): All required parameters must be populated in order to send to Azure. - :param filename_pattern: The pattern of the file names for the generated - output files. The following macros are supported in the file name: - {Basename} - The base name of the input video {Extension} - The + :param filename_pattern: Required. The pattern of the file names for the + generated output files. The following macros are supported in the file + name: {Basename} - The base name of the input video {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to @@ -35,6 +35,7 @@ class TransportStreamFormat(MultiBitrateFormat): """ _validation = { + 'filename_pattern': {'required': True}, 'odatatype': {'required': True}, } @@ -44,6 +45,6 @@ class TransportStreamFormat(MultiBitrateFormat): 'output_files': {'key': 'outputFiles', 'type': '[OutputFile]'}, } - def __init__(self, *, filename_pattern: str=None, output_files=None, **kwargs) -> None: + def __init__(self, *, filename_pattern: str, output_files=None, **kwargs) -> None: super(TransportStreamFormat, self).__init__(filename_pattern=filename_pattern, output_files=output_files, **kwargs) self.odatatype = '#Microsoft.Media.TransportStreamFormat' diff --git a/azure-mgmt-media/azure/mgmt/media/models/video_analyzer_preset.py b/azure-mgmt-media/azure/mgmt/media/models/video_analyzer_preset.py index 203a6501bd52..b02b85ac881a 100644 --- a/azure-mgmt-media/azure/mgmt/media/models/video_analyzer_preset.py +++ b/azure-mgmt-media/azure/mgmt/media/models/video_analyzer_preset.py @@ -21,15 +21,21 @@ class VideoAnalyzerPreset(AudioAnalyzerPreset): :param odatatype: Required. Constant filled by server. :type odatatype: str :param audio_language: The language for the audio payload in the input - using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). The list - of supported languages are, 'en-US', 'en-GB', 'es-ES', 'es-MX', 'fr-FR', - 'it-IT', 'ja-JP', 'pt-BR', 'zh-CN', 'de-DE', 'ar-EG', 'ru-RU', 'hi-IN'. If - not specified, automatic language detection would be employed. This + using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). The list + of supported languages are English ('en-US' and 'en-GB'), Spanish ('es-ES' + and 'es-MX'), French ('fr-FR'), Italian ('it-IT'), Japanese ('ja-JP'), + Portuguese ('pt-BR'), Chinese ('zh-CN'), German ('de-DE'), Arabic + ('ar-EG'), Russian ('ru-RU'), Hindi ('hi-IN'), and Korean ('ko-KR'). If + you know the language of your content, it is recommended that you specify + it. If the language isn't specified or set to null, automatic language + detection will choose the first language detected and process with the + selected language for the duration of the file. This language detection feature currently supports English, Chinese, French, German, Italian, - Japanese, Spanish, Russian, and Portuguese. The automatic detection works - best with audio recordings with clearly discernable speech. If automatic - detection fails to find the language, transcription would fallback to - English. + Japanese, Spanish, Russian, and Portuguese. It does not currently support + dynamically switching between languages after the first language is + detected. The automatic detection works best with audio recordings with + clearly discernable speech. If automatic detection fails to find the + language, transcription would fallback to 'en-US'." :type audio_language: str :param insights_to_extract: The type of insights to be extracted. If not set then based on the content the type will selected. If the content is diff --git a/azure-mgmt-media/azure/mgmt/media/models/video_analyzer_preset_py3.py b/azure-mgmt-media/azure/mgmt/media/models/video_analyzer_preset_py3.py index f54fa7bebc7e..ac8508da22a6 100644 --- a/azure-mgmt-media/azure/mgmt/media/models/video_analyzer_preset_py3.py +++ b/azure-mgmt-media/azure/mgmt/media/models/video_analyzer_preset_py3.py @@ -21,15 +21,21 @@ class VideoAnalyzerPreset(AudioAnalyzerPreset): :param odatatype: Required. Constant filled by server. :type odatatype: str :param audio_language: The language for the audio payload in the input - using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). The list - of supported languages are, 'en-US', 'en-GB', 'es-ES', 'es-MX', 'fr-FR', - 'it-IT', 'ja-JP', 'pt-BR', 'zh-CN', 'de-DE', 'ar-EG', 'ru-RU', 'hi-IN'. If - not specified, automatic language detection would be employed. This + using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). The list + of supported languages are English ('en-US' and 'en-GB'), Spanish ('es-ES' + and 'es-MX'), French ('fr-FR'), Italian ('it-IT'), Japanese ('ja-JP'), + Portuguese ('pt-BR'), Chinese ('zh-CN'), German ('de-DE'), Arabic + ('ar-EG'), Russian ('ru-RU'), Hindi ('hi-IN'), and Korean ('ko-KR'). If + you know the language of your content, it is recommended that you specify + it. If the language isn't specified or set to null, automatic language + detection will choose the first language detected and process with the + selected language for the duration of the file. This language detection feature currently supports English, Chinese, French, German, Italian, - Japanese, Spanish, Russian, and Portuguese. The automatic detection works - best with audio recordings with clearly discernable speech. If automatic - detection fails to find the language, transcription would fallback to - English. + Japanese, Spanish, Russian, and Portuguese. It does not currently support + dynamically switching between languages after the first language is + detected. The automatic detection works best with audio recordings with + clearly discernable speech. If automatic detection fails to find the + language, transcription would fallback to 'en-US'." :type audio_language: str :param insights_to_extract: The type of insights to be extracted. If not set then based on the content the type will selected. If the content is diff --git a/azure-mgmt-media/azure/mgmt/media/models/video_layer.py b/azure-mgmt-media/azure/mgmt/media/models/video_layer.py index 2f3ecf54083c..bc142466ce60 100644 --- a/azure-mgmt-media/azure/mgmt/media/models/video_layer.py +++ b/azure-mgmt-media/azure/mgmt/media/models/video_layer.py @@ -35,8 +35,9 @@ class VideoLayer(Layer): :type label: str :param odatatype: Required. Constant filled by server. :type odatatype: str - :param bitrate: The average bitrate in bits per second at which to encode - the input video when generating this layer. This is a required field. + :param bitrate: Required. The average bitrate in bits per second at which + to encode the input video when generating this layer. This is a required + field. :type bitrate: int :param max_bitrate: The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the @@ -65,6 +66,7 @@ class VideoLayer(Layer): _validation = { 'odatatype': {'required': True}, + 'bitrate': {'required': True}, } _attribute_map = { diff --git a/azure-mgmt-media/azure/mgmt/media/models/video_layer_py3.py b/azure-mgmt-media/azure/mgmt/media/models/video_layer_py3.py index e11e718b60b1..755d7b259531 100644 --- a/azure-mgmt-media/azure/mgmt/media/models/video_layer_py3.py +++ b/azure-mgmt-media/azure/mgmt/media/models/video_layer_py3.py @@ -35,8 +35,9 @@ class VideoLayer(Layer): :type label: str :param odatatype: Required. Constant filled by server. :type odatatype: str - :param bitrate: The average bitrate in bits per second at which to encode - the input video when generating this layer. This is a required field. + :param bitrate: Required. The average bitrate in bits per second at which + to encode the input video when generating this layer. This is a required + field. :type bitrate: int :param max_bitrate: The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the @@ -65,6 +66,7 @@ class VideoLayer(Layer): _validation = { 'odatatype': {'required': True}, + 'bitrate': {'required': True}, } _attribute_map = { @@ -84,7 +86,7 @@ class VideoLayer(Layer): 'odatatype': {'#Microsoft.Media.H264Layer': 'H264Layer'} } - def __init__(self, *, width: str=None, height: str=None, label: str=None, bitrate: int=None, max_bitrate: int=None, b_frames: int=None, frame_rate: str=None, slices: int=None, adaptive_bframe: bool=None, **kwargs) -> None: + def __init__(self, *, bitrate: int, width: str=None, height: str=None, label: str=None, max_bitrate: int=None, b_frames: int=None, frame_rate: str=None, slices: int=None, adaptive_bframe: bool=None, **kwargs) -> None: super(VideoLayer, self).__init__(width=width, height=height, label=label, **kwargs) self.bitrate = bitrate self.max_bitrate = max_bitrate diff --git a/azure-mgmt-media/azure/mgmt/media/models/video_overlay.py b/azure-mgmt-media/azure/mgmt/media/models/video_overlay.py index f24487335fbd..1128cfdd1296 100644 --- a/azure-mgmt-media/azure/mgmt/media/models/video_overlay.py +++ b/azure-mgmt-media/azure/mgmt/media/models/video_overlay.py @@ -17,11 +17,11 @@ class VideoOverlay(Overlay): All required parameters must be populated in order to send to Azure. - :param input_label: The label of the job input which is to be used as an - overlay. The Input must specify exactly one file. You can specify an image - file in JPG or PNG formats, or an audio file (such as a WAV, MP3, WMA or - M4A file), or a video file. See https://aka.ms/mesformats for the complete - list of supported audio and video file formats. + :param input_label: Required. The label of the job input which is to be + used as an overlay. The Input must specify exactly one file. You can + specify an image file in JPG or PNG formats, or an audio file (such as a + WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats + for the complete list of supported audio and video file formats. :type input_label: str :param start: The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For @@ -60,6 +60,7 @@ class VideoOverlay(Overlay): """ _validation = { + 'input_label': {'required': True}, 'odatatype': {'required': True}, } diff --git a/azure-mgmt-media/azure/mgmt/media/models/video_overlay_py3.py b/azure-mgmt-media/azure/mgmt/media/models/video_overlay_py3.py index 80158981e6dd..c0d65b8d749d 100644 --- a/azure-mgmt-media/azure/mgmt/media/models/video_overlay_py3.py +++ b/azure-mgmt-media/azure/mgmt/media/models/video_overlay_py3.py @@ -17,11 +17,11 @@ class VideoOverlay(Overlay): All required parameters must be populated in order to send to Azure. - :param input_label: The label of the job input which is to be used as an - overlay. The Input must specify exactly one file. You can specify an image - file in JPG or PNG formats, or an audio file (such as a WAV, MP3, WMA or - M4A file), or a video file. See https://aka.ms/mesformats for the complete - list of supported audio and video file formats. + :param input_label: Required. The label of the job input which is to be + used as an overlay. The Input must specify exactly one file. You can + specify an image file in JPG or PNG formats, or an audio file (such as a + WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats + for the complete list of supported audio and video file formats. :type input_label: str :param start: The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For @@ -60,6 +60,7 @@ class VideoOverlay(Overlay): """ _validation = { + 'input_label': {'required': True}, 'odatatype': {'required': True}, } @@ -76,7 +77,7 @@ class VideoOverlay(Overlay): 'crop_rectangle': {'key': 'cropRectangle', 'type': 'Rectangle'}, } - def __init__(self, *, input_label: str=None, start=None, end=None, fade_in_duration=None, fade_out_duration=None, audio_gain_level: float=None, position=None, opacity: float=None, crop_rectangle=None, **kwargs) -> None: + def __init__(self, *, input_label: str, start=None, end=None, fade_in_duration=None, fade_out_duration=None, audio_gain_level: float=None, position=None, opacity: float=None, crop_rectangle=None, **kwargs) -> None: super(VideoOverlay, self).__init__(input_label=input_label, start=start, end=end, fade_in_duration=fade_in_duration, fade_out_duration=fade_out_duration, audio_gain_level=audio_gain_level, **kwargs) self.position = position self.opacity = opacity diff --git a/azure-mgmt-media/azure/mgmt/media/version.py b/azure-mgmt-media/azure/mgmt/media/version.py index 24b9de3384da..8b09f47f15f0 100644 --- a/azure-mgmt-media/azure/mgmt/media/version.py +++ b/azure-mgmt-media/azure/mgmt/media/version.py @@ -9,5 +9,5 @@ # regenerated. # -------------------------------------------------------------------------- -VERSION = "1.1.0" +VERSION = "2018-07-01"