diff --git a/azure-batch/azure/batch/batch_service_client.py b/azure-batch/azure/batch/batch_service_client.py index fb6152594700..1d40e5a823a6 100644 --- a/azure-batch/azure/batch/batch_service_client.py +++ b/azure-batch/azure/batch/batch_service_client.py @@ -9,7 +9,7 @@ # regenerated. # -------------------------------------------------------------------------- -from msrest.service_client import ServiceClient +from msrest.service_client import SDKClient from msrest import Serializer, Deserializer from msrestazure import AzureConfiguration from .version import VERSION @@ -52,7 +52,7 @@ def __init__( self.credentials = credentials -class BatchServiceClient(object): +class BatchServiceClient(SDKClient): """A client for issuing REST requests to the Azure Batch service. :ivar config: Configuration for client. @@ -87,7 +87,7 @@ def __init__( self, credentials, base_url=None): self.config = BatchServiceClientConfiguration(credentials, base_url) - self._client = ServiceClient(self.config.credentials, self.config) + super(BatchServiceClient, self).__init__(self.config.credentials, self.config) client_models = {k: v for k, v in models.__dict__.items() if isinstance(v, type)} self.api_version = '2018-03-01.6.1' diff --git a/azure-batch/azure/batch/models/batch_service_client_enums.py b/azure-batch/azure/batch/models/batch_service_client_enums.py index f967ad757ae6..d67fac6edbb1 100644 --- a/azure-batch/azure/batch/models/batch_service_client_enums.py +++ b/azure-batch/azure/batch/models/batch_service_client_enums.py @@ -14,270 +14,270 @@ class OSType(str, Enum): - linux = "linux" - windows = "windows" + linux = "linux" #: The Linux operating system. + windows = "windows" #: The Windows operating system. class AccessScope(str, Enum): - job = "job" + job = "job" #: Grants access to perform all operations on the job containing the task. class CertificateState(str, Enum): - active = "active" - deleting = "deleting" - delete_failed = "deletefailed" + active = "active" #: The certificate is available for use in pools. + deleting = "deleting" #: The user has requested that the certificate be deleted, but the delete operation has not yet completed. You may not reference the certificate when creating or updating pools. + delete_failed = "deletefailed" #: The user requested that the certificate be deleted, but there are pools that still have references to the certificate, or it is still installed on one or more compute nodes. (The latter can occur if the certificate has been removed from the pool, but the node has not yet restarted. Nodes refresh their certificates only when they restart.) You may use the cancel certificate delete operation to cancel the delete, or the delete certificate operation to retry the delete. class CertificateFormat(str, Enum): - pfx = "pfx" - cer = "cer" + pfx = "pfx" #: The certificate is a PFX (PKCS#12) formatted certificate or certificate chain. + cer = "cer" #: The certificate is a base64-encoded X.509 certificate. class JobAction(str, Enum): - none = "none" - disable = "disable" - terminate = "terminate" + none = "none" #: Take no action. + disable = "disable" #: Disable the job. This is equivalent to calling the disable job API, with a disableTasks value of requeue. + terminate = "terminate" #: Terminate the job. The terminateReason in the job's executionInfo is set to "TaskFailed". class DependencyAction(str, Enum): - satisfy = "satisfy" - block = "block" + satisfy = "satisfy" #: Satisfy the task's dependencies. + block = "block" #: Block the task's dependencies. class AutoUserScope(str, Enum): - task = "task" - pool = "pool" + task = "task" #: Specifies that the service should create a new user for the task. + pool = "pool" #: Specifies that the task runs as the common auto user account which is created on every node in a pool. class ElevationLevel(str, Enum): - non_admin = "nonadmin" - admin = "admin" + non_admin = "nonadmin" #: The user is a standard user without elevated access. + admin = "admin" #: The user is a user with elevated access and operates with full Administrator permissions. class OutputFileUploadCondition(str, Enum): - task_success = "tasksuccess" - task_failure = "taskfailure" - task_completion = "taskcompletion" + task_success = "tasksuccess" #: Upload the file(s) only after the task process exits with an exit code of 0. + task_failure = "taskfailure" #: Upload the file(s) only after the task process exits with a nonzero exit code. + task_completion = "taskcompletion" #: Upload the file(s) after the task process exits, no matter what the exit code was. class ComputeNodeFillType(str, Enum): - spread = "spread" - pack = "pack" + spread = "spread" #: Tasks should be assigned evenly across all nodes in the pool. + pack = "pack" #: As many tasks as possible (maxTasksPerNode) should be assigned to each node in the pool before any tasks are assigned to the next node in the pool. class CertificateStoreLocation(str, Enum): - current_user = "currentuser" - local_machine = "localmachine" + current_user = "currentuser" #: Certificates should be installed to the CurrentUser certificate store. + local_machine = "localmachine" #: Certificates should be installed to the LocalMachine certificate store. class CertificateVisibility(str, Enum): - start_task = "starttask" - task = "task" - remote_user = "remoteuser" + start_task = "starttask" #: The certificate should be visible to the user account under which the start task is run. + task = "task" #: The certificate should be visibile to the user accounts under which job tasks are run. + remote_user = "remoteuser" #: The certificate should be visibile to the user accounts under which users remotely access the node. class CachingType(str, Enum): - none = "none" - read_only = "readonly" - read_write = "readwrite" + none = "none" #: The caching mode for the disk is not enabled. + read_only = "readonly" #: The caching mode for the disk is read only. + read_write = "readwrite" #: The caching mode for the disk is read and write. class StorageAccountType(str, Enum): - standard_lrs = "standard_lrs" - premium_lrs = "premium_lrs" + standard_lrs = "standard_lrs" #: The data disk should use standard locally redundant storage. + premium_lrs = "premium_lrs" #: The data disk should use premium locally redundant storage. class InboundEndpointProtocol(str, Enum): - tcp = "tcp" - udp = "udp" + tcp = "tcp" #: Use TCP for the endpoint. + udp = "udp" #: Use UDP for the endpoint. class NetworkSecurityGroupRuleAccess(str, Enum): - allow = "allow" - deny = "deny" + allow = "allow" #: Allow access. + deny = "deny" #: Deny access. class PoolLifetimeOption(str, Enum): - job_schedule = "jobschedule" - job = "job" + job_schedule = "jobschedule" #: The pool exists for the lifetime of the job schedule. The Batch Service creates the pool when it creates the first job on the schedule. You may apply this option only to job schedules, not to jobs. + job = "job" #: The pool exists for the lifetime of the job to which it is dedicated. The Batch service creates the pool when it creates the job. If the 'job' option is applied to a job schedule, the Batch service creates a new auto pool for every job created on the schedule. class OnAllTasksComplete(str, Enum): - no_action = "noaction" - terminate_job = "terminatejob" + no_action = "noaction" #: Do nothing. The job remains active unless terminated or disabled by some other means. + terminate_job = "terminatejob" #: Terminate the job. The job's terminateReason is set to 'AllTasksComplete'. class OnTaskFailure(str, Enum): - no_action = "noaction" - perform_exit_options_job_action = "performexitoptionsjobaction" + no_action = "noaction" #: Do nothing. The job remains active unless terminated or disabled by some other means. + perform_exit_options_job_action = "performexitoptionsjobaction" #: Take the action associated with the task exit condition in the task's exitConditions collection. (This may still result in no action being taken, if that is what the task specifies.) class JobScheduleState(str, Enum): - active = "active" - completed = "completed" - disabled = "disabled" - terminating = "terminating" - deleting = "deleting" + active = "active" #: The job schedule is active and will create jobs as per its schedule. + completed = "completed" #: The schedule has terminated, either by reaching its end time or by the user terminating it explicitly. + disabled = "disabled" #: The user has disabled the schedule. The scheduler will not initiate any new jobs will on this schedule, but any existing active job will continue to run. + terminating = "terminating" #: The schedule has no more work to do, or has been explicitly terminated by the user, but the termination operation is still in progress. The scheduler will not initiate any new jobs for this schedule, nor is any existing job active. + deleting = "deleting" #: The user has requested that the schedule be deleted, but the delete operation is still in progress. The scheduler will not initiate any new jobs for this schedule, and will delete any existing jobs and tasks under the schedule, including any active job. The schedule will be deleted when all jobs and tasks under the schedule have been deleted. class ErrorCategory(str, Enum): - user_error = "usererror" - server_error = "servererror" + user_error = "usererror" #: The error is due to a user issue, such as misconfiguration. + server_error = "servererror" #: The error is due to an internal server issue. class JobState(str, Enum): - active = "active" - disabling = "disabling" - disabled = "disabled" - enabling = "enabling" - terminating = "terminating" - completed = "completed" - deleting = "deleting" + active = "active" #: The job is available to have tasks scheduled. + disabling = "disabling" #: A user has requested that the job be disabled, but the disable operation is still in progress (for example, waiting for tasks to terminate). + disabled = "disabled" #: A user has disabled the job. No tasks are running, and no new tasks will be scheduled. + enabling = "enabling" #: A user has requested that the job be enabled, but the enable operation is still in progress. + terminating = "terminating" #: The job is about to complete, either because a Job Manager task has completed or because the user has terminated the job, but the terminate operation is still in progress (for example, because Job Release tasks are running). + completed = "completed" #: All tasks have terminated, and the system will not accept any more tasks or any further changes to the job. + deleting = "deleting" #: A user has requested that the job be deleted, but the delete operation is still in progress (for example, because the system is still terminating running tasks). class JobPreparationTaskState(str, Enum): - running = "running" - completed = "completed" + running = "running" #: The task is currently running (including retrying). + completed = "completed" #: The task has exited with exit code 0, or the task has exhausted its retry limit, or the Batch service was unable to start the task due to task preparation errors (such as resource file download failures). class TaskExecutionResult(str, Enum): - success = "success" - failure = "failure" + success = "success" #: The task ran successfully. + failure = "failure" #: There was an error during processing of the task. The failure may have occurred before the task process was launched, while the task process was executing, or after the task process exited. class JobReleaseTaskState(str, Enum): - running = "running" - completed = "completed" + running = "running" #: The task is currently running (including retrying). + completed = "completed" #: The task has exited with exit code 0, or the task has exhausted its retry limit, or the Batch service was unable to start the task due to task preparation errors (such as resource file download failures). class TaskCountValidationStatus(str, Enum): - validated = "validated" - unvalidated = "unvalidated" + validated = "validated" #: The Batch service has validated the state counts against the task states as reported in the List Tasks API. + unvalidated = "unvalidated" #: The Batch service has not been able to check state counts against the task states as reported in the List Tasks API. The validationStatus may be unvalidated if the job contains more than 200,000 tasks. class PoolState(str, Enum): - active = "active" - deleting = "deleting" - upgrading = "upgrading" + active = "active" #: The pool is available to run tasks subject to the availability of compute nodes. + deleting = "deleting" #: The user has requested that the pool be deleted, but the delete operation has not yet completed. + upgrading = "upgrading" #: The user has requested that the operating system of the pool's nodes be upgraded, but the upgrade operation has not yet completed (that is, some nodes in the pool have not yet been upgraded). While upgrading, the pool may be able to run tasks (with reduced capacity) but this is not guaranteed. class AllocationState(str, Enum): - steady = "steady" - resizing = "resizing" - stopping = "stopping" + steady = "steady" #: The pool is not resizing. There are no changes to the number of nodes in the pool in progress. A pool enters this state when it is created and when no operations are being performed on the pool to change the number of nodes. + resizing = "resizing" #: The pool is resizing; that is, compute nodes are being added to or removed from the pool. + stopping = "stopping" #: The pool was resizing, but the user has requested that the resize be stopped, but the stop request has not yet been completed. class TaskState(str, Enum): - active = "active" - preparing = "preparing" - running = "running" - completed = "completed" + active = "active" #: The task is queued and able to run, but is not currently assigned to a compute node. A task enters this state when it is created, when it is enabled after being disabled, or when it is awaiting a retry after a failed run. + preparing = "preparing" #: The task has been assigned to a compute node, but is waiting for a required Job Preparation task to complete on the node. If the Job Preparation task succeeds, the task will move to running. If the Job Preparation task fails, the task will return to active and will be eligible to be assigned to a different node. + running = "running" #: The task is running on a compute node. This includes task-level preparation such as downloading resource files or deploying application packages specified on the task - it does not necessarily mean that the task command line has started executing. + completed = "completed" #: The task is no longer eligible to run, usually because the task has finished successfully, or the task has finished unsuccessfully and has exhausted its retry limit. A task is also marked as completed if an error occurred launching the task, or when the task has been terminated. class TaskAddStatus(str, Enum): - success = "success" - client_error = "clienterror" - server_error = "servererror" + success = "success" #: The task was added successfully. + client_error = "clienterror" #: The task failed to add due to a client error and should not be retried without modifying the request as appropriate. + server_error = "servererror" #: Task failed to add due to a server error and can be retried without modification. class SubtaskState(str, Enum): - preparing = "preparing" - running = "running" - completed = "completed" + preparing = "preparing" #: The task has been assigned to a compute node, but is waiting for a required Job Preparation task to complete on the node. If the Job Preparation task succeeds, the task will move to running. If the Job Preparation task fails, the task will return to active and will be eligible to be assigned to a different node. + running = "running" #: The task is running on a compute node. This includes task-level preparation such as downloading resource files or deploying application packages specified on the task - it does not necessarily mean that the task command line has started executing. + completed = "completed" #: The task is no longer eligible to run, usually because the task has finished successfully, or the task has finished unsuccessfully and has exhausted its retry limit. A task is also marked as completed if an error occurred launching the task, or when the task has been terminated. class StartTaskState(str, Enum): - running = "running" - completed = "completed" + running = "running" #: The start task is currently running. + completed = "completed" #: The start task has exited with exit code 0, or the start task has failed and the retry limit has reached, or the start task process did not run due to task preparation errors (such as resource file download failures). class ComputeNodeState(str, Enum): - idle = "idle" - rebooting = "rebooting" - reimaging = "reimaging" - running = "running" - unusable = "unusable" - creating = "creating" - starting = "starting" - waiting_for_start_task = "waitingforstarttask" - start_task_failed = "starttaskfailed" - unknown = "unknown" - leaving_pool = "leavingpool" - offline = "offline" - preempted = "preempted" + idle = "idle" #: The node is not currently running a task. + rebooting = "rebooting" #: The node is rebooting. + reimaging = "reimaging" #: The node is reimaging. + running = "running" #: The node is running one or more tasks (other than a start task). + unusable = "unusable" #: The node cannot be used for task execution due to errors. + creating = "creating" #: The Batch service has obtained the underlying virtual machine from Azure Compute, but it has not yet started to join the pool. + starting = "starting" #: The Batch service is starting on the underlying virtual machine. + waiting_for_start_task = "waitingforstarttask" #: The start task has started running on the compute node, but waitForSuccess is set and the start task has not yet completed. + start_task_failed = "starttaskfailed" #: The start task has failed on the compute node (and exhausted all retries), and waitForSuccess is set. The node is not usable for running tasks. + unknown = "unknown" #: The Batch service has lost contact with the node, and does not know its true state. + leaving_pool = "leavingpool" #: The node is leaving the pool, either because the user explicitly removed it or because the pool is resizing or autoscaling down. + offline = "offline" #: The node is not currently running a task, and scheduling of new tasks to the node is disabled. + preempted = "preempted" #: The low-priority node has been preempted. Tasks which were running on the node when it was pre-empted will be rescheduled when another node becomes available. class SchedulingState(str, Enum): - enabled = "enabled" - disabled = "disabled" + enabled = "enabled" #: Tasks can be scheduled on the node. + disabled = "disabled" #: No new tasks will be scheduled on the node. Tasks already running on the node may still run to completion. All nodes start with scheduling enabled. class DisableJobOption(str, Enum): - requeue = "requeue" - terminate = "terminate" - wait = "wait" + requeue = "requeue" #: Terminate running tasks and requeue them. The tasks will run again when the job is enabled. + terminate = "terminate" #: Terminate running tasks. The tasks will be completed with failureInfo indicating that they were terminated, and will not run again. + wait = "wait" #: Allow currently running tasks to complete. class ComputeNodeDeallocationOption(str, Enum): - requeue = "requeue" - terminate = "terminate" - task_completion = "taskcompletion" - retained_data = "retaineddata" + requeue = "requeue" #: Terminate running task processes and requeue the tasks. The tasks will run again when a node is available. Remove nodes as soon as tasks have been terminated. + terminate = "terminate" #: Terminate running tasks. The tasks will be completed with failureInfo indicating that they were terminated, and will not run again. Remove nodes as soon as tasks have been terminated. + task_completion = "taskcompletion" #: Allow currently running tasks to complete. Schedule no new tasks while waiting. Remove nodes when all tasks have completed. + retained_data = "retaineddata" #: Allow currently running tasks to complete, then wait for all task data retention periods to expire. Schedule no new tasks while waiting. Remove nodes when all task retention periods have expired. class ComputeNodeRebootOption(str, Enum): - requeue = "requeue" - terminate = "terminate" - task_completion = "taskcompletion" - retained_data = "retaineddata" + requeue = "requeue" #: Terminate running task processes and requeue the tasks. The tasks will run again when a node is available. Restart the node as soon as tasks have been terminated. + terminate = "terminate" #: Terminate running tasks. The tasks will be completed with failureInfo indicating that they were terminated, and will not run again. Restart the node as soon as tasks have been terminated. + task_completion = "taskcompletion" #: Allow currently running tasks to complete. Schedule no new tasks while waiting. Restart the node when all tasks have completed. + retained_data = "retaineddata" #: Allow currently running tasks to complete, then wait for all task data retention periods to expire. Schedule no new tasks while waiting. Restart the node when all task retention periods have expired. class ComputeNodeReimageOption(str, Enum): - requeue = "requeue" - terminate = "terminate" - task_completion = "taskcompletion" - retained_data = "retaineddata" + requeue = "requeue" #: Terminate running task processes and requeue the tasks. The tasks will run again when a node is available. Reimage the node as soon as tasks have been terminated. + terminate = "terminate" #: Terminate running tasks. The tasks will be completed with failureInfo indicating that they were terminated, and will not run again. Reimage the node as soon as tasks have been terminated. + task_completion = "taskcompletion" #: Allow currently running tasks to complete. Schedule no new tasks while waiting. Reimage the node when all tasks have completed. + retained_data = "retaineddata" #: Allow currently running tasks to complete, then wait for all task data retention periods to expire. Schedule no new tasks while waiting. Reimage the node when all task retention periods have expired. class DisableComputeNodeSchedulingOption(str, Enum): - requeue = "requeue" - terminate = "terminate" - task_completion = "taskcompletion" + requeue = "requeue" #: Terminate running task processes and requeue the tasks. The tasks may run again on other compute nodes, or when task scheduling is re-enabled on this node. Enter offline state as soon as tasks have been terminated. + terminate = "terminate" #: Terminate running tasks. The tasks will be completed with failureInfo indicating that they were terminated, and will not run again. Enter offline state as soon as tasks have been terminated. + task_completion = "taskcompletion" #: Allow currently running tasks to complete. Schedule no new tasks while waiting. Enter offline state when all tasks have completed. diff --git a/azure-batch/azure/batch/models/cloud_job.py b/azure-batch/azure/batch/models/cloud_job.py index 8c3ca4d7f8e2..4e6cafbf48d9 100644 --- a/azure-batch/azure/batch/models/cloud_job.py +++ b/azure-batch/azure/batch/models/cloud_job.py @@ -99,7 +99,9 @@ class CloudJob(Model): :param execution_info: The execution information for the job. :type execution_info: ~azure.batch.models.JobExecutionInformation :param stats: Resource usage statistics for the entire lifetime of the - job. + job. The statistics may not be immediately available. The Batch service + performs periodic roll-up of statistics. The typical delay is about 30 + minutes. :type stats: ~azure.batch.models.JobStatistics """ diff --git a/azure-batch/azure/batch/models/cloud_job_py3.py b/azure-batch/azure/batch/models/cloud_job_py3.py index 7302e2787551..226cb3bb7cc3 100644 --- a/azure-batch/azure/batch/models/cloud_job_py3.py +++ b/azure-batch/azure/batch/models/cloud_job_py3.py @@ -99,7 +99,9 @@ class CloudJob(Model): :param execution_info: The execution information for the job. :type execution_info: ~azure.batch.models.JobExecutionInformation :param stats: Resource usage statistics for the entire lifetime of the - job. + job. The statistics may not be immediately available. The Batch service + performs periodic roll-up of statistics. The typical delay is about 30 + minutes. :type stats: ~azure.batch.models.JobStatistics """ diff --git a/azure-batch/azure/batch/models/cloud_job_schedule.py b/azure-batch/azure/batch/models/cloud_job_schedule.py index 5c52fda98069..1a2a33fda27d 100644 --- a/azure-batch/azure/batch/models/cloud_job_schedule.py +++ b/azure-batch/azure/batch/models/cloud_job_schedule.py @@ -64,6 +64,9 @@ class CloudJobSchedule(Model): is solely for the use of user code. :type metadata: list[~azure.batch.models.MetadataItem] :param stats: The lifetime resource usage statistics for the job schedule. + The statistics may not be immediately available. The Batch service + performs periodic roll-up of statistics. The typical delay is about 30 + minutes. :type stats: ~azure.batch.models.JobScheduleStatistics """ diff --git a/azure-batch/azure/batch/models/cloud_job_schedule_py3.py b/azure-batch/azure/batch/models/cloud_job_schedule_py3.py index 773c843b3c89..1542fe56acbe 100644 --- a/azure-batch/azure/batch/models/cloud_job_schedule_py3.py +++ b/azure-batch/azure/batch/models/cloud_job_schedule_py3.py @@ -64,6 +64,9 @@ class CloudJobSchedule(Model): is solely for the use of user code. :type metadata: list[~azure.batch.models.MetadataItem] :param stats: The lifetime resource usage statistics for the job schedule. + The statistics may not be immediately available. The Batch service + performs periodic roll-up of statistics. The typical delay is about 30 + minutes. :type stats: ~azure.batch.models.JobScheduleStatistics """ diff --git a/azure-batch/azure/batch/models/cloud_pool.py b/azure-batch/azure/batch/models/cloud_pool.py index 202170371869..7e1602157653 100644 --- a/azure-batch/azure/batch/models/cloud_pool.py +++ b/azure-batch/azure/batch/models/cloud_pool.py @@ -56,7 +56,7 @@ class CloudPool(Model): machines in a pool are the same size. For information about available sizes of virtual machines for Cloud Services pools (pools created with cloudServiceConfiguration), see Sizes for Cloud Services - (http://azure.microsoft.com/documentation/articles/cloud-services-sizes-specs/). + (https://azure.microsoft.com/documentation/articles/cloud-services-sizes-specs/). Batch supports all Cloud Services VM sizes except ExtraSmall, A1V2 and A2V2. For information about available VM sizes for pools using images from the Virtual Machines Marketplace (pools created with @@ -165,7 +165,9 @@ class CloudPool(Model): metadata. :type metadata: list[~azure.batch.models.MetadataItem] :param stats: Utilization and resource usage statistics for the entire - lifetime of the pool. + lifetime of the pool. The statistics may not be immediately available. The + Batch service performs periodic roll-up of statistics. The typical delay + is about 30 minutes. :type stats: ~azure.batch.models.PoolStatistics """ diff --git a/azure-batch/azure/batch/models/cloud_pool_py3.py b/azure-batch/azure/batch/models/cloud_pool_py3.py index cbc8c0d22d7c..87ce5a9d5724 100644 --- a/azure-batch/azure/batch/models/cloud_pool_py3.py +++ b/azure-batch/azure/batch/models/cloud_pool_py3.py @@ -56,7 +56,7 @@ class CloudPool(Model): machines in a pool are the same size. For information about available sizes of virtual machines for Cloud Services pools (pools created with cloudServiceConfiguration), see Sizes for Cloud Services - (http://azure.microsoft.com/documentation/articles/cloud-services-sizes-specs/). + (https://azure.microsoft.com/documentation/articles/cloud-services-sizes-specs/). Batch supports all Cloud Services VM sizes except ExtraSmall, A1V2 and A2V2. For information about available VM sizes for pools using images from the Virtual Machines Marketplace (pools created with @@ -165,7 +165,9 @@ class CloudPool(Model): metadata. :type metadata: list[~azure.batch.models.MetadataItem] :param stats: Utilization and resource usage statistics for the entire - lifetime of the pool. + lifetime of the pool. The statistics may not be immediately available. The + Batch service performs periodic roll-up of statistics. The typical delay + is about 30 minutes. :type stats: ~azure.batch.models.PoolStatistics """ diff --git a/azure-batch/azure/batch/models/cloud_task.py b/azure-batch/azure/batch/models/cloud_task.py index d9cc213d4512..f6857c58a82b 100644 --- a/azure-batch/azure/batch/models/cloud_task.py +++ b/azure-batch/azure/batch/models/cloud_task.py @@ -15,6 +15,17 @@ class CloudTask(Model): """An Azure Batch task. + Batch will retry tasks when a recovery operation is triggered on a compute + node. Examples of recovery operations include (but are not limited to) when + an unhealthy compute node is rebooted or a compute node disappeared due to + host failure. Retries due to recovery operations are independent of and are + not counted against the maxTaskRetryCount. Even if the maxTaskRetryCount is + 0, an internal retry due to a recovery operation may occur. Because of + this, all tasks should be idempotent. This means tasks need to tolerate + being interrupted and restarted without causing any corruption or duplicate + data. The best practice for long running tasks is to use some form of + checkpointing. + :param id: A string that uniquely identifies the task within the job. The ID can contain any combination of alphanumeric characters including hyphens and underscores, and cannot contain more than 64 characters. @@ -59,7 +70,10 @@ class CloudTask(Model): take advantage of shell features such as environment variable expansion. If you want to take advantage of such features, you should invoke the shell in the command line, for example using "cmd /c MyCommand" in Windows - or "/bin/sh -c MyCommand" in Linux. + or "/bin/sh -c MyCommand" in Linux. If the command line refers to file + paths, it should use a relative path (relative to the task working + directory), or use the Batch provided environment variable + (https://docs.microsoft.com/en-us/azure/batch/batch-compute-node-environment-variables). :type command_line: str :param container_settings: The settings for the container under which the task runs. If the pool that will run this task has containerConfiguration diff --git a/azure-batch/azure/batch/models/cloud_task_py3.py b/azure-batch/azure/batch/models/cloud_task_py3.py index 9277bc1287da..312d49d6a34b 100644 --- a/azure-batch/azure/batch/models/cloud_task_py3.py +++ b/azure-batch/azure/batch/models/cloud_task_py3.py @@ -15,6 +15,17 @@ class CloudTask(Model): """An Azure Batch task. + Batch will retry tasks when a recovery operation is triggered on a compute + node. Examples of recovery operations include (but are not limited to) when + an unhealthy compute node is rebooted or a compute node disappeared due to + host failure. Retries due to recovery operations are independent of and are + not counted against the maxTaskRetryCount. Even if the maxTaskRetryCount is + 0, an internal retry due to a recovery operation may occur. Because of + this, all tasks should be idempotent. This means tasks need to tolerate + being interrupted and restarted without causing any corruption or duplicate + data. The best practice for long running tasks is to use some form of + checkpointing. + :param id: A string that uniquely identifies the task within the job. The ID can contain any combination of alphanumeric characters including hyphens and underscores, and cannot contain more than 64 characters. @@ -59,7 +70,10 @@ class CloudTask(Model): take advantage of shell features such as environment variable expansion. If you want to take advantage of such features, you should invoke the shell in the command line, for example using "cmd /c MyCommand" in Windows - or "/bin/sh -c MyCommand" in Linux. + or "/bin/sh -c MyCommand" in Linux. If the command line refers to file + paths, it should use a relative path (relative to the task working + directory), or use the Batch provided environment variable + (https://docs.microsoft.com/en-us/azure/batch/batch-compute-node-environment-variables). :type command_line: str :param container_settings: The settings for the container under which the task runs. If the pool that will run this task has containerConfiguration diff --git a/azure-batch/azure/batch/models/compute_node.py b/azure-batch/azure/batch/models/compute_node.py index fb172dde3b35..930e8bf6ec6b 100644 --- a/azure-batch/azure/batch/models/compute_node.py +++ b/azure-batch/azure/batch/models/compute_node.py @@ -56,7 +56,7 @@ class ComputeNode(Model): For information about available sizes of virtual machines for Cloud Services pools (pools created with cloudServiceConfiguration), see Sizes for Cloud Services - (http://azure.microsoft.com/documentation/articles/cloud-services-sizes-specs/). + (https://azure.microsoft.com/documentation/articles/cloud-services-sizes-specs/). Batch supports all Cloud Services VM sizes except ExtraSmall, A1V2 and A2V2. For information about available VM sizes for pools using images from the Virtual Machines Marketplace (pools created with diff --git a/azure-batch/azure/batch/models/compute_node_py3.py b/azure-batch/azure/batch/models/compute_node_py3.py index c0ddf2a717f1..8f0f6bf39ce1 100644 --- a/azure-batch/azure/batch/models/compute_node_py3.py +++ b/azure-batch/azure/batch/models/compute_node_py3.py @@ -56,7 +56,7 @@ class ComputeNode(Model): For information about available sizes of virtual machines for Cloud Services pools (pools created with cloudServiceConfiguration), see Sizes for Cloud Services - (http://azure.microsoft.com/documentation/articles/cloud-services-sizes-specs/). + (https://azure.microsoft.com/documentation/articles/cloud-services-sizes-specs/). Batch supports all Cloud Services VM sizes except ExtraSmall, A1V2 and A2V2. For information about available VM sizes for pools using images from the Virtual Machines Marketplace (pools created with diff --git a/azure-batch/azure/batch/models/job_manager_task.py b/azure-batch/azure/batch/models/job_manager_task.py index 9df2ebb82b08..efafa8227135 100644 --- a/azure-batch/azure/batch/models/job_manager_task.py +++ b/azure-batch/azure/batch/models/job_manager_task.py @@ -28,6 +28,16 @@ class JobManagerTask(Model): have priority over tasks in other jobs. Across jobs, only job level priorities are observed. For example, if a Job Manager in a priority 0 job needs to be restarted, it will not displace tasks of a priority 1 job. + Batch will retry tasks when a recovery operation is triggered on a compute + node. Examples of recovery operations include (but are not limited to) when + an unhealthy compute node is rebooted or a compute node disappeared due to + host failure. Retries due to recovery operations are independent of and are + not counted against the maxTaskRetryCount. Even if the maxTaskRetryCount is + 0, an internal retry due to a recovery operation may occur. Because of + this, all tasks should be idempotent. This means tasks need to tolerate + being interrupted and restarted without causing any corruption or duplicate + data. The best practice for long running tasks is to use some form of + checkpointing. All required parameters must be populated in order to send to Azure. @@ -45,7 +55,10 @@ class JobManagerTask(Model): advantage of shell features such as environment variable expansion. If you want to take advantage of such features, you should invoke the shell in the command line, for example using "cmd /c MyCommand" in Windows or - "/bin/sh -c MyCommand" in Linux. + "/bin/sh -c MyCommand" in Linux. If the command line refers to file paths, + it should use a relative path (relative to the task working directory), or + use the Batch provided environment variable + (https://docs.microsoft.com/en-us/azure/batch/batch-compute-node-environment-variables). :type command_line: str :param container_settings: The settings for the container under which the Job Manager task runs. If the pool that will run this task has @@ -120,7 +133,7 @@ class JobManagerTask(Model): :type authentication_token_settings: ~azure.batch.models.AuthenticationTokenSettings :param allow_low_priority_node: Whether the Job Manager task may run on a - low-priority compute node. The default value is false. + low-priority compute node. The default value is true. :type allow_low_priority_node: bool """ diff --git a/azure-batch/azure/batch/models/job_manager_task_py3.py b/azure-batch/azure/batch/models/job_manager_task_py3.py index 8f51704877c4..7167413b8777 100644 --- a/azure-batch/azure/batch/models/job_manager_task_py3.py +++ b/azure-batch/azure/batch/models/job_manager_task_py3.py @@ -28,6 +28,16 @@ class JobManagerTask(Model): have priority over tasks in other jobs. Across jobs, only job level priorities are observed. For example, if a Job Manager in a priority 0 job needs to be restarted, it will not displace tasks of a priority 1 job. + Batch will retry tasks when a recovery operation is triggered on a compute + node. Examples of recovery operations include (but are not limited to) when + an unhealthy compute node is rebooted or a compute node disappeared due to + host failure. Retries due to recovery operations are independent of and are + not counted against the maxTaskRetryCount. Even if the maxTaskRetryCount is + 0, an internal retry due to a recovery operation may occur. Because of + this, all tasks should be idempotent. This means tasks need to tolerate + being interrupted and restarted without causing any corruption or duplicate + data. The best practice for long running tasks is to use some form of + checkpointing. All required parameters must be populated in order to send to Azure. @@ -45,7 +55,10 @@ class JobManagerTask(Model): advantage of shell features such as environment variable expansion. If you want to take advantage of such features, you should invoke the shell in the command line, for example using "cmd /c MyCommand" in Windows or - "/bin/sh -c MyCommand" in Linux. + "/bin/sh -c MyCommand" in Linux. If the command line refers to file paths, + it should use a relative path (relative to the task working directory), or + use the Batch provided environment variable + (https://docs.microsoft.com/en-us/azure/batch/batch-compute-node-environment-variables). :type command_line: str :param container_settings: The settings for the container under which the Job Manager task runs. If the pool that will run this task has @@ -120,7 +133,7 @@ class JobManagerTask(Model): :type authentication_token_settings: ~azure.batch.models.AuthenticationTokenSettings :param allow_low_priority_node: Whether the Job Manager task may run on a - low-priority compute node. The default value is false. + low-priority compute node. The default value is true. :type allow_low_priority_node: bool """ diff --git a/azure-batch/azure/batch/models/job_preparation_task.py b/azure-batch/azure/batch/models/job_preparation_task.py index c65f28ccf3af..6d6025a1dd5a 100644 --- a/azure-batch/azure/batch/models/job_preparation_task.py +++ b/azure-batch/azure/batch/models/job_preparation_task.py @@ -32,7 +32,17 @@ class JobPreparationTask(Model): run again on the node before scheduling any other task of the job, if rerunOnNodeRebootAfterSuccess is true or if the Job Preparation task did not previously complete. If the compute node is reimaged, the Job - Preparation task is run again before scheduling any task of the job. + Preparation task is run again before scheduling any task of the job. Batch + will retry tasks when a recovery operation is triggered on a compute node. + Examples of recovery operations include (but are not limited to) when an + unhealthy compute node is rebooted or a compute node disappeared due to + host failure. Retries due to recovery operations are independent of and are + not counted against the maxTaskRetryCount. Even if the maxTaskRetryCount is + 0, an internal retry due to a recovery operation may occur. Because of + this, all tasks should be idempotent. This means tasks need to tolerate + being interrupted and restarted without causing any corruption or duplicate + data. The best practice for long running tasks is to use some form of + checkpointing. All required parameters must be populated in order to send to Azure. @@ -51,7 +61,10 @@ class JobPreparationTask(Model): take advantage of shell features such as environment variable expansion. If you want to take advantage of such features, you should invoke the shell in the command line, for example using "cmd /c MyCommand" in Windows - or "/bin/sh -c MyCommand" in Linux. + or "/bin/sh -c MyCommand" in Linux. If the command line refers to file + paths, it should use a relative path (relative to the task working + directory), or use the Batch provided environment variable + (https://docs.microsoft.com/en-us/azure/batch/batch-compute-node-environment-variables). :type command_line: str :param container_settings: The settings for the container under which the Job Preparation task runs. When this is specified, all directories diff --git a/azure-batch/azure/batch/models/job_preparation_task_py3.py b/azure-batch/azure/batch/models/job_preparation_task_py3.py index fb3462b31b7e..fb8017e1d5fd 100644 --- a/azure-batch/azure/batch/models/job_preparation_task_py3.py +++ b/azure-batch/azure/batch/models/job_preparation_task_py3.py @@ -32,7 +32,17 @@ class JobPreparationTask(Model): run again on the node before scheduling any other task of the job, if rerunOnNodeRebootAfterSuccess is true or if the Job Preparation task did not previously complete. If the compute node is reimaged, the Job - Preparation task is run again before scheduling any task of the job. + Preparation task is run again before scheduling any task of the job. Batch + will retry tasks when a recovery operation is triggered on a compute node. + Examples of recovery operations include (but are not limited to) when an + unhealthy compute node is rebooted or a compute node disappeared due to + host failure. Retries due to recovery operations are independent of and are + not counted against the maxTaskRetryCount. Even if the maxTaskRetryCount is + 0, an internal retry due to a recovery operation may occur. Because of + this, all tasks should be idempotent. This means tasks need to tolerate + being interrupted and restarted without causing any corruption or duplicate + data. The best practice for long running tasks is to use some form of + checkpointing. All required parameters must be populated in order to send to Azure. @@ -51,7 +61,10 @@ class JobPreparationTask(Model): take advantage of shell features such as environment variable expansion. If you want to take advantage of such features, you should invoke the shell in the command line, for example using "cmd /c MyCommand" in Windows - or "/bin/sh -c MyCommand" in Linux. + or "/bin/sh -c MyCommand" in Linux. If the command line refers to file + paths, it should use a relative path (relative to the task working + directory), or use the Batch provided environment variable + (https://docs.microsoft.com/en-us/azure/batch/batch-compute-node-environment-variables). :type command_line: str :param container_settings: The settings for the container under which the Job Preparation task runs. When this is specified, all directories diff --git a/azure-batch/azure/batch/models/job_release_task.py b/azure-batch/azure/batch/models/job_release_task.py index 1bb4f4be80d3..bd6214db1abd 100644 --- a/azure-batch/azure/batch/models/job_release_task.py +++ b/azure-batch/azure/batch/models/job_release_task.py @@ -50,7 +50,10 @@ class JobReleaseTask(Model): advantage of shell features such as environment variable expansion. If you want to take advantage of such features, you should invoke the shell in the command line, for example using "cmd /c MyCommand" in Windows or - "/bin/sh -c MyCommand" in Linux. + "/bin/sh -c MyCommand" in Linux. If the command line refers to file paths, + it should use a relative path (relative to the task working directory), or + use the Batch provided environment variable + (https://docs.microsoft.com/en-us/azure/batch/batch-compute-node-environment-variables). :type command_line: str :param container_settings: The settings for the container under which the Job Release task runs. When this is specified, all directories recursively diff --git a/azure-batch/azure/batch/models/job_release_task_py3.py b/azure-batch/azure/batch/models/job_release_task_py3.py index 8bd854296f20..18a35b8f2c81 100644 --- a/azure-batch/azure/batch/models/job_release_task_py3.py +++ b/azure-batch/azure/batch/models/job_release_task_py3.py @@ -50,7 +50,10 @@ class JobReleaseTask(Model): advantage of shell features such as environment variable expansion. If you want to take advantage of such features, you should invoke the shell in the command line, for example using "cmd /c MyCommand" in Windows or - "/bin/sh -c MyCommand" in Linux. + "/bin/sh -c MyCommand" in Linux. If the command line refers to file paths, + it should use a relative path (relative to the task working directory), or + use the Batch provided environment variable + (https://docs.microsoft.com/en-us/azure/batch/batch-compute-node-environment-variables). :type command_line: str :param container_settings: The settings for the container under which the Job Release task runs. When this is specified, all directories recursively diff --git a/azure-batch/azure/batch/models/node_counts.py b/azure-batch/azure/batch/models/node_counts.py index 836ead9c0a22..de54c0c01b1c 100644 --- a/azure-batch/azure/batch/models/node_counts.py +++ b/azure-batch/azure/batch/models/node_counts.py @@ -36,6 +36,9 @@ class NodeCounts(Model): :param start_task_failed: Required. The number of nodes in the startTaskFailed state. :type start_task_failed: int + :param leaving_pool: Required. The number of nodes in the leavingPool + state. + :type leaving_pool: int :param unknown: Required. The number of nodes in the unknown state. :type unknown: int :param unusable: Required. The number of nodes in the unusable state. @@ -57,6 +60,7 @@ class NodeCounts(Model): 'running': {'required': True}, 'starting': {'required': True}, 'start_task_failed': {'required': True}, + 'leaving_pool': {'required': True}, 'unknown': {'required': True}, 'unusable': {'required': True}, 'waiting_for_start_task': {'required': True}, @@ -73,6 +77,7 @@ class NodeCounts(Model): 'running': {'key': 'running', 'type': 'int'}, 'starting': {'key': 'starting', 'type': 'int'}, 'start_task_failed': {'key': 'startTaskFailed', 'type': 'int'}, + 'leaving_pool': {'key': 'leavingPool', 'type': 'int'}, 'unknown': {'key': 'unknown', 'type': 'int'}, 'unusable': {'key': 'unusable', 'type': 'int'}, 'waiting_for_start_task': {'key': 'waitingForStartTask', 'type': 'int'}, @@ -90,6 +95,7 @@ def __init__(self, **kwargs): self.running = kwargs.get('running', None) self.starting = kwargs.get('starting', None) self.start_task_failed = kwargs.get('start_task_failed', None) + self.leaving_pool = kwargs.get('leaving_pool', None) self.unknown = kwargs.get('unknown', None) self.unusable = kwargs.get('unusable', None) self.waiting_for_start_task = kwargs.get('waiting_for_start_task', None) diff --git a/azure-batch/azure/batch/models/node_counts_py3.py b/azure-batch/azure/batch/models/node_counts_py3.py index 5e158cb6e1be..bfeca712f86b 100644 --- a/azure-batch/azure/batch/models/node_counts_py3.py +++ b/azure-batch/azure/batch/models/node_counts_py3.py @@ -36,6 +36,9 @@ class NodeCounts(Model): :param start_task_failed: Required. The number of nodes in the startTaskFailed state. :type start_task_failed: int + :param leaving_pool: Required. The number of nodes in the leavingPool + state. + :type leaving_pool: int :param unknown: Required. The number of nodes in the unknown state. :type unknown: int :param unusable: Required. The number of nodes in the unusable state. @@ -57,6 +60,7 @@ class NodeCounts(Model): 'running': {'required': True}, 'starting': {'required': True}, 'start_task_failed': {'required': True}, + 'leaving_pool': {'required': True}, 'unknown': {'required': True}, 'unusable': {'required': True}, 'waiting_for_start_task': {'required': True}, @@ -73,13 +77,14 @@ class NodeCounts(Model): 'running': {'key': 'running', 'type': 'int'}, 'starting': {'key': 'starting', 'type': 'int'}, 'start_task_failed': {'key': 'startTaskFailed', 'type': 'int'}, + 'leaving_pool': {'key': 'leavingPool', 'type': 'int'}, 'unknown': {'key': 'unknown', 'type': 'int'}, 'unusable': {'key': 'unusable', 'type': 'int'}, 'waiting_for_start_task': {'key': 'waitingForStartTask', 'type': 'int'}, 'total': {'key': 'total', 'type': 'int'}, } - def __init__(self, *, creating: int, idle: int, offline: int, preempted: int, rebooting: int, reimaging: int, running: int, starting: int, start_task_failed: int, unknown: int, unusable: int, waiting_for_start_task: int, total: int, **kwargs) -> None: + def __init__(self, *, creating: int, idle: int, offline: int, preempted: int, rebooting: int, reimaging: int, running: int, starting: int, start_task_failed: int, leaving_pool: int, unknown: int, unusable: int, waiting_for_start_task: int, total: int, **kwargs) -> None: super(NodeCounts, self).__init__(**kwargs) self.creating = creating self.idle = idle @@ -90,6 +95,7 @@ def __init__(self, *, creating: int, idle: int, offline: int, preempted: int, re self.running = running self.starting = starting self.start_task_failed = start_task_failed + self.leaving_pool = leaving_pool self.unknown = unknown self.unusable = unusable self.waiting_for_start_task = waiting_for_start_task diff --git a/azure-batch/azure/batch/models/pool_add_parameter.py b/azure-batch/azure/batch/models/pool_add_parameter.py index f66caf9f16f3..7d21c67bc843 100644 --- a/azure-batch/azure/batch/models/pool_add_parameter.py +++ b/azure-batch/azure/batch/models/pool_add_parameter.py @@ -31,7 +31,7 @@ class PoolAddParameter(Model): virtual machines in a pool are the same size. For information about available sizes of virtual machines for Cloud Services pools (pools created with cloudServiceConfiguration), see Sizes for Cloud Services - (http://azure.microsoft.com/documentation/articles/cloud-services-sizes-specs/). + (https://azure.microsoft.com/documentation/articles/cloud-services-sizes-specs/). Batch supports all Cloud Services VM sizes except ExtraSmall, A1V2 and A2V2. For information about available VM sizes for pools using images from the Virtual Machines Marketplace (pools created with diff --git a/azure-batch/azure/batch/models/pool_add_parameter_py3.py b/azure-batch/azure/batch/models/pool_add_parameter_py3.py index 1a6d0da0ddd7..170963fa94ce 100644 --- a/azure-batch/azure/batch/models/pool_add_parameter_py3.py +++ b/azure-batch/azure/batch/models/pool_add_parameter_py3.py @@ -31,7 +31,7 @@ class PoolAddParameter(Model): virtual machines in a pool are the same size. For information about available sizes of virtual machines for Cloud Services pools (pools created with cloudServiceConfiguration), see Sizes for Cloud Services - (http://azure.microsoft.com/documentation/articles/cloud-services-sizes-specs/). + (https://azure.microsoft.com/documentation/articles/cloud-services-sizes-specs/). Batch supports all Cloud Services VM sizes except ExtraSmall, A1V2 and A2V2. For information about available VM sizes for pools using images from the Virtual Machines Marketplace (pools created with diff --git a/azure-batch/azure/batch/models/pool_specification.py b/azure-batch/azure/batch/models/pool_specification.py index 76272afd7efc..f3262aa2bfd2 100644 --- a/azure-batch/azure/batch/models/pool_specification.py +++ b/azure-batch/azure/batch/models/pool_specification.py @@ -25,7 +25,7 @@ class PoolSpecification(Model): All virtual machines in a pool are the same size. For information about available sizes of virtual machines for Cloud Services pools (pools created with cloudServiceConfiguration), see Sizes for Cloud Services - (http://azure.microsoft.com/documentation/articles/cloud-services-sizes-specs/). + (https://azure.microsoft.com/documentation/articles/cloud-services-sizes-specs/). Batch supports all Cloud Services VM sizes except ExtraSmall, A1V2 and A2V2. For information about available VM sizes for pools using images from the Virtual Machines Marketplace (pools created with diff --git a/azure-batch/azure/batch/models/pool_specification_py3.py b/azure-batch/azure/batch/models/pool_specification_py3.py index eb12b15da848..0adee1101104 100644 --- a/azure-batch/azure/batch/models/pool_specification_py3.py +++ b/azure-batch/azure/batch/models/pool_specification_py3.py @@ -25,7 +25,7 @@ class PoolSpecification(Model): All virtual machines in a pool are the same size. For information about available sizes of virtual machines for Cloud Services pools (pools created with cloudServiceConfiguration), see Sizes for Cloud Services - (http://azure.microsoft.com/documentation/articles/cloud-services-sizes-specs/). + (https://azure.microsoft.com/documentation/articles/cloud-services-sizes-specs/). Batch supports all Cloud Services VM sizes except ExtraSmall, A1V2 and A2V2. For information about available VM sizes for pools using images from the Virtual Machines Marketplace (pools created with diff --git a/azure-batch/azure/batch/models/start_task.py b/azure-batch/azure/batch/models/start_task.py index bb99fe2a17ad..f5e303863367 100644 --- a/azure-batch/azure/batch/models/start_task.py +++ b/azure-batch/azure/batch/models/start_task.py @@ -16,6 +16,17 @@ class StartTask(Model): """A task which is run when a compute node joins a pool in the Azure Batch service, or when the compute node is rebooted or reimaged. + Batch will retry tasks when a recovery operation is triggered on a compute + node. Examples of recovery operations include (but are not limited to) when + an unhealthy compute node is rebooted or a compute node disappeared due to + host failure. Retries due to recovery operations are independent of and are + not counted against the maxTaskRetryCount. Even if the maxTaskRetryCount is + 0, an internal retry due to a recovery operation may occur. Because of + this, all tasks should be idempotent. This means tasks need to tolerate + being interrupted and restarted without causing any corruption or duplicate + data. The best practice for long running tasks is to use some form of + checkpointing. + All required parameters must be populated in order to send to Azure. :param command_line: Required. The command line of the start task. The @@ -23,7 +34,10 @@ class StartTask(Model): advantage of shell features such as environment variable expansion. If you want to take advantage of such features, you should invoke the shell in the command line, for example using "cmd /c MyCommand" in Windows or - "/bin/sh -c MyCommand" in Linux. + "/bin/sh -c MyCommand" in Linux. If the command line refers to file paths, + it should use a relative path (relative to the task working directory), or + use the Batch provided environment variable + (https://docs.microsoft.com/en-us/azure/batch/batch-compute-node-environment-variables). :type command_line: str :param container_settings: The settings for the container under which the start task runs. When this is specified, all directories recursively below diff --git a/azure-batch/azure/batch/models/start_task_py3.py b/azure-batch/azure/batch/models/start_task_py3.py index 995c2809d30d..1f839732ff31 100644 --- a/azure-batch/azure/batch/models/start_task_py3.py +++ b/azure-batch/azure/batch/models/start_task_py3.py @@ -16,6 +16,17 @@ class StartTask(Model): """A task which is run when a compute node joins a pool in the Azure Batch service, or when the compute node is rebooted or reimaged. + Batch will retry tasks when a recovery operation is triggered on a compute + node. Examples of recovery operations include (but are not limited to) when + an unhealthy compute node is rebooted or a compute node disappeared due to + host failure. Retries due to recovery operations are independent of and are + not counted against the maxTaskRetryCount. Even if the maxTaskRetryCount is + 0, an internal retry due to a recovery operation may occur. Because of + this, all tasks should be idempotent. This means tasks need to tolerate + being interrupted and restarted without causing any corruption or duplicate + data. The best practice for long running tasks is to use some form of + checkpointing. + All required parameters must be populated in order to send to Azure. :param command_line: Required. The command line of the start task. The @@ -23,7 +34,10 @@ class StartTask(Model): advantage of shell features such as environment variable expansion. If you want to take advantage of such features, you should invoke the shell in the command line, for example using "cmd /c MyCommand" in Windows or - "/bin/sh -c MyCommand" in Linux. + "/bin/sh -c MyCommand" in Linux. If the command line refers to file paths, + it should use a relative path (relative to the task working directory), or + use the Batch provided environment variable + (https://docs.microsoft.com/en-us/azure/batch/batch-compute-node-environment-variables). :type command_line: str :param container_settings: The settings for the container under which the start task runs. When this is specified, all directories recursively below diff --git a/azure-batch/azure/batch/models/task_add_parameter.py b/azure-batch/azure/batch/models/task_add_parameter.py index 746afddd2595..6d9facc6be98 100644 --- a/azure-batch/azure/batch/models/task_add_parameter.py +++ b/azure-batch/azure/batch/models/task_add_parameter.py @@ -15,6 +15,17 @@ class TaskAddParameter(Model): """An Azure Batch task to add. + Batch will retry tasks when a recovery operation is triggered on a compute + node. Examples of recovery operations include (but are not limited to) when + an unhealthy compute node is rebooted or a compute node disappeared due to + host failure. Retries due to recovery operations are independent of and are + not counted against the maxTaskRetryCount. Even if the maxTaskRetryCount is + 0, an internal retry due to a recovery operation may occur. Because of + this, all tasks should be idempotent. This means tasks need to tolerate + being interrupted and restarted without causing any corruption or duplicate + data. The best practice for long running tasks is to use some form of + checkpointing. + All required parameters must be populated in order to send to Azure. :param id: Required. A string that uniquely identifies the task within the @@ -34,7 +45,10 @@ class TaskAddParameter(Model): and therefore cannot take advantage of shell features such as environment variable expansion. If you want to take advantage of such features, you should invoke the shell in the command line, for example using "cmd /c - MyCommand" in Windows or "/bin/sh -c MyCommand" in Linux. + MyCommand" in Windows or "/bin/sh -c MyCommand" in Linux. If the command + line refers to file paths, it should use a relative path (relative to the + task working directory), or use the Batch provided environment variable + (https://docs.microsoft.com/en-us/azure/batch/batch-compute-node-environment-variables). :type command_line: str :param container_settings: The settings for the container under which the task runs. If the pool that will run this task has containerConfiguration diff --git a/azure-batch/azure/batch/models/task_add_parameter_py3.py b/azure-batch/azure/batch/models/task_add_parameter_py3.py index d5be52ac15a0..ddf83f0b701b 100644 --- a/azure-batch/azure/batch/models/task_add_parameter_py3.py +++ b/azure-batch/azure/batch/models/task_add_parameter_py3.py @@ -15,6 +15,17 @@ class TaskAddParameter(Model): """An Azure Batch task to add. + Batch will retry tasks when a recovery operation is triggered on a compute + node. Examples of recovery operations include (but are not limited to) when + an unhealthy compute node is rebooted or a compute node disappeared due to + host failure. Retries due to recovery operations are independent of and are + not counted against the maxTaskRetryCount. Even if the maxTaskRetryCount is + 0, an internal retry due to a recovery operation may occur. Because of + this, all tasks should be idempotent. This means tasks need to tolerate + being interrupted and restarted without causing any corruption or duplicate + data. The best practice for long running tasks is to use some form of + checkpointing. + All required parameters must be populated in order to send to Azure. :param id: Required. A string that uniquely identifies the task within the @@ -34,7 +45,10 @@ class TaskAddParameter(Model): and therefore cannot take advantage of shell features such as environment variable expansion. If you want to take advantage of such features, you should invoke the shell in the command line, for example using "cmd /c - MyCommand" in Windows or "/bin/sh -c MyCommand" in Linux. + MyCommand" in Windows or "/bin/sh -c MyCommand" in Linux. If the command + line refers to file paths, it should use a relative path (relative to the + task working directory), or use the Batch provided environment variable + (https://docs.microsoft.com/en-us/azure/batch/batch-compute-node-environment-variables). :type command_line: str :param container_settings: The settings for the container under which the task runs. If the pool that will run this task has containerConfiguration diff --git a/azure-batch/azure/batch/models/task_constraints.py b/azure-batch/azure/batch/models/task_constraints.py index 502ceeeb0332..22898fad6ff2 100644 --- a/azure-batch/azure/batch/models/task_constraints.py +++ b/azure-batch/azure/batch/models/task_constraints.py @@ -28,12 +28,15 @@ class TaskConstraints(Model): :type retention_time: timedelta :param max_task_retry_count: The maximum number of times the task may be retried. The Batch service retries a task if its exit code is nonzero. - Note that this value specifically controls the number of retries. The - Batch service will try the task once, and may then retry up to this limit. - For example, if the maximum retry count is 3, Batch tries the task up to 4 - times (one initial try and 3 retries). If the maximum retry count is 0, - the Batch service does not retry the task. If the maximum retry count is - -1, the Batch service retries the task without limit. + Note that this value specifically controls the number of retries for the + task executable due to a nonzero exit code. The Batch service will try the + task once, and may then retry up to this limit. For example, if the + maximum retry count is 3, Batch tries the task up to 4 times (one initial + try and 3 retries). If the maximum retry count is 0, the Batch service + does not retry the task after the first attempt. If the maximum retry + count is -1, the Batch service retries the task without limit. Resource + files and application packages are only downloaded again if the task is + retried on a new compute node. :type max_task_retry_count: int """ diff --git a/azure-batch/azure/batch/models/task_constraints_py3.py b/azure-batch/azure/batch/models/task_constraints_py3.py index 9ee03677f75d..2070d09669c6 100644 --- a/azure-batch/azure/batch/models/task_constraints_py3.py +++ b/azure-batch/azure/batch/models/task_constraints_py3.py @@ -28,12 +28,15 @@ class TaskConstraints(Model): :type retention_time: timedelta :param max_task_retry_count: The maximum number of times the task may be retried. The Batch service retries a task if its exit code is nonzero. - Note that this value specifically controls the number of retries. The - Batch service will try the task once, and may then retry up to this limit. - For example, if the maximum retry count is 3, Batch tries the task up to 4 - times (one initial try and 3 retries). If the maximum retry count is 0, - the Batch service does not retry the task. If the maximum retry count is - -1, the Batch service retries the task without limit. + Note that this value specifically controls the number of retries for the + task executable due to a nonzero exit code. The Batch service will try the + task once, and may then retry up to this limit. For example, if the + maximum retry count is 3, Batch tries the task up to 4 times (one initial + try and 3 retries). If the maximum retry count is 0, the Batch service + does not retry the task after the first attempt. If the maximum retry + count is -1, the Batch service retries the task without limit. Resource + files and application packages are only downloaded again if the task is + retried on a new compute node. :type max_task_retry_count: int """ diff --git a/azure-batch/azure/batch/models/user_identity.py b/azure-batch/azure/batch/models/user_identity.py index 1301c24e4418..b75dfd7306b3 100644 --- a/azure-batch/azure/batch/models/user_identity.py +++ b/azure-batch/azure/batch/models/user_identity.py @@ -15,7 +15,10 @@ class UserIdentity(Model): """The definition of the user identity under which the task is run. - Specify either the userName or autoUser property, but not both. + Specify either the userName or autoUser property, but not both. On + CloudServiceConfiguration pools, this user is logged in with the + INTERACTIVE flag. On Windows VirtualMachineConfiguration pools, this user + is logged in with the BATCH flag. :param user_name: The name of the user identity under which the task is run. The userName and autoUser properties are mutually exclusive; you must diff --git a/azure-batch/azure/batch/models/user_identity_py3.py b/azure-batch/azure/batch/models/user_identity_py3.py index 379928526110..e566f58cd51c 100644 --- a/azure-batch/azure/batch/models/user_identity_py3.py +++ b/azure-batch/azure/batch/models/user_identity_py3.py @@ -15,7 +15,10 @@ class UserIdentity(Model): """The definition of the user identity under which the task is run. - Specify either the userName or autoUser property, but not both. + Specify either the userName or autoUser property, but not both. On + CloudServiceConfiguration pools, this user is logged in with the + INTERACTIVE flag. On Windows VirtualMachineConfiguration pools, this user + is logged in with the BATCH flag. :param user_name: The name of the user identity under which the task is run. The userName and autoUser properties are mutually exclusive; you must diff --git a/azure-batch/azure/batch/operations/job_operations.py b/azure-batch/azure/batch/operations/job_operations.py index a913c85588fb..ae900846e26e 100644 --- a/azure-batch/azure/batch/operations/job_operations.py +++ b/azure-batch/azure/batch/operations/job_operations.py @@ -43,7 +43,9 @@ def get_all_lifetime_statistics( Statistics are aggregated across all jobs that have ever existed in the account, from account creation to the last update time of the - statistics. + statistics. The statistics may not be immediately available. The Batch + service performs periodic roll-up of statistics. The typical delay is + about 30 minutes. :param job_get_all_lifetime_statistics_options: Additional parameters for the operation diff --git a/azure-batch/azure/batch/operations/pool_operations.py b/azure-batch/azure/batch/operations/pool_operations.py index 0cef8de72862..8c48f54e4edd 100644 --- a/azure-batch/azure/batch/operations/pool_operations.py +++ b/azure-batch/azure/batch/operations/pool_operations.py @@ -156,7 +156,9 @@ def get_all_lifetime_statistics( Statistics are aggregated across all pools that have ever existed in the account, from account creation to the last update time of the - statistics. + statistics. The statistics may not be immediately available. The Batch + service performs periodic roll-up of statistics. The typical delay is + about 30 minutes. :param pool_get_all_lifetime_statistics_options: Additional parameters for the operation diff --git a/azure-batch/build.json b/azure-batch/build.json index 1f3e5361005e..3cf6b3a9732d 100644 --- a/azure-batch/build.json +++ b/azure-batch/build.json @@ -4,15 +4,16 @@ "resolvedInfo": null, "packageMetadata": { "name": "@microsoft.azure/autorest-core", - "version": "2.0.4259", + "version": "2.0.4275", "engines": { "node": ">=7.10.0" }, - "dependencies": {}, + "dependencies": { + "typescript": "2.6.2" + }, "optionalDependencies": {}, "devDependencies": { "@types/commonmark": "^0.27.0", - "@types/js-yaml": "^3.10.0", "@types/jsonpath": "^0.1.29", "@types/node": "^8.0.53", "@types/source-map": "0.5.0", @@ -28,63 +29,63 @@ "bundleDependencies": false, "peerDependencies": {}, "deprecated": false, - "_resolved": "/tmp/.autorest/@microsoft.azure_autorest-core@2.0.4259/node_modules/@microsoft.azure/autorest-core", + "_resolved": "/root/.autorest/@microsoft.azure_autorest-core@2.0.4275/node_modules/@microsoft.azure/autorest-core", "_integrity": null, - "_shasum": "d7d762d9e7ed3b5c097aeb8d9a5804d768adee6a", + "_shasum": "fdcfc270c8be0ddec8a8996951d975b9c76c333d", "_shrinkwrap": null, "bin": { "autorest-core": "./dist/app.js", "autorest-language-service": "dist/language-service/language-service.js" }, - "_id": "@microsoft.azure/autorest-core@2.0.4259", - "_from": "file:/tmp/.autorest/@microsoft.azure_autorest-core@2.0.4259/node_modules/@microsoft.azure/autorest-core", + "_id": "@microsoft.azure/autorest-core@2.0.4275", + "_from": "file:/root/.autorest/@microsoft.azure_autorest-core@2.0.4275/node_modules/@microsoft.azure/autorest-core", "_requested": { "type": "directory", - "where": "/tmp/.autorest/@microsoft.azure_autorest-core@2.0.4259/node_modules/@microsoft.azure/autorest-core", - "raw": "/tmp/.autorest/@microsoft.azure_autorest-core@2.0.4259/node_modules/@microsoft.azure/autorest-core", - "rawSpec": "/tmp/.autorest/@microsoft.azure_autorest-core@2.0.4259/node_modules/@microsoft.azure/autorest-core", - "saveSpec": "file:/tmp/.autorest/@microsoft.azure_autorest-core@2.0.4259/node_modules/@microsoft.azure/autorest-core", - "fetchSpec": "/tmp/.autorest/@microsoft.azure_autorest-core@2.0.4259/node_modules/@microsoft.azure/autorest-core" + "where": "/root/.autorest/@microsoft.azure_autorest-core@2.0.4275/node_modules/@microsoft.azure/autorest-core", + "raw": "/root/.autorest/@microsoft.azure_autorest-core@2.0.4275/node_modules/@microsoft.azure/autorest-core", + "rawSpec": "/root/.autorest/@microsoft.azure_autorest-core@2.0.4275/node_modules/@microsoft.azure/autorest-core", + "saveSpec": "file:/root/.autorest/@microsoft.azure_autorest-core@2.0.4275/node_modules/@microsoft.azure/autorest-core", + "fetchSpec": "/root/.autorest/@microsoft.azure_autorest-core@2.0.4275/node_modules/@microsoft.azure/autorest-core" }, - "_spec": "/tmp/.autorest/@microsoft.azure_autorest-core@2.0.4259/node_modules/@microsoft.azure/autorest-core", - "_where": "/tmp/.autorest/@microsoft.azure_autorest-core@2.0.4259/node_modules/@microsoft.azure/autorest-core" + "_spec": "/root/.autorest/@microsoft.azure_autorest-core@2.0.4275/node_modules/@microsoft.azure/autorest-core", + "_where": "/root/.autorest/@microsoft.azure_autorest-core@2.0.4275/node_modules/@microsoft.azure/autorest-core" }, "extensionManager": { - "installationPath": "/tmp/.autorest", + "installationPath": "/root/.autorest", "sharedLock": { - "name": "/tmp/.autorest", + "name": "/root/.autorest", "exclusiveLock": { - "name": "_tmp_.autorest.exclusive-lock", + "name": "_root_.autorest.exclusive-lock", "options": { - "port": 5245, - "host": "2130716018", + "port": 45234, + "host": "2130706813", "exclusive": true }, - "pipe": "/tmp/pipe__tmp_.autorest.exclusive-lock:5245" + "pipe": "/tmp/pipe__root_.autorest.exclusive-lock:45234" }, "busyLock": { - "name": "_tmp_.autorest.busy-lock", + "name": "_root_.autorest.busy-lock", "options": { - "port": 46071, - "host": "2130735832", + "port": 37199, + "host": "2130756895", "exclusive": true }, - "pipe": "/tmp/pipe__tmp_.autorest.busy-lock:46071" + "pipe": "/tmp/pipe__root_.autorest.busy-lock:37199" }, "personalLock": { - "name": "_tmp_.autorest.4890.15425350853.personal-lock", + "name": "_root_.autorest.8216.386719855544.personal-lock", "options": { - "port": 22892, - "host": "2130771569", + "port": 63645, + "host": "2130724149", "exclusive": true }, - "pipe": "/tmp/pipe__tmp_.autorest.4890.15425350853.personal-lock:22892" + "pipe": "/tmp/pipe__root_.autorest.8216.386719855544.personal-lock:63645" }, - "file": "/tmp/_tmp_.autorest.lock" + "file": "/tmp/_root_.autorest.lock" }, "dotnetPath": "/root/.dotnet" }, - "installationPath": "/tmp/.autorest" + "installationPath": "/root/.autorest" }, { "resolvedInfo": null, @@ -114,60 +115,60 @@ "bundleDependencies": false, "peerDependencies": {}, "deprecated": false, - "_resolved": "/tmp/.autorest/@microsoft.azure_autorest.modeler@2.0.21/node_modules/@microsoft.azure/autorest.modeler", + "_resolved": "/root/.autorest/@microsoft.azure_autorest.modeler@2.0.21/node_modules/@microsoft.azure/autorest.modeler", "_integrity": null, "_shasum": "3ce7d3939124b31830be15e5de99b9b7768afb90", "_shrinkwrap": null, "bin": null, "_id": "@microsoft.azure/autorest.modeler@2.0.21", - "_from": "file:/tmp/.autorest/@microsoft.azure_autorest.modeler@2.0.21/node_modules/@microsoft.azure/autorest.modeler", + "_from": "file:/root/.autorest/@microsoft.azure_autorest.modeler@2.0.21/node_modules/@microsoft.azure/autorest.modeler", "_requested": { "type": "directory", - "where": "/tmp/.autorest/@microsoft.azure_autorest.modeler@2.0.21/node_modules/@microsoft.azure/autorest.modeler", - "raw": "/tmp/.autorest/@microsoft.azure_autorest.modeler@2.0.21/node_modules/@microsoft.azure/autorest.modeler", - "rawSpec": "/tmp/.autorest/@microsoft.azure_autorest.modeler@2.0.21/node_modules/@microsoft.azure/autorest.modeler", - "saveSpec": "file:/tmp/.autorest/@microsoft.azure_autorest.modeler@2.0.21/node_modules/@microsoft.azure/autorest.modeler", - "fetchSpec": "/tmp/.autorest/@microsoft.azure_autorest.modeler@2.0.21/node_modules/@microsoft.azure/autorest.modeler" + "where": "/root/.autorest/@microsoft.azure_autorest.modeler@2.0.21/node_modules/@microsoft.azure/autorest.modeler", + "raw": "/root/.autorest/@microsoft.azure_autorest.modeler@2.0.21/node_modules/@microsoft.azure/autorest.modeler", + "rawSpec": "/root/.autorest/@microsoft.azure_autorest.modeler@2.0.21/node_modules/@microsoft.azure/autorest.modeler", + "saveSpec": "file:/root/.autorest/@microsoft.azure_autorest.modeler@2.0.21/node_modules/@microsoft.azure/autorest.modeler", + "fetchSpec": "/root/.autorest/@microsoft.azure_autorest.modeler@2.0.21/node_modules/@microsoft.azure/autorest.modeler" }, - "_spec": "/tmp/.autorest/@microsoft.azure_autorest.modeler@2.0.21/node_modules/@microsoft.azure/autorest.modeler", - "_where": "/tmp/.autorest/@microsoft.azure_autorest.modeler@2.0.21/node_modules/@microsoft.azure/autorest.modeler" + "_spec": "/root/.autorest/@microsoft.azure_autorest.modeler@2.0.21/node_modules/@microsoft.azure/autorest.modeler", + "_where": "/root/.autorest/@microsoft.azure_autorest.modeler@2.0.21/node_modules/@microsoft.azure/autorest.modeler" }, "extensionManager": { - "installationPath": "/tmp/.autorest", + "installationPath": "/root/.autorest", "sharedLock": { - "name": "/tmp/.autorest", + "name": "/root/.autorest", "exclusiveLock": { - "name": "_tmp_.autorest.exclusive-lock", + "name": "_root_.autorest.exclusive-lock", "options": { - "port": 5245, - "host": "2130716018", + "port": 45234, + "host": "2130706813", "exclusive": true }, - "pipe": "/tmp/pipe__tmp_.autorest.exclusive-lock:5245" + "pipe": "/tmp/pipe__root_.autorest.exclusive-lock:45234" }, "busyLock": { - "name": "_tmp_.autorest.busy-lock", + "name": "_root_.autorest.busy-lock", "options": { - "port": 46071, - "host": "2130735832", + "port": 37199, + "host": "2130756895", "exclusive": true }, - "pipe": "/tmp/pipe__tmp_.autorest.busy-lock:46071" + "pipe": "/tmp/pipe__root_.autorest.busy-lock:37199" }, "personalLock": { - "name": "_tmp_.autorest.4890.15425350853.personal-lock", + "name": "_root_.autorest.8216.386719855544.personal-lock", "options": { - "port": 22892, - "host": "2130771569", + "port": 63645, + "host": "2130724149", "exclusive": true }, - "pipe": "/tmp/pipe__tmp_.autorest.4890.15425350853.personal-lock:22892" + "pipe": "/tmp/pipe__root_.autorest.8216.386719855544.personal-lock:63645" }, - "file": "/tmp/_tmp_.autorest.lock" + "file": "/tmp/_root_.autorest.lock" }, "dotnetPath": "/root/.dotnet" }, - "installationPath": "/tmp/.autorest" + "installationPath": "/root/.autorest" }, { "resolvedInfo": null, @@ -199,157 +200,72 @@ "bundleDependencies": false, "peerDependencies": {}, "deprecated": false, - "_resolved": "/tmp/.autorest/@microsoft.azure_autorest.modeler@2.3.44/node_modules/@microsoft.azure/autorest.modeler", + "_resolved": "/root/.autorest/@microsoft.azure_autorest.modeler@2.3.44/node_modules/@microsoft.azure/autorest.modeler", "_integrity": null, "_shasum": "9b5a880a77467be33a77f002f03230d3ccc21266", "_shrinkwrap": null, "bin": null, "_id": "@microsoft.azure/autorest.modeler@2.3.44", - "_from": "file:/tmp/.autorest/@microsoft.azure_autorest.modeler@2.3.44/node_modules/@microsoft.azure/autorest.modeler", - "_requested": { - "type": "directory", - "where": "/tmp/.autorest/@microsoft.azure_autorest.modeler@2.3.44/node_modules/@microsoft.azure/autorest.modeler", - "raw": "/tmp/.autorest/@microsoft.azure_autorest.modeler@2.3.44/node_modules/@microsoft.azure/autorest.modeler", - "rawSpec": "/tmp/.autorest/@microsoft.azure_autorest.modeler@2.3.44/node_modules/@microsoft.azure/autorest.modeler", - "saveSpec": "file:/tmp/.autorest/@microsoft.azure_autorest.modeler@2.3.44/node_modules/@microsoft.azure/autorest.modeler", - "fetchSpec": "/tmp/.autorest/@microsoft.azure_autorest.modeler@2.3.44/node_modules/@microsoft.azure/autorest.modeler" - }, - "_spec": "/tmp/.autorest/@microsoft.azure_autorest.modeler@2.3.44/node_modules/@microsoft.azure/autorest.modeler", - "_where": "/tmp/.autorest/@microsoft.azure_autorest.modeler@2.3.44/node_modules/@microsoft.azure/autorest.modeler" - }, - "extensionManager": { - "installationPath": "/tmp/.autorest", - "sharedLock": { - "name": "/tmp/.autorest", - "exclusiveLock": { - "name": "_tmp_.autorest.exclusive-lock", - "options": { - "port": 5245, - "host": "2130716018", - "exclusive": true - }, - "pipe": "/tmp/pipe__tmp_.autorest.exclusive-lock:5245" - }, - "busyLock": { - "name": "_tmp_.autorest.busy-lock", - "options": { - "port": 46071, - "host": "2130735832", - "exclusive": true - }, - "pipe": "/tmp/pipe__tmp_.autorest.busy-lock:46071" - }, - "personalLock": { - "name": "_tmp_.autorest.4890.15425350853.personal-lock", - "options": { - "port": 22892, - "host": "2130771569", - "exclusive": true - }, - "pipe": "/tmp/pipe__tmp_.autorest.4890.15425350853.personal-lock:22892" - }, - "file": "/tmp/_tmp_.autorest.lock" - }, - "dotnetPath": "/root/.dotnet" - }, - "installationPath": "/tmp/.autorest" - }, - { - "resolvedInfo": null, - "packageMetadata": { - "name": "@microsoft.azure/autorest.python", - "version": "2.1.40", - "dependencies": { - "dotnet-2.0.0": "^1.4.4" - }, - "optionalDependencies": {}, - "devDependencies": { - "@microsoft.azure/autorest.testserver": "^2.4.0", - "autorest": "^2.0.4203", - "coffee-script": "^1.11.1", - "dotnet-sdk-2.0.0": "^1.4.4", - "gulp": "^3.9.1", - "gulp-filter": "^5.0.0", - "gulp-line-ending-corrector": "^1.0.1", - "iced-coffee-script": "^108.0.11", - "marked": "^0.3.6", - "marked-terminal": "^2.0.0", - "moment": "^2.17.1", - "run-sequence": "*", - "shx": "^0.2.2", - "through2-parallel": "^0.1.3", - "yargs": "^8.0.2", - "yarn": "^1.0.2" - }, - "bundleDependencies": false, - "peerDependencies": {}, - "deprecated": false, - "_resolved": "/tmp/.autorest/@microsoft.azure_autorest.python@2.1.40/node_modules/@microsoft.azure/autorest.python", - "_integrity": null, - "_shasum": "9b3f08c892d725ac571b3a7dc8f781d76da64397", - "_shrinkwrap": null, - "bin": null, - "_id": "@microsoft.azure/autorest.python@2.1.40", - "_from": "file:/tmp/.autorest/@microsoft.azure_autorest.python@2.1.40/node_modules/@microsoft.azure/autorest.python", + "_from": "file:/root/.autorest/@microsoft.azure_autorest.modeler@2.3.44/node_modules/@microsoft.azure/autorest.modeler", "_requested": { "type": "directory", - "where": "/tmp/.autorest/@microsoft.azure_autorest.python@2.1.40/node_modules/@microsoft.azure/autorest.python", - "raw": "/tmp/.autorest/@microsoft.azure_autorest.python@2.1.40/node_modules/@microsoft.azure/autorest.python", - "rawSpec": "/tmp/.autorest/@microsoft.azure_autorest.python@2.1.40/node_modules/@microsoft.azure/autorest.python", - "saveSpec": "file:/tmp/.autorest/@microsoft.azure_autorest.python@2.1.40/node_modules/@microsoft.azure/autorest.python", - "fetchSpec": "/tmp/.autorest/@microsoft.azure_autorest.python@2.1.40/node_modules/@microsoft.azure/autorest.python" + "where": "/root/.autorest/@microsoft.azure_autorest.modeler@2.3.44/node_modules/@microsoft.azure/autorest.modeler", + "raw": "/root/.autorest/@microsoft.azure_autorest.modeler@2.3.44/node_modules/@microsoft.azure/autorest.modeler", + "rawSpec": "/root/.autorest/@microsoft.azure_autorest.modeler@2.3.44/node_modules/@microsoft.azure/autorest.modeler", + "saveSpec": "file:/root/.autorest/@microsoft.azure_autorest.modeler@2.3.44/node_modules/@microsoft.azure/autorest.modeler", + "fetchSpec": "/root/.autorest/@microsoft.azure_autorest.modeler@2.3.44/node_modules/@microsoft.azure/autorest.modeler" }, - "_spec": "/tmp/.autorest/@microsoft.azure_autorest.python@2.1.40/node_modules/@microsoft.azure/autorest.python", - "_where": "/tmp/.autorest/@microsoft.azure_autorest.python@2.1.40/node_modules/@microsoft.azure/autorest.python" + "_spec": "/root/.autorest/@microsoft.azure_autorest.modeler@2.3.44/node_modules/@microsoft.azure/autorest.modeler", + "_where": "/root/.autorest/@microsoft.azure_autorest.modeler@2.3.44/node_modules/@microsoft.azure/autorest.modeler" }, "extensionManager": { - "installationPath": "/tmp/.autorest", + "installationPath": "/root/.autorest", "sharedLock": { - "name": "/tmp/.autorest", + "name": "/root/.autorest", "exclusiveLock": { - "name": "_tmp_.autorest.exclusive-lock", + "name": "_root_.autorest.exclusive-lock", "options": { - "port": 5245, - "host": "2130716018", + "port": 45234, + "host": "2130706813", "exclusive": true }, - "pipe": "/tmp/pipe__tmp_.autorest.exclusive-lock:5245" + "pipe": "/tmp/pipe__root_.autorest.exclusive-lock:45234" }, "busyLock": { - "name": "_tmp_.autorest.busy-lock", + "name": "_root_.autorest.busy-lock", "options": { - "port": 46071, - "host": "2130735832", + "port": 37199, + "host": "2130756895", "exclusive": true }, - "pipe": "/tmp/pipe__tmp_.autorest.busy-lock:46071" + "pipe": "/tmp/pipe__root_.autorest.busy-lock:37199" }, "personalLock": { - "name": "_tmp_.autorest.4890.15425350853.personal-lock", + "name": "_root_.autorest.8216.386719855544.personal-lock", "options": { - "port": 22892, - "host": "2130771569", + "port": 63645, + "host": "2130724149", "exclusive": true }, - "pipe": "/tmp/pipe__tmp_.autorest.4890.15425350853.personal-lock:22892" + "pipe": "/tmp/pipe__root_.autorest.8216.386719855544.personal-lock:63645" }, - "file": "/tmp/_tmp_.autorest.lock" + "file": "/tmp/_root_.autorest.lock" }, "dotnetPath": "/root/.dotnet" }, - "installationPath": "/tmp/.autorest" + "installationPath": "/root/.autorest" }, { "resolvedInfo": null, "packageMetadata": { "name": "@microsoft.azure/autorest.python", - "version": "3.0.41", + "version": "3.0.50", "dependencies": { "dotnet-2.0.0": "^1.4.4" }, "optionalDependencies": {}, "devDependencies": { - "@microsoft.azure/autorest.testserver": "^2.4.0", + "@microsoft.azure/autorest.testserver": "^2.5.6", "autorest": "^2.0.4203", "coffee-script": "^1.11.1", "dotnet-sdk-2.0.0": "^1.4.4", @@ -369,60 +285,60 @@ "bundleDependencies": false, "peerDependencies": {}, "deprecated": false, - "_resolved": "/tmp/.autorest/@microsoft.azure_autorest.python@3.0.41/node_modules/@microsoft.azure/autorest.python", + "_resolved": "/root/.autorest/@microsoft.azure_autorest.python@3.0.50/node_modules/@microsoft.azure/autorest.python", "_integrity": null, - "_shasum": "184d947713655b7e1fcc3b269466953b45d5bddf", + "_shasum": "2faedd2ee31addf11cb56d96d14adee4bd5587a0", "_shrinkwrap": null, "bin": null, - "_id": "@microsoft.azure/autorest.python@3.0.41", - "_from": "file:/tmp/.autorest/@microsoft.azure_autorest.python@3.0.41/node_modules/@microsoft.azure/autorest.python", + "_id": "@microsoft.azure/autorest.python@3.0.50", + "_from": "file:/root/.autorest/@microsoft.azure_autorest.python@3.0.50/node_modules/@microsoft.azure/autorest.python", "_requested": { "type": "directory", - "where": "/tmp/.autorest/@microsoft.azure_autorest.python@3.0.41/node_modules/@microsoft.azure/autorest.python", - "raw": "/tmp/.autorest/@microsoft.azure_autorest.python@3.0.41/node_modules/@microsoft.azure/autorest.python", - "rawSpec": "/tmp/.autorest/@microsoft.azure_autorest.python@3.0.41/node_modules/@microsoft.azure/autorest.python", - "saveSpec": "file:/tmp/.autorest/@microsoft.azure_autorest.python@3.0.41/node_modules/@microsoft.azure/autorest.python", - "fetchSpec": "/tmp/.autorest/@microsoft.azure_autorest.python@3.0.41/node_modules/@microsoft.azure/autorest.python" + "where": "/root/.autorest/@microsoft.azure_autorest.python@3.0.50/node_modules/@microsoft.azure/autorest.python", + "raw": "/root/.autorest/@microsoft.azure_autorest.python@3.0.50/node_modules/@microsoft.azure/autorest.python", + "rawSpec": "/root/.autorest/@microsoft.azure_autorest.python@3.0.50/node_modules/@microsoft.azure/autorest.python", + "saveSpec": "file:/root/.autorest/@microsoft.azure_autorest.python@3.0.50/node_modules/@microsoft.azure/autorest.python", + "fetchSpec": "/root/.autorest/@microsoft.azure_autorest.python@3.0.50/node_modules/@microsoft.azure/autorest.python" }, - "_spec": "/tmp/.autorest/@microsoft.azure_autorest.python@3.0.41/node_modules/@microsoft.azure/autorest.python", - "_where": "/tmp/.autorest/@microsoft.azure_autorest.python@3.0.41/node_modules/@microsoft.azure/autorest.python" + "_spec": "/root/.autorest/@microsoft.azure_autorest.python@3.0.50/node_modules/@microsoft.azure/autorest.python", + "_where": "/root/.autorest/@microsoft.azure_autorest.python@3.0.50/node_modules/@microsoft.azure/autorest.python" }, "extensionManager": { - "installationPath": "/tmp/.autorest", + "installationPath": "/root/.autorest", "sharedLock": { - "name": "/tmp/.autorest", + "name": "/root/.autorest", "exclusiveLock": { - "name": "_tmp_.autorest.exclusive-lock", + "name": "_root_.autorest.exclusive-lock", "options": { - "port": 5245, - "host": "2130716018", + "port": 45234, + "host": "2130706813", "exclusive": true }, - "pipe": "/tmp/pipe__tmp_.autorest.exclusive-lock:5245" + "pipe": "/tmp/pipe__root_.autorest.exclusive-lock:45234" }, "busyLock": { - "name": "_tmp_.autorest.busy-lock", + "name": "_root_.autorest.busy-lock", "options": { - "port": 46071, - "host": "2130735832", + "port": 37199, + "host": "2130756895", "exclusive": true }, - "pipe": "/tmp/pipe__tmp_.autorest.busy-lock:46071" + "pipe": "/tmp/pipe__root_.autorest.busy-lock:37199" }, "personalLock": { - "name": "_tmp_.autorest.4890.15425350853.personal-lock", + "name": "_root_.autorest.8216.386719855544.personal-lock", "options": { - "port": 22892, - "host": "2130771569", + "port": 63645, + "host": "2130724149", "exclusive": true }, - "pipe": "/tmp/pipe__tmp_.autorest.4890.15425350853.personal-lock:22892" + "pipe": "/tmp/pipe__root_.autorest.8216.386719855544.personal-lock:63645" }, - "file": "/tmp/_tmp_.autorest.lock" + "file": "/tmp/_root_.autorest.lock" }, "dotnetPath": "/root/.dotnet" }, - "installationPath": "/tmp/.autorest" + "installationPath": "/root/.autorest" } ], "autorest_bootstrap": {}