Skip to content
Closed
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 1 addition & 1 deletion sdk/batch/azure-batch/MANIFEST.in
Original file line number Diff line number Diff line change
Expand Up @@ -3,4 +3,4 @@ include LICENSE
include azure/batch/py.typed
recursive-include tests *.py
recursive-include samples *.py *.md
include azure/__init__.py
include azure/__init__.py
6 changes: 6 additions & 0 deletions sdk/batch/azure-batch/_meta.json
Original file line number Diff line number Diff line change
@@ -0,0 +1,6 @@
{
"commit": "167d2be6a8c000e27109c75f9133218e92c64724",
"repository_url": "https://github.com/Azure/azure-rest-api-specs",
"typespec_src": "specification/batch/Azure.Batch",
"@azure-tools/typespec-python": "0.42.2"
}
286 changes: 286 additions & 0 deletions sdk/batch/azure-batch/apiview-properties.json

Large diffs are not rendered by default.

5 changes: 3 additions & 2 deletions sdk/batch/azure-batch/azure/batch/_model_base.py
Original file line number Diff line number Diff line change
Expand Up @@ -2,8 +2,9 @@
# coding=utf-8
# --------------------------------------------------------------------------
# Copyright (c) Microsoft Corporation. All rights reserved.
# Licensed under the MIT License. See License.txt in the project root for
# license information.
# Licensed under the MIT License. See License.txt in the project root for license information.
# Code generated by Microsoft (R) Python Code Generator.
# Changes may cause incorrect behavior and will be lost if the code is regenerated.
# --------------------------------------------------------------------------
# pylint: disable=protected-access, broad-except

Expand Down
62 changes: 30 additions & 32 deletions sdk/batch/azure-batch/azure/batch/_operations/_operations.py
Original file line number Diff line number Diff line change
Expand Up @@ -478,11 +478,11 @@ def build_batch_enable_pool_auto_scale_request( # pylint: disable=name-too-long
# Construct headers
if ocpdate is not None:
_headers["ocp-date"] = _SERIALIZER.header("ocpdate", ocpdate, "rfc-1123")
_headers["content-type"] = _SERIALIZER.header("content_type", content_type, "str")
if if_modified_since is not None:
_headers["If-Modified-Since"] = _SERIALIZER.header("if_modified_since", if_modified_since, "rfc-1123")
if if_unmodified_since is not None:
_headers["If-Unmodified-Since"] = _SERIALIZER.header("if_unmodified_since", if_unmodified_since, "rfc-1123")
_headers["content-type"] = _SERIALIZER.header("content_type", content_type, "str")
_headers["Accept"] = _SERIALIZER.header("accept", accept, "str")
if_match = prep_if_match(etag, match_condition)
if if_match is not None:
Expand Down Expand Up @@ -560,11 +560,11 @@ def build_batch_resize_pool_request(
# Construct headers
if ocpdate is not None:
_headers["ocp-date"] = _SERIALIZER.header("ocpdate", ocpdate, "rfc-1123")
_headers["content-type"] = _SERIALIZER.header("content_type", content_type, "str")
if if_modified_since is not None:
_headers["If-Modified-Since"] = _SERIALIZER.header("if_modified_since", if_modified_since, "rfc-1123")
if if_unmodified_since is not None:
_headers["If-Unmodified-Since"] = _SERIALIZER.header("if_unmodified_since", if_unmodified_since, "rfc-1123")
_headers["content-type"] = _SERIALIZER.header("content_type", content_type, "str")
_headers["Accept"] = _SERIALIZER.header("accept", accept, "str")
if_match = prep_if_match(etag, match_condition)
if if_match is not None:
Expand Down Expand Up @@ -690,11 +690,11 @@ def build_batch_remove_nodes_request(
# Construct headers
if ocpdate is not None:
_headers["ocp-date"] = _SERIALIZER.header("ocpdate", ocpdate, "rfc-1123")
_headers["content-type"] = _SERIALIZER.header("content_type", content_type, "str")
if if_modified_since is not None:
_headers["If-Modified-Since"] = _SERIALIZER.header("if_modified_since", if_modified_since, "rfc-1123")
if if_unmodified_since is not None:
_headers["If-Unmodified-Since"] = _SERIALIZER.header("if_unmodified_since", if_unmodified_since, "rfc-1123")
_headers["content-type"] = _SERIALIZER.header("content_type", content_type, "str")
_headers["Accept"] = _SERIALIZER.header("accept", accept, "str")
if_match = prep_if_match(etag, match_condition)
if if_match is not None:
Expand Down Expand Up @@ -1013,11 +1013,11 @@ def build_batch_disable_job_request(
# Construct headers
if ocpdate is not None:
_headers["ocp-date"] = _SERIALIZER.header("ocpdate", ocpdate, "rfc-1123")
_headers["content-type"] = _SERIALIZER.header("content_type", content_type, "str")
if if_modified_since is not None:
_headers["If-Modified-Since"] = _SERIALIZER.header("if_modified_since", if_modified_since, "rfc-1123")
if if_unmodified_since is not None:
_headers["If-Unmodified-Since"] = _SERIALIZER.header("if_unmodified_since", if_unmodified_since, "rfc-1123")
_headers["content-type"] = _SERIALIZER.header("content_type", content_type, "str")
_headers["Accept"] = _SERIALIZER.header("accept", accept, "str")
if_match = prep_if_match(etag, match_condition)
if if_match is not None:
Expand Down Expand Up @@ -1114,11 +1114,11 @@ def build_batch_terminate_job_request(
# Construct headers
if ocpdate is not None:
_headers["ocp-date"] = _SERIALIZER.header("ocpdate", ocpdate, "rfc-1123")
_headers["content-type"] = _SERIALIZER.header("content_type", content_type, "str")
if if_modified_since is not None:
_headers["If-Modified-Since"] = _SERIALIZER.header("if_modified_since", if_modified_since, "rfc-1123")
if if_unmodified_since is not None:
_headers["If-Unmodified-Since"] = _SERIALIZER.header("if_unmodified_since", if_unmodified_since, "rfc-1123")
_headers["content-type"] = _SERIALIZER.header("content_type", content_type, "str")
_headers["Accept"] = _SERIALIZER.header("accept", accept, "str")
if_match = prep_if_match(etag, match_condition)
if if_match is not None:
Expand Down Expand Up @@ -3387,7 +3387,7 @@ def prepare_request(next_link=None):

def extract_data(pipeline_response):
deserialized = pipeline_response.http_response.json()
list_of_elem = _deserialize(List[_models.BatchApplication], deserialized["value"])
list_of_elem = _deserialize(List[_models.BatchApplication], deserialized.get("value", []))
if cls:
list_of_elem = cls(list_of_elem) # type: ignore
return deserialized.get("odata.nextLink") or None, iter(list_of_elem)
Expand Down Expand Up @@ -3610,7 +3610,7 @@ def prepare_request(next_link=None):

def extract_data(pipeline_response):
deserialized = pipeline_response.http_response.json()
list_of_elem = _deserialize(List[_models.BatchPoolUsageMetrics], deserialized["value"])
list_of_elem = _deserialize(List[_models.BatchPoolUsageMetrics], deserialized.get("value", []))
if cls:
list_of_elem = cls(list_of_elem) # type: ignore
return deserialized.get("odata.nextLink") or None, iter(list_of_elem)
Expand Down Expand Up @@ -3814,7 +3814,7 @@ def prepare_request(next_link=None):

def extract_data(pipeline_response):
deserialized = pipeline_response.http_response.json()
list_of_elem = _deserialize(List[_models.BatchPool], deserialized["value"])
list_of_elem = _deserialize(List[_models.BatchPool], deserialized.get("value", []))
if cls:
list_of_elem = cls(list_of_elem) # type: ignore
return deserialized.get("odata.nextLink") or None, iter(list_of_elem)
Expand Down Expand Up @@ -5101,7 +5101,7 @@ def prepare_request(next_link=None):

def extract_data(pipeline_response):
deserialized = pipeline_response.http_response.json()
list_of_elem = _deserialize(List[_models.BatchSupportedImage], deserialized["value"])
list_of_elem = _deserialize(List[_models.BatchSupportedImage], deserialized.get("value", []))
if cls:
list_of_elem = cls(list_of_elem) # type: ignore
return deserialized.get("odata.nextLink") or None, iter(list_of_elem)
Expand Down Expand Up @@ -5214,7 +5214,7 @@ def prepare_request(next_link=None):

def extract_data(pipeline_response):
deserialized = pipeline_response.http_response.json()
list_of_elem = _deserialize(List[_models.BatchPoolNodeCounts], deserialized["value"])
list_of_elem = _deserialize(List[_models.BatchPoolNodeCounts], deserialized.get("value", []))
if cls:
list_of_elem = cls(list_of_elem) # type: ignore
return deserialized.get("odata.nextLink") or None, iter(list_of_elem)
Expand Down Expand Up @@ -6249,7 +6249,7 @@ def prepare_request(next_link=None):

def extract_data(pipeline_response):
deserialized = pipeline_response.http_response.json()
list_of_elem = _deserialize(List[_models.BatchJob], deserialized["value"])
list_of_elem = _deserialize(List[_models.BatchJob], deserialized.get("value", []))
if cls:
list_of_elem = cls(list_of_elem) # type: ignore
return deserialized.get("odata.nextLink") or None, iter(list_of_elem)
Expand Down Expand Up @@ -6375,7 +6375,7 @@ def prepare_request(next_link=None):

def extract_data(pipeline_response):
deserialized = pipeline_response.http_response.json()
list_of_elem = _deserialize(List[_models.BatchJob], deserialized["value"])
list_of_elem = _deserialize(List[_models.BatchJob], deserialized.get("value", []))
if cls:
list_of_elem = cls(list_of_elem) # type: ignore
return deserialized.get("odata.nextLink") or None, iter(list_of_elem)
Expand Down Expand Up @@ -6503,7 +6503,9 @@ def prepare_request(next_link=None):

def extract_data(pipeline_response):
deserialized = pipeline_response.http_response.json()
list_of_elem = _deserialize(List[_models.BatchJobPreparationAndReleaseTaskStatus], deserialized["value"])
list_of_elem = _deserialize(
List[_models.BatchJobPreparationAndReleaseTaskStatus], deserialized.get("value", [])
)
if cls:
list_of_elem = cls(list_of_elem) # type: ignore
return deserialized.get("odata.nextLink") or None, iter(list_of_elem)
Expand Down Expand Up @@ -6785,7 +6787,7 @@ def prepare_request(next_link=None):

def extract_data(pipeline_response):
deserialized = pipeline_response.http_response.json()
list_of_elem = _deserialize(List[_models.BatchCertificate], deserialized["value"])
list_of_elem = _deserialize(List[_models.BatchCertificate], deserialized.get("value", []))
if cls:
list_of_elem = cls(list_of_elem) # type: ignore
return deserialized.get("odata.nextLink") or None, iter(list_of_elem)
Expand Down Expand Up @@ -6991,7 +6993,7 @@ def get_certificate(
ocpdate: Optional[datetime.datetime] = None,
select: Optional[List[str]] = None,
**kwargs: Any
) -> _models.GetCertificateResponse:
) -> _models.BatchCertificate:
"""Gets information about the specified Certificate.

:param thumbprint_algorithm: The algorithm used to derive the thumbprint parameter. This must
Expand All @@ -7009,8 +7011,8 @@ def get_certificate(
:paramtype ocpdate: ~datetime.datetime
:keyword select: An OData $select clause. Default value is None.
:paramtype select: list[str]
:return: GetCertificateResponse. The GetCertificateResponse is compatible with MutableMapping
:rtype: ~azure.batch.models.GetCertificateResponse
:return: BatchCertificate. The BatchCertificate is compatible with MutableMapping
:rtype: ~azure.batch.models.BatchCertificate
:raises ~azure.core.exceptions.HttpResponseError:
"""
error_map: MutableMapping = {
Expand All @@ -7024,7 +7026,7 @@ def get_certificate(
_headers = kwargs.pop("headers", {}) or {}
_params = kwargs.pop("params", {}) or {}

cls: ClsType[_models.GetCertificateResponse] = kwargs.pop("cls", None)
cls: ClsType[_models.BatchCertificate] = kwargs.pop("cls", None)

_request = build_batch_get_certificate_request(
thumbprint_algorithm=thumbprint_algorithm,
Expand Down Expand Up @@ -7067,7 +7069,7 @@ def get_certificate(
if _stream:
deserialized = response.iter_bytes()
else:
deserialized = _deserialize(_models.GetCertificateResponse, response.json())
deserialized = _deserialize(_models.BatchCertificate, response.json())

if cls:
return cls(pipeline_response, deserialized, response_headers) # type: ignore
Expand Down Expand Up @@ -8143,7 +8145,7 @@ def prepare_request(next_link=None):

def extract_data(pipeline_response):
deserialized = pipeline_response.http_response.json()
list_of_elem = _deserialize(List[_models.BatchJobSchedule], deserialized["value"])
list_of_elem = _deserialize(List[_models.BatchJobSchedule], deserialized.get("value", []))
if cls:
list_of_elem = cls(list_of_elem) # type: ignore
return deserialized.get("odata.nextLink") or None, iter(list_of_elem)
Expand Down Expand Up @@ -8357,7 +8359,7 @@ def prepare_request(next_link=None):

def extract_data(pipeline_response):
deserialized = pipeline_response.http_response.json()
list_of_elem = _deserialize(List[_models.BatchTask], deserialized["value"])
list_of_elem = _deserialize(List[_models.BatchTask], deserialized.get("value", []))
if cls:
list_of_elem = cls(list_of_elem) # type: ignore
return deserialized.get("odata.nextLink") or None, iter(list_of_elem)
Expand Down Expand Up @@ -8936,7 +8938,7 @@ def prepare_request(next_link=None):

def extract_data(pipeline_response):
deserialized = pipeline_response.http_response.json()
list_of_elem = _deserialize(List[_models.BatchSubtask], deserialized["value"])
list_of_elem = _deserialize(List[_models.BatchSubtask], deserialized.get("value", []))
if cls:
list_of_elem = cls(list_of_elem) # type: ignore
return deserialized.get("odata.nextLink") or None, iter(list_of_elem)
Expand Down Expand Up @@ -9389,10 +9391,8 @@ def get_task_file(

return deserialized # type: ignore

# manually renamed
# rename will be through typespec in next version
@distributed_trace
def _get_task_file_properties_internal(
def get_task_file_properties(
self,
job_id: str,
task_id: str,
Expand Down Expand Up @@ -9597,7 +9597,7 @@ def prepare_request(next_link=None):

def extract_data(pipeline_response):
deserialized = pipeline_response.http_response.json()
list_of_elem = _deserialize(List[_models.BatchNodeFile], deserialized["value"])
list_of_elem = _deserialize(List[_models.BatchNodeFile], deserialized.get("value", []))
if cls:
list_of_elem = cls(list_of_elem) # type: ignore
return deserialized.get("odata.nextLink") or None, iter(list_of_elem)
Expand Down Expand Up @@ -10811,7 +10811,7 @@ def prepare_request(next_link=None):

def extract_data(pipeline_response):
deserialized = pipeline_response.http_response.json()
list_of_elem = _deserialize(List[_models.BatchNode], deserialized["value"])
list_of_elem = _deserialize(List[_models.BatchNode], deserialized.get("value", []))
if cls:
list_of_elem = cls(list_of_elem) # type: ignore
return deserialized.get("odata.nextLink") or None, iter(list_of_elem)
Expand Down Expand Up @@ -11028,7 +11028,7 @@ def prepare_request(next_link=None):

def extract_data(pipeline_response):
deserialized = pipeline_response.http_response.json()
list_of_elem = _deserialize(List[_models.BatchNodeVMExtension], deserialized["value"])
list_of_elem = _deserialize(List[_models.BatchNodeVMExtension], deserialized.get("value", []))
if cls:
list_of_elem = cls(list_of_elem) # type: ignore
return deserialized.get("odata.nextLink") or None, iter(list_of_elem)
Expand Down Expand Up @@ -11256,10 +11256,8 @@ def get_node_file(

return deserialized # type: ignore

# manually renamed
# rename will be through typespec in next version
@distributed_trace
def _get_node_file_properties_internal(
def get_node_file_properties(
self,
pool_id: str,
node_id: str,
Expand Down Expand Up @@ -11462,7 +11460,7 @@ def prepare_request(next_link=None):

def extract_data(pipeline_response):
deserialized = pipeline_response.http_response.json()
list_of_elem = _deserialize(List[_models.BatchNodeFile], deserialized["value"])
list_of_elem = _deserialize(List[_models.BatchNodeFile], deserialized.get("value", []))
if cls:
list_of_elem = cls(list_of_elem) # type: ignore
return deserialized.get("odata.nextLink") or None, iter(list_of_elem)
Expand Down
Loading