Skip to content

Commit b431fff

Browse files
author
Andrew Omondi
committed
chore: revert typing changes
1 parent dd1a4e1 commit b431fff

File tree

7 files changed

+38
-54
lines changed

7 files changed

+38
-54
lines changed

src/msgraph_core/models/large_file_upload_session.py

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,8 +1,9 @@
11
from __future__ import annotations
22

33
import datetime
4+
from collections.abc import Callable
45
from dataclasses import dataclass, field
5-
from typing import Any, Callable, Optional
6+
from typing import Any, Optional
67

78
from kiota_abstractions.serialization import (
89
AdditionalDataHolder,

src/msgraph_core/models/page_result.py

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -11,8 +11,9 @@
1111
"""
1212
from __future__ import annotations
1313

14+
from collections.abc import Callable
1415
from dataclasses import dataclass
15-
from typing import Callable, Optional, TypeVar
16+
from typing import Optional, TypeVar
1617

1718
from kiota_abstractions.serialization.parsable import Parsable
1819
from kiota_abstractions.serialization.parse_node import ParseNode

src/msgraph_core/models/upload_result.py

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,7 @@
1+
from collections.abc import Callable
12
from dataclasses import dataclass
23
from datetime import datetime
3-
from typing import Any, Callable, Generic, Optional, TypeVar
4+
from typing import Any, Generic, Optional, TypeVar
45

56
from kiota_abstractions.serialization import (
67
AdditionalDataHolder,

src/msgraph_core/requests/batch_response_content.py

Lines changed: 8 additions & 13 deletions
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,7 @@
11
import base64
2+
from collections.abc import Callable
23
from io import BytesIO
3-
from typing import Callable, Optional, Type, TypeVar, Union
4+
from typing import Optional, Type, TypeVar, Union
45

56
from kiota_abstractions.serialization import (
67
Parsable,
@@ -107,20 +108,17 @@ def response_body(self, request_id: str, type: Type[T]) -> Optional[T]:
107108
raise ValueError(f"No response found for id: {request_id}")
108109

109110
if not issubclass(type, Parsable):
110-
raise ValueError(
111-
"Type passed must implement the Parsable interface")
111+
raise ValueError("Type passed must implement the Parsable interface")
112112

113113
response = self.get_response_by_id(request_id)
114114
if response is not None:
115115
content_type = response.content_type
116116
else:
117117
raise ValueError(
118-
f"Unable to get content-type header in response item for request Id: {
119-
request_id}"
118+
f"Unable to get content-type header in response item for request Id: {request_id}"
120119
)
121120
if not content_type:
122-
raise RuntimeError(
123-
"Unable to get content-type header in response item")
121+
raise RuntimeError("Unable to get content-type header in response item")
124122

125123
response_body = response.body or BytesIO()
126124
try:
@@ -130,17 +128,15 @@ def response_body(self, request_id: str, type: Type[T]) -> Optional[T]:
130128
)
131129
except Exception:
132130
response_body.seek(0)
133-
base64_decoded_body = BytesIO(
134-
base64.b64decode(response_body.read()))
131+
base64_decoded_body = BytesIO(base64.b64decode(response_body.read()))
135132
parse_node = ParseNodeFactoryRegistry().get_root_parse_node(
136133
content_type, base64_decoded_body
137134
)
138135
response.body = base64_decoded_body
139136
return parse_node.get_object_value(type)
140137
except Exception:
141138
raise ValueError(
142-
f"Unable to deserialize batch response for request Id: {
143-
request_id} to {type}"
139+
f"Unable to deserialize batch response for request Id: {request_id} to {type}"
144140
)
145141

146142
def get_field_deserializers(self) -> dict[str, Callable[[ParseNode], None]]:
@@ -165,8 +161,7 @@ def serialize(self, writer: SerializationWriter) -> None:
165161
:param writer: The writer to write to
166162
"""
167163
if self._responses is not None:
168-
writer.write_collection_of_object_values(
169-
'responses', list(self._responses.values()))
164+
writer.write_collection_of_object_values('responses', list(self._responses.values()))
170165
else:
171166
writer.write_collection_of_object_values('responses', [])
172167

src/msgraph_core/requests/batch_response_content_collection.py

Lines changed: 4 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -1,4 +1,5 @@
1-
from typing import Callable
1+
from collections.abc import Callable
2+
23
from kiota_abstractions.serialization import Parsable, ParseNode, SerializationWriter
34

45
from .batch_response_content import BatchResponseContent
@@ -51,8 +52,7 @@ async def responses_status_codes(self) -> dict[str, int]:
5152
else:
5253
raise ValueError("Response ID cannot be None")
5354
else:
54-
raise TypeError(
55-
"Invalid type: Collection must be of type BatchResponseContent")
55+
raise TypeError("Invalid type: Collection must be of type BatchResponseContent")
5656
return status_codes
5757

5858
def get_field_deserializers(self) -> dict[str, Callable[[ParseNode], None]]:
@@ -65,8 +65,7 @@ def get_field_deserializers(self) -> dict[str, Callable[[ParseNode], None]]:
6565
return {
6666
'responses':
6767
lambda n:
68-
setattr(self, "_responses",
69-
n.get_collection_of_object_values(BatchResponseItem))
68+
setattr(self, "_responses", n.get_collection_of_object_values(BatchResponseItem))
7069
}
7170

7271
def serialize(self, writer: SerializationWriter) -> None:

src/msgraph_core/tasks/large_file_upload.py

Lines changed: 15 additions & 26 deletions
Original file line numberDiff line numberDiff line change
@@ -1,9 +1,10 @@
11
import logging
22
import os
33
from asyncio import Future
4+
from collections.abc import Callable
45
from datetime import datetime, timedelta, timezone
56
from io import BytesIO
6-
from typing import Any, Callable, Optional, Tuple, TypeVar, Union
7+
from typing import Any, Optional, Tuple, TypeVar, Union
78

89
from kiota_abstractions.headers_collection import HeadersCollection
910
from kiota_abstractions.method import Method
@@ -37,13 +38,11 @@ def __init__(
3738
self.max_chunk_size = max_chunk_size
3839
self.factory = parsable_factory
3940
cleaned_value = self.check_value_exists(
40-
upload_session, 'get_next_expected_range', [
41-
'next_expected_range', 'NextExpectedRange']
41+
upload_session, 'get_next_expected_range', ['next_expected_range', 'NextExpectedRange']
4242
)
4343
self.next_range = cleaned_value[0]
4444
self._chunks = int((self.file_size / max_chunk_size) + 0.5)
45-
self.on_chunk_upload_complete: Optional[Callable[[
46-
list[int]], None]] = None
45+
self.on_chunk_upload_complete: Optional[Callable[[list[int]], None]] = None
4746

4847
@property
4948
def upload_session(self):
@@ -69,8 +68,7 @@ def upload_session_expired(self, upload_session: Optional[Parsable] = None) -> b
6968
now = datetime.now(timezone.utc)
7069
upload_session = upload_session or self.upload_session
7170
if not hasattr(upload_session, "expiration_date_time"):
72-
raise ValueError(
73-
"Upload session does not have an expiration date time")
71+
raise ValueError("Upload session does not have an expiration date time")
7472
expiry = getattr(upload_session, 'expiration_date_time')
7573
if expiry is None:
7674
raise ValueError("Expiry is None")
@@ -95,16 +93,13 @@ async def upload(self, after_chunk_upload: Optional[Callable] = None):
9593

9694
self.on_chunk_upload_complete = after_chunk_upload or self.on_chunk_upload_complete
9795
session: LargeFileUploadSession = await self.next_chunk(
98-
self.stream, 0, max(
99-
0, min(self.max_chunk_size - 1, self.file_size - 1))
96+
self.stream, 0, max(0, min(self.max_chunk_size - 1, self.file_size - 1))
10097
)
10198
process_next = session
10299
# determine the range to be uploaded
103100
# even when resuming existing upload sessions.
104-
range_parts = self.next_range[0].split(
105-
"-") if self.next_range else ['0', '0']
106-
end = min(int(range_parts[0]) +
107-
self.max_chunk_size - 1, self.file_size)
101+
range_parts = self.next_range[0].split("-") if self.next_range else ['0', '0']
102+
end = min(int(range_parts[0]) + self.max_chunk_size - 1, self.file_size)
108103
uploaded_range = [range_parts[0], end]
109104
response = None
110105

@@ -129,13 +124,12 @@ async def upload(self, after_chunk_upload: Optional[Callable] = None):
129124
if not next_range:
130125
continue
131126
range_parts = str(next_range[0]).split("-")
132-
end = min(int(range_parts[0]) +
133-
self.max_chunk_size, self.file_size)
127+
end = min(int(range_parts[0]) + self.max_chunk_size, self.file_size)
134128
uploaded_range = [range_parts[0], end]
135129
self.next_range = next_range[0] + "-"
136130
process_next = await self.next_chunk(self.stream)
137131

138-
except Exception as error: # pylint: disable=broad-except
132+
except Exception as error: #pylint: disable=broad-except
139133
logging.error("Error uploading chunk %s", error)
140134
finally:
141135
self.chunks -= 1
@@ -182,8 +176,7 @@ async def next_chunk(
182176
chunk_data = file.read(end - start + 1)
183177
info.headers = HeadersCollection()
184178

185-
info.headers.try_add(
186-
'Content-Range', f'bytes {start}-{end}/{self.file_size}')
179+
info.headers.try_add('Content-Range', f'bytes {start}-{end}/{self.file_size}')
187180
info.headers.try_add('Content-Length', str(len(chunk_data)))
188181
info.headers.try_add("Content-Type", "application/octet-stream")
189182
info.set_stream_content(bytes(chunk_data))
@@ -223,8 +216,7 @@ async def last_chunk(
223216
chunk_data = file.read(end - start + 1)
224217
info.headers = HeadersCollection()
225218

226-
info.headers.try_add(
227-
'Content-Range', f'bytes {start}-{end}/{self.file_size}')
219+
info.headers.try_add('Content-Range', f'bytes {start}-{end}/{self.file_size}')
228220
info.headers.try_add('Content-Length', str(len(chunk_data)))
229221
info.headers.try_add("Content-Type", "application/octet-stream")
230222
info.set_stream_content(bytes(chunk_data))
@@ -239,8 +231,7 @@ def get_file(self) -> BytesIO:
239231

240232
async def cancel(self) -> Parsable:
241233
upload_url = self.get_validated_upload_url(self.upload_session)
242-
request_information = RequestInformation(
243-
method=Method.DELETE, url_template=upload_url)
234+
request_information = RequestInformation(method=Method.DELETE, url_template=upload_url)
244235

245236
await self.request_adapter.send_no_response_content_async(request_information)
246237

@@ -263,8 +254,7 @@ def additional_data_contains(self, parsable: Parsable,
263254
'AdditionalDataHolder'
264255
)
265256
if not hasattr(parsable, 'additional_data'):
266-
raise ValueError(
267-
'The object passed does not contain an additional_data property')
257+
raise ValueError('The object passed does not contain an additional_data property')
268258
additional_data = parsable.additional_data
269259
for property_candidate in property_candidates:
270260
if property_candidate in additional_data:
@@ -308,8 +298,7 @@ async def resume(self) -> Future:
308298

309299
def get_validated_upload_url(self, upload_session: Parsable) -> str:
310300
if not hasattr(upload_session, 'upload_url'):
311-
raise RuntimeError(
312-
'The upload session does not contain a valid upload url')
301+
raise RuntimeError('The upload session does not contain a valid upload url')
313302
result = upload_session.upload_url
314303

315304
if result is None or result.strip() == '':

src/msgraph_core/tasks/page_iterator.py

Lines changed: 5 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -17,7 +17,8 @@
1717
and models modules.
1818
"""
1919

20-
from typing import Callable, Optional, Type, TypeVar, Union
20+
from collections.abc import Callable
21+
from typing import Optional, Type, TypeVar, Union
2122

2223
from kiota_abstractions.headers_collection import HeadersCollection
2324
from kiota_abstractions.method import Method
@@ -151,8 +152,7 @@ async def next(self) -> Optional[PageResult]:
151152
next_link = response.odata_next_link if response and hasattr(
152153
response, 'odata_next_link'
153154
) else None
154-
value = response.value if response and hasattr(
155-
response, 'value') else None
155+
value = response.value if response and hasattr(response, 'value') else None
156156
return PageResult(next_link, value)
157157

158158
@staticmethod
@@ -180,8 +180,7 @@ def convert_to_page(response: Union[T, list, object]) -> PageResult:
180180
value = getattr(response, 'value', [])
181181
if value is None:
182182
raise ValueError('The response does not contain a value.')
183-
parsable_page = response if isinstance(
184-
response, dict) else vars(response)
183+
parsable_page = response if isinstance(response, dict) else vars(response)
185184
next_link = parsable_page.get('odata_next_link', '') if isinstance(
186185
parsable_page, dict
187186
) else getattr(parsable_page, 'odata_next_link', '')
@@ -231,8 +230,7 @@ def enumerate(self, callback: Optional[Callable] = None) -> bool:
231230
if not page_items:
232231
return False
233232
for i in range(self.pause_index, len(page_items)):
234-
keep_iterating = callback(
235-
page_items[i]) if callback is not None else True
233+
keep_iterating = callback(page_items[i]) if callback is not None else True
236234
if not keep_iterating:
237235
self.pause_index = i + 1
238236
break

0 commit comments

Comments
 (0)