|
19 | 19 | from openai import OpenAI, AsyncOpenAI, APIResponseValidationError |
20 | 20 | from openai._client import OpenAI, AsyncOpenAI |
21 | 21 | from openai._models import BaseModel, FinalRequestOptions |
22 | | -from openai._response import APIResponse, AsyncAPIResponse |
23 | 22 | from openai._constants import RAW_RESPONSE_HEADER |
24 | 23 | from openai._streaming import Stream, AsyncStream |
25 | 24 | from openai._exceptions import OpenAIError, APIStatusError, APITimeoutError, APIResponseValidationError |
@@ -665,33 +664,6 @@ def test_parse_retry_after_header(self, remaining_retries: int, retry_after: str |
665 | 664 | calculated = client._calculate_retry_timeout(remaining_retries, options, headers) |
666 | 665 | assert calculated == pytest.approx(timeout, 0.5 * 0.875) # pyright: ignore[reportUnknownMemberType] |
667 | 666 |
|
668 | | - @mock.patch("openai._base_client.BaseClient._calculate_retry_timeout", _low_retry_timeout) |
669 | | - @pytest.mark.respx(base_url=base_url) |
670 | | - def test_streaming_response(self) -> None: |
671 | | - response = self.client.post( |
672 | | - "/chat/completions", |
673 | | - body=dict( |
674 | | - messages=[ |
675 | | - { |
676 | | - "role": "user", |
677 | | - "content": "Say this is a test", |
678 | | - } |
679 | | - ], |
680 | | - model="gpt-3.5-turbo", |
681 | | - ), |
682 | | - cast_to=APIResponse[bytes], |
683 | | - options={"headers": {RAW_RESPONSE_HEADER: "stream"}}, |
684 | | - ) |
685 | | - |
686 | | - assert not cast(Any, response.is_closed) |
687 | | - assert _get_open_connections(self.client) == 1 |
688 | | - |
689 | | - for _ in response.iter_bytes(): |
690 | | - ... |
691 | | - |
692 | | - assert cast(Any, response.is_closed) |
693 | | - assert _get_open_connections(self.client) == 0 |
694 | | - |
695 | 667 | @mock.patch("openai._base_client.BaseClient._calculate_retry_timeout", _low_retry_timeout) |
696 | 668 | @pytest.mark.respx(base_url=base_url) |
697 | 669 | def test_retrying_timeout_errors_doesnt_leak(self, respx_mock: MockRouter) -> None: |
@@ -1372,33 +1344,6 @@ async def test_parse_retry_after_header(self, remaining_retries: int, retry_afte |
1372 | 1344 | calculated = client._calculate_retry_timeout(remaining_retries, options, headers) |
1373 | 1345 | assert calculated == pytest.approx(timeout, 0.5 * 0.875) # pyright: ignore[reportUnknownMemberType] |
1374 | 1346 |
|
1375 | | - @mock.patch("openai._base_client.BaseClient._calculate_retry_timeout", _low_retry_timeout) |
1376 | | - @pytest.mark.respx(base_url=base_url) |
1377 | | - async def test_streaming_response(self) -> None: |
1378 | | - response = await self.client.post( |
1379 | | - "/chat/completions", |
1380 | | - body=dict( |
1381 | | - messages=[ |
1382 | | - { |
1383 | | - "role": "user", |
1384 | | - "content": "Say this is a test", |
1385 | | - } |
1386 | | - ], |
1387 | | - model="gpt-3.5-turbo", |
1388 | | - ), |
1389 | | - cast_to=AsyncAPIResponse[bytes], |
1390 | | - options={"headers": {RAW_RESPONSE_HEADER: "stream"}}, |
1391 | | - ) |
1392 | | - |
1393 | | - assert not cast(Any, response.is_closed) |
1394 | | - assert _get_open_connections(self.client) == 1 |
1395 | | - |
1396 | | - async for _ in response.iter_bytes(): |
1397 | | - ... |
1398 | | - |
1399 | | - assert cast(Any, response.is_closed) |
1400 | | - assert _get_open_connections(self.client) == 0 |
1401 | | - |
1402 | 1347 | @mock.patch("openai._base_client.BaseClient._calculate_retry_timeout", _low_retry_timeout) |
1403 | 1348 | @pytest.mark.respx(base_url=base_url) |
1404 | 1349 | async def test_retrying_timeout_errors_doesnt_leak(self, respx_mock: MockRouter) -> None: |
|
0 commit comments