Skip to content

Commit aa49f62

Browse files
feat(api): comparison filter in/not in
1 parent aa445ca commit aa49f62

23 files changed

+129
-11
lines changed

.stats.yml

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -1,4 +1,4 @@
11
configured_endpoints: 136
2-
openapi_spec_url: https://storage.googleapis.com/stainless-sdk-openapi-specs/openai%2Fopenai-d64cf80d2ebddf175c5578f68226a3d5bbd3f7fd8d62ccac2205f3fc05a355ee.yml
3-
openapi_spec_hash: d51e0d60d0c536f210b597a211bc5af0
4-
config_hash: e7c42016df9c6bd7bd6ff15101b9bc9b
2+
openapi_spec_url: https://storage.googleapis.com/stainless-sdk-openapi-specs/openai%2Fopenai-e66e85fb7f72477256dca1acb6b23396989d381c5c1b318de564195436bcb93f.yml
3+
openapi_spec_hash: 0a4bbb5aa0ae532a072bd6b3854e70b1
4+
config_hash: 89bf7bb3a1f9439ffc6ea0e7dc57ba9b

src/openai/resources/beta/assistants.py

Lines changed: 12 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -102,6 +102,9 @@ def create(
102102
effort can result in faster responses and fewer tokens used on reasoning in a
103103
response.
104104
105+
Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
106+
effort.
107+
105108
response_format: Specifies the format that the model must output. Compatible with
106109
[GPT-4o](https://platform.openai.com/docs/models#gpt-4o),
107110
[GPT-4 Turbo](https://platform.openai.com/docs/models#gpt-4-turbo-and-gpt-4),
@@ -309,6 +312,9 @@ def update(
309312
effort can result in faster responses and fewer tokens used on reasoning in a
310313
response.
311314
315+
Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
316+
effort.
317+
312318
response_format: Specifies the format that the model must output. Compatible with
313319
[GPT-4o](https://platform.openai.com/docs/models#gpt-4o),
314320
[GPT-4 Turbo](https://platform.openai.com/docs/models#gpt-4-turbo-and-gpt-4),
@@ -555,6 +561,9 @@ async def create(
555561
effort can result in faster responses and fewer tokens used on reasoning in a
556562
response.
557563
564+
Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
565+
effort.
566+
558567
response_format: Specifies the format that the model must output. Compatible with
559568
[GPT-4o](https://platform.openai.com/docs/models#gpt-4o),
560569
[GPT-4 Turbo](https://platform.openai.com/docs/models#gpt-4-turbo-and-gpt-4),
@@ -762,6 +771,9 @@ async def update(
762771
effort can result in faster responses and fewer tokens used on reasoning in a
763772
response.
764773
774+
Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
775+
effort.
776+
765777
response_format: Specifies the format that the model must output. Compatible with
766778
[GPT-4o](https://platform.openai.com/docs/models#gpt-4o),
767779
[GPT-4 Turbo](https://platform.openai.com/docs/models#gpt-4-turbo-and-gpt-4),

src/openai/resources/beta/threads/runs/runs.py

Lines changed: 18 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -173,6 +173,9 @@ def create(
173173
effort can result in faster responses and fewer tokens used on reasoning in a
174174
response.
175175
176+
Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
177+
effort.
178+
176179
response_format: Specifies the format that the model must output. Compatible with
177180
[GPT-4o](https://platform.openai.com/docs/models#gpt-4o),
178181
[GPT-4 Turbo](https://platform.openai.com/docs/models#gpt-4-turbo-and-gpt-4),
@@ -327,6 +330,9 @@ def create(
327330
effort can result in faster responses and fewer tokens used on reasoning in a
328331
response.
329332
333+
Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
334+
effort.
335+
330336
response_format: Specifies the format that the model must output. Compatible with
331337
[GPT-4o](https://platform.openai.com/docs/models#gpt-4o),
332338
[GPT-4 Turbo](https://platform.openai.com/docs/models#gpt-4-turbo-and-gpt-4),
@@ -477,6 +483,9 @@ def create(
477483
effort can result in faster responses and fewer tokens used on reasoning in a
478484
response.
479485
486+
Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
487+
effort.
488+
480489
response_format: Specifies the format that the model must output. Compatible with
481490
[GPT-4o](https://platform.openai.com/docs/models#gpt-4o),
482491
[GPT-4 Turbo](https://platform.openai.com/docs/models#gpt-4-turbo-and-gpt-4),
@@ -1603,6 +1612,9 @@ async def create(
16031612
effort can result in faster responses and fewer tokens used on reasoning in a
16041613
response.
16051614
1615+
Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
1616+
effort.
1617+
16061618
response_format: Specifies the format that the model must output. Compatible with
16071619
[GPT-4o](https://platform.openai.com/docs/models#gpt-4o),
16081620
[GPT-4 Turbo](https://platform.openai.com/docs/models#gpt-4-turbo-and-gpt-4),
@@ -1757,6 +1769,9 @@ async def create(
17571769
effort can result in faster responses and fewer tokens used on reasoning in a
17581770
response.
17591771
1772+
Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
1773+
effort.
1774+
17601775
response_format: Specifies the format that the model must output. Compatible with
17611776
[GPT-4o](https://platform.openai.com/docs/models#gpt-4o),
17621777
[GPT-4 Turbo](https://platform.openai.com/docs/models#gpt-4-turbo-and-gpt-4),
@@ -1907,6 +1922,9 @@ async def create(
19071922
effort can result in faster responses and fewer tokens used on reasoning in a
19081923
response.
19091924
1925+
Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
1926+
effort.
1927+
19101928
response_format: Specifies the format that the model must output. Compatible with
19111929
[GPT-4o](https://platform.openai.com/docs/models#gpt-4o),
19121930
[GPT-4 Turbo](https://platform.openai.com/docs/models#gpt-4-turbo-and-gpt-4),

src/openai/resources/chat/completions/completions.py

Lines changed: 18 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -407,6 +407,9 @@ def create(
407407
effort can result in faster responses and fewer tokens used on reasoning in a
408408
response.
409409
410+
Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
411+
effort.
412+
410413
response_format: An object specifying the format that the model must output.
411414
412415
Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured
@@ -704,6 +707,9 @@ def create(
704707
effort can result in faster responses and fewer tokens used on reasoning in a
705708
response.
706709
710+
Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
711+
effort.
712+
707713
response_format: An object specifying the format that the model must output.
708714
709715
Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured
@@ -992,6 +998,9 @@ def create(
992998
effort can result in faster responses and fewer tokens used on reasoning in a
993999
response.
9941000
1001+
Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
1002+
effort.
1003+
9951004
response_format: An object specifying the format that the model must output.
9961005
9971006
Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured
@@ -1845,6 +1854,9 @@ async def create(
18451854
effort can result in faster responses and fewer tokens used on reasoning in a
18461855
response.
18471856
1857+
Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
1858+
effort.
1859+
18481860
response_format: An object specifying the format that the model must output.
18491861
18501862
Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured
@@ -2142,6 +2154,9 @@ async def create(
21422154
effort can result in faster responses and fewer tokens used on reasoning in a
21432155
response.
21442156
2157+
Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
2158+
effort.
2159+
21452160
response_format: An object specifying the format that the model must output.
21462161
21472162
Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured
@@ -2430,6 +2445,9 @@ async def create(
24302445
effort can result in faster responses and fewer tokens used on reasoning in a
24312446
response.
24322447
2448+
Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
2449+
effort.
2450+
24332451
response_format: An object specifying the format that the model must output.
24342452
24352453
Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured

src/openai/resources/files.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -236,7 +236,7 @@ def delete(
236236
timeout: float | httpx.Timeout | None | NotGiven = not_given,
237237
) -> FileDeleted:
238238
"""
239-
Delete a file.
239+
Delete a file and remove it from all vector stores.
240240
241241
Args:
242242
extra_headers: Send extra headers
@@ -553,7 +553,7 @@ async def delete(
553553
timeout: float | httpx.Timeout | None | NotGiven = not_given,
554554
) -> FileDeleted:
555555
"""
556-
Delete a file.
556+
Delete a file and remove it from all vector stores.
557557
558558
Args:
559559
extra_headers: Send extra headers

src/openai/types/beta/assistant_create_params.py

Lines changed: 3 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -65,6 +65,9 @@ class AssistantCreateParams(TypedDict, total=False):
6565
supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning
6666
effort can result in faster responses and fewer tokens used on reasoning in a
6767
response.
68+
69+
Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
70+
effort.
6871
"""
6972

7073
response_format: Optional[AssistantResponseFormatOptionParam]

src/openai/types/beta/assistant_update_params.py

Lines changed: 3 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -100,6 +100,9 @@ class AssistantUpdateParams(TypedDict, total=False):
100100
supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning
101101
effort can result in faster responses and fewer tokens used on reasoning in a
102102
response.
103+
104+
Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
105+
effort.
103106
"""
104107

105108
response_format: Optional[AssistantResponseFormatOptionParam]

src/openai/types/beta/threads/run_create_params.py

Lines changed: 3 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -114,6 +114,9 @@ class RunCreateParamsBase(TypedDict, total=False):
114114
supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning
115115
effort can result in faster responses and fewer tokens used on reasoning in a
116116
response.
117+
118+
Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
119+
effort.
117120
"""
118121

119122
response_format: Optional[AssistantResponseFormatOptionParam]

src/openai/types/chat/completion_create_params.py

Lines changed: 3 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -192,6 +192,9 @@ class CompletionCreateParamsBase(TypedDict, total=False):
192192
supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning
193193
effort can result in faster responses and fewer tokens used on reasoning in a
194194
response.
195+
196+
Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
197+
effort.
195198
"""
196199

197200
response_format: ResponseFormat

src/openai/types/evals/create_eval_completions_run_data_source.py

Lines changed: 3 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -175,6 +175,9 @@ class SamplingParams(BaseModel):
175175
supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning
176176
effort can result in faster responses and fewer tokens used on reasoning in a
177177
response.
178+
179+
Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
180+
effort.
178181
"""
179182

180183
response_format: Optional[SamplingParamsResponseFormat] = None

0 commit comments

Comments
 (0)