@@ -85,6 +85,7 @@ def create(
85
85
metadata : Optional [Metadata ] | NotGiven = NOT_GIVEN ,
86
86
model : Union [str , ChatModel , None ] | NotGiven = NOT_GIVEN ,
87
87
parallel_tool_calls : bool | NotGiven = NOT_GIVEN ,
88
+ reasoning_effort : Optional [Literal ["low" , "medium" , "high" ]] | NotGiven = NOT_GIVEN ,
88
89
response_format : Optional [AssistantResponseFormatOptionParam ] | NotGiven = NOT_GIVEN ,
89
90
stream : Optional [Literal [False ]] | NotGiven = NOT_GIVEN ,
90
91
temperature : Optional [float ] | NotGiven = NOT_GIVEN ,
@@ -153,6 +154,13 @@ def create(
153
154
[parallel function calling](https://platform.openai.com/docs/guides/function-calling#configuring-parallel-function-calling)
154
155
during tool use.
155
156
157
+ reasoning_effort: **o1 and o3-mini models only**
158
+
159
+ Constrains effort on reasoning for
160
+ [reasoning models](https://platform.openai.com/docs/guides/reasoning). Currently
161
+ supported values are `low`, `medium`, and `high`. Reducing reasoning effort can
162
+ result in faster responses and fewer tokens used on reasoning in a response.
163
+
156
164
response_format: Specifies the format that the model must output. Compatible with
157
165
[GPT-4o](https://platform.openai.com/docs/models#gpt-4o),
158
166
[GPT-4 Turbo](https://platform.openai.com/docs/models#gpt-4-turbo-and-gpt-4),
@@ -228,6 +236,7 @@ def create(
228
236
metadata : Optional [Metadata ] | NotGiven = NOT_GIVEN ,
229
237
model : Union [str , ChatModel , None ] | NotGiven = NOT_GIVEN ,
230
238
parallel_tool_calls : bool | NotGiven = NOT_GIVEN ,
239
+ reasoning_effort : Optional [Literal ["low" , "medium" , "high" ]] | NotGiven = NOT_GIVEN ,
231
240
response_format : Optional [AssistantResponseFormatOptionParam ] | NotGiven = NOT_GIVEN ,
232
241
temperature : Optional [float ] | NotGiven = NOT_GIVEN ,
233
242
tool_choice : Optional [AssistantToolChoiceOptionParam ] | NotGiven = NOT_GIVEN ,
@@ -299,6 +308,13 @@ def create(
299
308
[parallel function calling](https://platform.openai.com/docs/guides/function-calling#configuring-parallel-function-calling)
300
309
during tool use.
301
310
311
+ reasoning_effort: **o1 and o3-mini models only**
312
+
313
+ Constrains effort on reasoning for
314
+ [reasoning models](https://platform.openai.com/docs/guides/reasoning). Currently
315
+ supported values are `low`, `medium`, and `high`. Reducing reasoning effort can
316
+ result in faster responses and fewer tokens used on reasoning in a response.
317
+
302
318
response_format: Specifies the format that the model must output. Compatible with
303
319
[GPT-4o](https://platform.openai.com/docs/models#gpt-4o),
304
320
[GPT-4 Turbo](https://platform.openai.com/docs/models#gpt-4-turbo-and-gpt-4),
@@ -370,6 +386,7 @@ def create(
370
386
metadata : Optional [Metadata ] | NotGiven = NOT_GIVEN ,
371
387
model : Union [str , ChatModel , None ] | NotGiven = NOT_GIVEN ,
372
388
parallel_tool_calls : bool | NotGiven = NOT_GIVEN ,
389
+ reasoning_effort : Optional [Literal ["low" , "medium" , "high" ]] | NotGiven = NOT_GIVEN ,
373
390
response_format : Optional [AssistantResponseFormatOptionParam ] | NotGiven = NOT_GIVEN ,
374
391
temperature : Optional [float ] | NotGiven = NOT_GIVEN ,
375
392
tool_choice : Optional [AssistantToolChoiceOptionParam ] | NotGiven = NOT_GIVEN ,
@@ -441,6 +458,13 @@ def create(
441
458
[parallel function calling](https://platform.openai.com/docs/guides/function-calling#configuring-parallel-function-calling)
442
459
during tool use.
443
460
461
+ reasoning_effort: **o1 and o3-mini models only**
462
+
463
+ Constrains effort on reasoning for
464
+ [reasoning models](https://platform.openai.com/docs/guides/reasoning). Currently
465
+ supported values are `low`, `medium`, and `high`. Reducing reasoning effort can
466
+ result in faster responses and fewer tokens used on reasoning in a response.
467
+
444
468
response_format: Specifies the format that the model must output. Compatible with
445
469
[GPT-4o](https://platform.openai.com/docs/models#gpt-4o),
446
470
[GPT-4 Turbo](https://platform.openai.com/docs/models#gpt-4-turbo-and-gpt-4),
@@ -511,6 +535,7 @@ def create(
511
535
metadata : Optional [Metadata ] | NotGiven = NOT_GIVEN ,
512
536
model : Union [str , ChatModel , None ] | NotGiven = NOT_GIVEN ,
513
537
parallel_tool_calls : bool | NotGiven = NOT_GIVEN ,
538
+ reasoning_effort : Optional [Literal ["low" , "medium" , "high" ]] | NotGiven = NOT_GIVEN ,
514
539
response_format : Optional [AssistantResponseFormatOptionParam ] | NotGiven = NOT_GIVEN ,
515
540
stream : Optional [Literal [False ]] | Literal [True ] | NotGiven = NOT_GIVEN ,
516
541
temperature : Optional [float ] | NotGiven = NOT_GIVEN ,
@@ -541,6 +566,7 @@ def create(
541
566
"metadata" : metadata ,
542
567
"model" : model ,
543
568
"parallel_tool_calls" : parallel_tool_calls ,
569
+ "reasoning_effort" : reasoning_effort ,
544
570
"response_format" : response_format ,
545
571
"stream" : stream ,
546
572
"temperature" : temperature ,
@@ -941,6 +967,7 @@ async def create(
941
967
metadata : Optional [Metadata ] | NotGiven = NOT_GIVEN ,
942
968
model : Union [str , ChatModel , None ] | NotGiven = NOT_GIVEN ,
943
969
parallel_tool_calls : bool | NotGiven = NOT_GIVEN ,
970
+ reasoning_effort : Optional [Literal ["low" , "medium" , "high" ]] | NotGiven = NOT_GIVEN ,
944
971
response_format : Optional [AssistantResponseFormatOptionParam ] | NotGiven = NOT_GIVEN ,
945
972
stream : Optional [Literal [False ]] | NotGiven = NOT_GIVEN ,
946
973
temperature : Optional [float ] | NotGiven = NOT_GIVEN ,
@@ -1009,6 +1036,13 @@ async def create(
1009
1036
[parallel function calling](https://platform.openai.com/docs/guides/function-calling#configuring-parallel-function-calling)
1010
1037
during tool use.
1011
1038
1039
+ reasoning_effort: **o1 and o3-mini models only**
1040
+
1041
+ Constrains effort on reasoning for
1042
+ [reasoning models](https://platform.openai.com/docs/guides/reasoning). Currently
1043
+ supported values are `low`, `medium`, and `high`. Reducing reasoning effort can
1044
+ result in faster responses and fewer tokens used on reasoning in a response.
1045
+
1012
1046
response_format: Specifies the format that the model must output. Compatible with
1013
1047
[GPT-4o](https://platform.openai.com/docs/models#gpt-4o),
1014
1048
[GPT-4 Turbo](https://platform.openai.com/docs/models#gpt-4-turbo-and-gpt-4),
@@ -1084,6 +1118,7 @@ async def create(
1084
1118
metadata : Optional [Metadata ] | NotGiven = NOT_GIVEN ,
1085
1119
model : Union [str , ChatModel , None ] | NotGiven = NOT_GIVEN ,
1086
1120
parallel_tool_calls : bool | NotGiven = NOT_GIVEN ,
1121
+ reasoning_effort : Optional [Literal ["low" , "medium" , "high" ]] | NotGiven = NOT_GIVEN ,
1087
1122
response_format : Optional [AssistantResponseFormatOptionParam ] | NotGiven = NOT_GIVEN ,
1088
1123
temperature : Optional [float ] | NotGiven = NOT_GIVEN ,
1089
1124
tool_choice : Optional [AssistantToolChoiceOptionParam ] | NotGiven = NOT_GIVEN ,
@@ -1155,6 +1190,13 @@ async def create(
1155
1190
[parallel function calling](https://platform.openai.com/docs/guides/function-calling#configuring-parallel-function-calling)
1156
1191
during tool use.
1157
1192
1193
+ reasoning_effort: **o1 and o3-mini models only**
1194
+
1195
+ Constrains effort on reasoning for
1196
+ [reasoning models](https://platform.openai.com/docs/guides/reasoning). Currently
1197
+ supported values are `low`, `medium`, and `high`. Reducing reasoning effort can
1198
+ result in faster responses and fewer tokens used on reasoning in a response.
1199
+
1158
1200
response_format: Specifies the format that the model must output. Compatible with
1159
1201
[GPT-4o](https://platform.openai.com/docs/models#gpt-4o),
1160
1202
[GPT-4 Turbo](https://platform.openai.com/docs/models#gpt-4-turbo-and-gpt-4),
@@ -1226,6 +1268,7 @@ async def create(
1226
1268
metadata : Optional [Metadata ] | NotGiven = NOT_GIVEN ,
1227
1269
model : Union [str , ChatModel , None ] | NotGiven = NOT_GIVEN ,
1228
1270
parallel_tool_calls : bool | NotGiven = NOT_GIVEN ,
1271
+ reasoning_effort : Optional [Literal ["low" , "medium" , "high" ]] | NotGiven = NOT_GIVEN ,
1229
1272
response_format : Optional [AssistantResponseFormatOptionParam ] | NotGiven = NOT_GIVEN ,
1230
1273
temperature : Optional [float ] | NotGiven = NOT_GIVEN ,
1231
1274
tool_choice : Optional [AssistantToolChoiceOptionParam ] | NotGiven = NOT_GIVEN ,
@@ -1297,6 +1340,13 @@ async def create(
1297
1340
[parallel function calling](https://platform.openai.com/docs/guides/function-calling#configuring-parallel-function-calling)
1298
1341
during tool use.
1299
1342
1343
+ reasoning_effort: **o1 and o3-mini models only**
1344
+
1345
+ Constrains effort on reasoning for
1346
+ [reasoning models](https://platform.openai.com/docs/guides/reasoning). Currently
1347
+ supported values are `low`, `medium`, and `high`. Reducing reasoning effort can
1348
+ result in faster responses and fewer tokens used on reasoning in a response.
1349
+
1300
1350
response_format: Specifies the format that the model must output. Compatible with
1301
1351
[GPT-4o](https://platform.openai.com/docs/models#gpt-4o),
1302
1352
[GPT-4 Turbo](https://platform.openai.com/docs/models#gpt-4-turbo-and-gpt-4),
@@ -1367,6 +1417,7 @@ async def create(
1367
1417
metadata : Optional [Metadata ] | NotGiven = NOT_GIVEN ,
1368
1418
model : Union [str , ChatModel , None ] | NotGiven = NOT_GIVEN ,
1369
1419
parallel_tool_calls : bool | NotGiven = NOT_GIVEN ,
1420
+ reasoning_effort : Optional [Literal ["low" , "medium" , "high" ]] | NotGiven = NOT_GIVEN ,
1370
1421
response_format : Optional [AssistantResponseFormatOptionParam ] | NotGiven = NOT_GIVEN ,
1371
1422
stream : Optional [Literal [False ]] | Literal [True ] | NotGiven = NOT_GIVEN ,
1372
1423
temperature : Optional [float ] | NotGiven = NOT_GIVEN ,
@@ -1397,6 +1448,7 @@ async def create(
1397
1448
"metadata" : metadata ,
1398
1449
"model" : model ,
1399
1450
"parallel_tool_calls" : parallel_tool_calls ,
1451
+ "reasoning_effort" : reasoning_effort ,
1400
1452
"response_format" : response_format ,
1401
1453
"stream" : stream ,
1402
1454
"temperature" : temperature ,
0 commit comments