@@ -950,6 +950,7 @@ def stream(
950
950
max_prompt_tokens : Optional [int ] | NotGiven = NOT_GIVEN ,
951
951
metadata : Optional [object ] | NotGiven = NOT_GIVEN ,
952
952
model : Union [str , ChatModel , None ] | NotGiven = NOT_GIVEN ,
953
+ parallel_tool_calls : bool | NotGiven = NOT_GIVEN ,
953
954
response_format : Optional [AssistantResponseFormatOptionParam ] | NotGiven = NOT_GIVEN ,
954
955
temperature : Optional [float ] | NotGiven = NOT_GIVEN ,
955
956
tool_choice : Optional [AssistantToolChoiceOptionParam ] | NotGiven = NOT_GIVEN ,
@@ -979,6 +980,7 @@ def stream(
979
980
max_prompt_tokens : Optional [int ] | NotGiven = NOT_GIVEN ,
980
981
metadata : Optional [object ] | NotGiven = NOT_GIVEN ,
981
982
model : Union [str , ChatModel , None ] | NotGiven = NOT_GIVEN ,
983
+ parallel_tool_calls : bool | NotGiven = NOT_GIVEN ,
982
984
response_format : Optional [AssistantResponseFormatOptionParam ] | NotGiven = NOT_GIVEN ,
983
985
temperature : Optional [float ] | NotGiven = NOT_GIVEN ,
984
986
tool_choice : Optional [AssistantToolChoiceOptionParam ] | NotGiven = NOT_GIVEN ,
@@ -1008,6 +1010,7 @@ def stream(
1008
1010
max_prompt_tokens : Optional [int ] | NotGiven = NOT_GIVEN ,
1009
1011
metadata : Optional [object ] | NotGiven = NOT_GIVEN ,
1010
1012
model : Union [str , ChatModel , None ] | NotGiven = NOT_GIVEN ,
1013
+ parallel_tool_calls : bool | NotGiven = NOT_GIVEN ,
1011
1014
response_format : Optional [AssistantResponseFormatOptionParam ] | NotGiven = NOT_GIVEN ,
1012
1015
temperature : Optional [float ] | NotGiven = NOT_GIVEN ,
1013
1016
tool_choice : Optional [AssistantToolChoiceOptionParam ] | NotGiven = NOT_GIVEN ,
@@ -1051,6 +1054,7 @@ def stream(
1051
1054
"tool_choice" : tool_choice ,
1052
1055
"stream" : True ,
1053
1056
"tools" : tools ,
1057
+ "parallel_tool_calls" : parallel_tool_calls ,
1054
1058
"truncation_strategy" : truncation_strategy ,
1055
1059
"top_p" : top_p ,
1056
1060
},
@@ -2246,6 +2250,7 @@ def stream(
2246
2250
max_prompt_tokens : Optional [int ] | NotGiven = NOT_GIVEN ,
2247
2251
metadata : Optional [object ] | NotGiven = NOT_GIVEN ,
2248
2252
model : Union [str , ChatModel , None ] | NotGiven = NOT_GIVEN ,
2253
+ parallel_tool_calls : bool | NotGiven = NOT_GIVEN ,
2249
2254
response_format : Optional [AssistantResponseFormatOptionParam ] | NotGiven = NOT_GIVEN ,
2250
2255
temperature : Optional [float ] | NotGiven = NOT_GIVEN ,
2251
2256
tool_choice : Optional [AssistantToolChoiceOptionParam ] | NotGiven = NOT_GIVEN ,
@@ -2275,6 +2280,7 @@ def stream(
2275
2280
max_prompt_tokens : Optional [int ] | NotGiven = NOT_GIVEN ,
2276
2281
metadata : Optional [object ] | NotGiven = NOT_GIVEN ,
2277
2282
model : Union [str , ChatModel , None ] | NotGiven = NOT_GIVEN ,
2283
+ parallel_tool_calls : bool | NotGiven = NOT_GIVEN ,
2278
2284
response_format : Optional [AssistantResponseFormatOptionParam ] | NotGiven = NOT_GIVEN ,
2279
2285
temperature : Optional [float ] | NotGiven = NOT_GIVEN ,
2280
2286
tool_choice : Optional [AssistantToolChoiceOptionParam ] | NotGiven = NOT_GIVEN ,
@@ -2304,6 +2310,7 @@ def stream(
2304
2310
max_prompt_tokens : Optional [int ] | NotGiven = NOT_GIVEN ,
2305
2311
metadata : Optional [object ] | NotGiven = NOT_GIVEN ,
2306
2312
model : Union [str , ChatModel , None ] | NotGiven = NOT_GIVEN ,
2313
+ parallel_tool_calls : bool | NotGiven = NOT_GIVEN ,
2307
2314
response_format : Optional [AssistantResponseFormatOptionParam ] | NotGiven = NOT_GIVEN ,
2308
2315
temperature : Optional [float ] | NotGiven = NOT_GIVEN ,
2309
2316
tool_choice : Optional [AssistantToolChoiceOptionParam ] | NotGiven = NOT_GIVEN ,
@@ -2349,6 +2356,7 @@ def stream(
2349
2356
"tool_choice" : tool_choice ,
2350
2357
"stream" : True ,
2351
2358
"tools" : tools ,
2359
+ "parallel_tool_calls" : parallel_tool_calls ,
2352
2360
"truncation_strategy" : truncation_strategy ,
2353
2361
"top_p" : top_p ,
2354
2362
},
0 commit comments