@@ -136,22 +136,6 @@ type Assistant struct {
136
136
// assistant. Tools can be of types `code_interpreter`, `file_search`, or
137
137
// `function`.
138
138
Tools []AssistantTool `json:"tools,required"`
139
- // Specifies the format that the model must output. Compatible with
140
- // [GPT-4o](https://platform.openai.com/docs/models/gpt-4o),
141
- // [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-turbo-and-gpt-4),
142
- // and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
143
- //
144
- // Setting to `{ "type": "json_object" }` enables JSON mode, which guarantees the
145
- // message the model generates is valid JSON.
146
- //
147
- // **Important:** when using JSON mode, you **must** also instruct the model to
148
- // produce JSON yourself via a system or user message. Without this, the model may
149
- // generate an unending stream of whitespace until the generation reaches the token
150
- // limit, resulting in a long-running and seemingly "stuck" request. Also note that
151
- // the message content may be partially cut off if `finish_reason="length"`, which
152
- // indicates the generation exceeded `max_tokens` or the conversation exceeded the
153
- // max context length.
154
- ResponseFormat AssistantResponseFormatOptionUnion `json:"response_format,nullable"`
155
139
// What sampling temperature to use, between 0 and 2. Higher values like 0.8 will
156
140
// make the output more random, while lower values like 0.2 will make it more
157
141
// focused and deterministic.
@@ -172,21 +156,20 @@ type Assistant struct {
172
156
173
157
// assistantJSON contains the JSON metadata for the struct [Assistant]
174
158
type assistantJSON struct {
175
- ID apijson.Field
176
- CreatedAt apijson.Field
177
- Description apijson.Field
178
- Instructions apijson.Field
179
- Metadata apijson.Field
180
- Model apijson.Field
181
- Name apijson.Field
182
- Object apijson.Field
183
- Tools apijson.Field
184
- ResponseFormat apijson.Field
185
- Temperature apijson.Field
186
- ToolResources apijson.Field
187
- TopP apijson.Field
188
- raw string
189
- ExtraFields map [string ]apijson.Field
159
+ ID apijson.Field
160
+ CreatedAt apijson.Field
161
+ Description apijson.Field
162
+ Instructions apijson.Field
163
+ Metadata apijson.Field
164
+ Model apijson.Field
165
+ Name apijson.Field
166
+ Object apijson.Field
167
+ Tools apijson.Field
168
+ Temperature apijson.Field
169
+ ToolResources apijson.Field
170
+ TopP apijson.Field
171
+ raw string
172
+ ExtraFields map [string ]apijson.Field
190
173
}
191
174
192
175
func (r * Assistant ) UnmarshalJSON (data []byte ) (err error ) {
@@ -1869,8 +1852,8 @@ func (r FileSearchToolType) IsKnown() bool {
1869
1852
// Overrides for the file search tool.
1870
1853
type FileSearchToolFileSearch struct {
1871
1854
// The maximum number of results the file search tool should output. The default is
1872
- // 20 for gpt-4\* models and 5 for gpt-3.5-turbo. This number should be between 1
1873
- // and 50 inclusive.
1855
+ // 20 for ` gpt-4*` models and 5 for ` gpt-3.5-turbo` . This number should be between
1856
+ // 1 and 50 inclusive.
1874
1857
//
1875
1858
// Note that the file search tool may output fewer than `max_num_results` results.
1876
1859
// See the
@@ -1914,8 +1897,8 @@ func (r FileSearchToolParam) implementsBetaThreadNewAndRunParamsToolUnion() {}
1914
1897
// Overrides for the file search tool.
1915
1898
type FileSearchToolFileSearchParam struct {
1916
1899
// The maximum number of results the file search tool should output. The default is
1917
- // 20 for gpt-4\* models and 5 for gpt-3.5-turbo. This number should be between 1
1918
- // and 50 inclusive.
1900
+ // 20 for ` gpt-4*` models and 5 for ` gpt-3.5-turbo` . This number should be between
1901
+ // 1 and 50 inclusive.
1919
1902
//
1920
1903
// Note that the file search tool may output fewer than `max_num_results` results.
1921
1904
// See the
@@ -2001,22 +1984,6 @@ type BetaAssistantNewParams struct {
2001
1984
Metadata param.Field [interface {}] `json:"metadata"`
2002
1985
// The name of the assistant. The maximum length is 256 characters.
2003
1986
Name param.Field [string ] `json:"name"`
2004
- // Specifies the format that the model must output. Compatible with
2005
- // [GPT-4o](https://platform.openai.com/docs/models/gpt-4o),
2006
- // [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-turbo-and-gpt-4),
2007
- // and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
2008
- //
2009
- // Setting to `{ "type": "json_object" }` enables JSON mode, which guarantees the
2010
- // message the model generates is valid JSON.
2011
- //
2012
- // **Important:** when using JSON mode, you **must** also instruct the model to
2013
- // produce JSON yourself via a system or user message. Without this, the model may
2014
- // generate an unending stream of whitespace until the generation reaches the token
2015
- // limit, resulting in a long-running and seemingly "stuck" request. Also note that
2016
- // the message content may be partially cut off if `finish_reason="length"`, which
2017
- // indicates the generation exceeded `max_tokens` or the conversation exceeded the
2018
- // max context length.
2019
- ResponseFormat param.Field [AssistantResponseFormatOptionUnionParam ] `json:"response_format"`
2020
1987
// What sampling temperature to use, between 0 and 2. Higher values like 0.8 will
2021
1988
// make the output more random, while lower values like 0.2 will make it more
2022
1989
// focused and deterministic.
@@ -2234,22 +2201,6 @@ type BetaAssistantUpdateParams struct {
2234
2201
Model param.Field [string ] `json:"model"`
2235
2202
// The name of the assistant. The maximum length is 256 characters.
2236
2203
Name param.Field [string ] `json:"name"`
2237
- // Specifies the format that the model must output. Compatible with
2238
- // [GPT-4o](https://platform.openai.com/docs/models/gpt-4o),
2239
- // [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-turbo-and-gpt-4),
2240
- // and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
2241
- //
2242
- // Setting to `{ "type": "json_object" }` enables JSON mode, which guarantees the
2243
- // message the model generates is valid JSON.
2244
- //
2245
- // **Important:** when using JSON mode, you **must** also instruct the model to
2246
- // produce JSON yourself via a system or user message. Without this, the model may
2247
- // generate an unending stream of whitespace until the generation reaches the token
2248
- // limit, resulting in a long-running and seemingly "stuck" request. Also note that
2249
- // the message content may be partially cut off if `finish_reason="length"`, which
2250
- // indicates the generation exceeded `max_tokens` or the conversation exceeded the
2251
- // max context length.
2252
- ResponseFormat param.Field [AssistantResponseFormatOptionUnionParam ] `json:"response_format"`
2253
2204
// What sampling temperature to use, between 0 and 2. Higher values like 0.8 will
2254
2205
// make the output more random, while lower values like 0.2 will make it more
2255
2206
// focused and deterministic.
@@ -2335,7 +2286,7 @@ type BetaAssistantListParams struct {
2335
2286
// `url.Values`.
2336
2287
func (r BetaAssistantListParams ) URLQuery () (v url.Values ) {
2337
2288
return apiquery .MarshalWithSettings (r , apiquery.QuerySettings {
2338
- ArrayFormat : apiquery .ArrayQueryFormatComma ,
2289
+ ArrayFormat : apiquery .ArrayQueryFormatBrackets ,
2339
2290
NestedFormat : apiquery .NestedQueryFormatBrackets ,
2340
2291
})
2341
2292
}
0 commit comments