Skip to content

Commit 2d4da01

Browse files
feat(api): update urls for endpoints
1 parent 36b6036 commit 2d4da01

19 files changed

+761
-210
lines changed

.stats.yml

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -1,4 +1,4 @@
11
configured_endpoints: 193
2-
openapi_spec_url: https://storage.googleapis.com/stainless-sdk-openapi-specs/digitalocean%2Fgradient-2344b44246a44d39ad5b74d3077bd2958745aad67feb15970756532fa0b3f9d6.yml
3-
openapi_spec_hash: a1913979235ce152a8dc380fabe5362e
4-
config_hash: 13e570f98198e8fd1dfcb7ca59d73e0d
2+
openapi_spec_url: https://storage.googleapis.com/stainless-sdk-openapi-specs/digitalocean%2Fgradient-93df5cb3d61c1bbcab2a4bbf3f0775454bd185a400bc140c61dbf36e108dbf89.yml
3+
openapi_spec_hash: 0126adf782c5feac3d8e682e466c3cf7
4+
config_hash: 3f968a57adb20643373c134efc9af01a

src/gradient/resources/agents/chat/completions.py

Lines changed: 28 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -64,6 +64,7 @@ def create(
6464
metadata: Optional[Dict[str, str]] | Omit = omit,
6565
n: Optional[int] | Omit = omit,
6666
presence_penalty: Optional[float] | Omit = omit,
67+
reasoning_effort: Optional[Literal["none", "minimal", "low", "medium", "high", "xhigh"]] | Omit = omit,
6768
stop: Union[Optional[str], SequenceNotStr[str], None] | Omit = omit,
6869
stream: Optional[Literal[False]] | Omit = omit,
6970
stream_options: Optional[completion_create_params.StreamOptions] | Omit = omit,
@@ -129,6 +130,9 @@ def create(
129130
whether they appear in the text so far, increasing the model's likelihood to
130131
talk about new topics.
131132
133+
reasoning_effort: Constrains effort on reasoning for reasoning models. Reducing reasoning effort
134+
can result in faster responses and fewer tokens used on reasoning in a response.
135+
132136
stop: Up to 4 sequences where the API will stop generating further tokens. The
133137
returned text will not contain the stop sequence.
134138
@@ -193,6 +197,7 @@ def create(
193197
metadata: Optional[Dict[str, str]] | Omit = omit,
194198
n: Optional[int] | Omit = omit,
195199
presence_penalty: Optional[float] | Omit = omit,
200+
reasoning_effort: Optional[Literal["none", "minimal", "low", "medium", "high", "xhigh"]] | Omit = omit,
196201
stop: Union[Optional[str], SequenceNotStr[str], None] | Omit = omit,
197202
stream_options: Optional[completion_create_params.StreamOptions] | Omit = omit,
198203
temperature: Optional[float] | Omit = omit,
@@ -260,6 +265,9 @@ def create(
260265
whether they appear in the text so far, increasing the model's likelihood to
261266
talk about new topics.
262267
268+
reasoning_effort: Constrains effort on reasoning for reasoning models. Reducing reasoning effort
269+
can result in faster responses and fewer tokens used on reasoning in a response.
270+
263271
stop: Up to 4 sequences where the API will stop generating further tokens. The
264272
returned text will not contain the stop sequence.
265273
@@ -321,6 +329,7 @@ def create(
321329
metadata: Optional[Dict[str, str]] | Omit = omit,
322330
n: Optional[int] | Omit = omit,
323331
presence_penalty: Optional[float] | Omit = omit,
332+
reasoning_effort: Optional[Literal["none", "minimal", "low", "medium", "high", "xhigh"]] | Omit = omit,
324333
stop: Union[Optional[str], SequenceNotStr[str], None] | Omit = omit,
325334
stream_options: Optional[completion_create_params.StreamOptions] | Omit = omit,
326335
temperature: Optional[float] | Omit = omit,
@@ -388,6 +397,9 @@ def create(
388397
whether they appear in the text so far, increasing the model's likelihood to
389398
talk about new topics.
390399
400+
reasoning_effort: Constrains effort on reasoning for reasoning models. Reducing reasoning effort
401+
can result in faster responses and fewer tokens used on reasoning in a response.
402+
391403
stop: Up to 4 sequences where the API will stop generating further tokens. The
392404
returned text will not contain the stop sequence.
393405
@@ -451,6 +463,7 @@ def create(
451463
metadata: Optional[Dict[str, str]] | Omit = omit,
452464
n: Optional[int] | Omit = omit,
453465
presence_penalty: Optional[float] | Omit = omit,
466+
reasoning_effort: Optional[Literal["none", "minimal", "low", "medium", "high", "xhigh"]] | Omit = omit,
454467
stop: Union[Optional[str], SequenceNotStr[str], None] | Omit = omit,
455468
stream: Optional[Literal[False]] | Literal[True] | Omit = omit,
456469
stream_options: Optional[completion_create_params.StreamOptions] | Omit = omit,
@@ -491,6 +504,7 @@ def create(
491504
"metadata": metadata,
492505
"n": n,
493506
"presence_penalty": presence_penalty,
507+
"reasoning_effort": reasoning_effort,
494508
"stop": stop,
495509
"stream": stream,
496510
"stream_options": stream_options,
@@ -557,6 +571,7 @@ async def create(
557571
metadata: Optional[Dict[str, str]] | Omit = omit,
558572
n: Optional[int] | Omit = omit,
559573
presence_penalty: Optional[float] | Omit = omit,
574+
reasoning_effort: Optional[Literal["none", "minimal", "low", "medium", "high", "xhigh"]] | Omit = omit,
560575
stop: Union[Optional[str], SequenceNotStr[str], None] | Omit = omit,
561576
stream: Optional[Literal[False]] | Omit = omit,
562577
stream_options: Optional[completion_create_params.StreamOptions] | Omit = omit,
@@ -622,6 +637,9 @@ async def create(
622637
whether they appear in the text so far, increasing the model's likelihood to
623638
talk about new topics.
624639
640+
reasoning_effort: Constrains effort on reasoning for reasoning models. Reducing reasoning effort
641+
can result in faster responses and fewer tokens used on reasoning in a response.
642+
625643
stop: Up to 4 sequences where the API will stop generating further tokens. The
626644
returned text will not contain the stop sequence.
627645
@@ -686,6 +704,7 @@ async def create(
686704
metadata: Optional[Dict[str, str]] | Omit = omit,
687705
n: Optional[int] | Omit = omit,
688706
presence_penalty: Optional[float] | Omit = omit,
707+
reasoning_effort: Optional[Literal["none", "minimal", "low", "medium", "high", "xhigh"]] | Omit = omit,
689708
stop: Union[Optional[str], SequenceNotStr[str], None] | Omit = omit,
690709
stream_options: Optional[completion_create_params.StreamOptions] | Omit = omit,
691710
temperature: Optional[float] | Omit = omit,
@@ -753,6 +772,9 @@ async def create(
753772
whether they appear in the text so far, increasing the model's likelihood to
754773
talk about new topics.
755774
775+
reasoning_effort: Constrains effort on reasoning for reasoning models. Reducing reasoning effort
776+
can result in faster responses and fewer tokens used on reasoning in a response.
777+
756778
stop: Up to 4 sequences where the API will stop generating further tokens. The
757779
returned text will not contain the stop sequence.
758780
@@ -814,6 +836,7 @@ async def create(
814836
metadata: Optional[Dict[str, str]] | Omit = omit,
815837
n: Optional[int] | Omit = omit,
816838
presence_penalty: Optional[float] | Omit = omit,
839+
reasoning_effort: Optional[Literal["none", "minimal", "low", "medium", "high", "xhigh"]] | Omit = omit,
817840
stop: Union[Optional[str], SequenceNotStr[str], None] | Omit = omit,
818841
stream_options: Optional[completion_create_params.StreamOptions] | Omit = omit,
819842
temperature: Optional[float] | Omit = omit,
@@ -881,6 +904,9 @@ async def create(
881904
whether they appear in the text so far, increasing the model's likelihood to
882905
talk about new topics.
883906
907+
reasoning_effort: Constrains effort on reasoning for reasoning models. Reducing reasoning effort
908+
can result in faster responses and fewer tokens used on reasoning in a response.
909+
884910
stop: Up to 4 sequences where the API will stop generating further tokens. The
885911
returned text will not contain the stop sequence.
886912
@@ -941,6 +967,7 @@ async def create(
941967
metadata: Optional[Dict[str, str]] | Omit = omit,
942968
n: Optional[int] | Omit = omit,
943969
presence_penalty: Optional[float] | Omit = omit,
970+
reasoning_effort: Optional[Literal["none", "minimal", "low", "medium", "high", "xhigh"]] | Omit = omit,
944971
stop: Union[Optional[str], SequenceNotStr[str], None] | Omit = omit,
945972
stream: Optional[Literal[False]] | Literal[True] | Omit = omit,
946973
stream_options: Optional[completion_create_params.StreamOptions] | Omit = omit,
@@ -981,6 +1008,7 @@ async def create(
9811008
"metadata": metadata,
9821009
"n": n,
9831010
"presence_penalty": presence_penalty,
1011+
"reasoning_effort": reasoning_effort,
9841012
"stop": stop,
9851013
"stream": stream,
9861014
"stream_options": stream_options,

src/gradient/resources/chat/completions.py

Lines changed: 28 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -64,6 +64,7 @@ def create(
6464
metadata: Optional[Dict[str, str]] | Omit = omit,
6565
n: Optional[int] | Omit = omit,
6666
presence_penalty: Optional[float] | Omit = omit,
67+
reasoning_effort: Optional[Literal["none", "minimal", "low", "medium", "high", "xhigh"]] | Omit = omit,
6768
stop: Union[Optional[str], SequenceNotStr[str], None] | Omit = omit,
6869
stream: Optional[Literal[False]] | Omit = omit,
6970
stream_options: Optional[completion_create_params.StreamOptions] | Omit = omit,
@@ -129,6 +130,9 @@ def create(
129130
whether they appear in the text so far, increasing the model's likelihood to
130131
talk about new topics.
131132
133+
reasoning_effort: Constrains effort on reasoning for reasoning models. Reducing reasoning effort
134+
can result in faster responses and fewer tokens used on reasoning in a response.
135+
132136
stop: Up to 4 sequences where the API will stop generating further tokens. The
133137
returned text will not contain the stop sequence.
134138
@@ -194,6 +198,7 @@ def create(
194198
metadata: Optional[Dict[str, str]] | Omit = omit,
195199
n: Optional[int] | Omit = omit,
196200
presence_penalty: Optional[float] | Omit = omit,
201+
reasoning_effort: Optional[Literal["none", "minimal", "low", "medium", "high", "xhigh"]] | Omit = omit,
197202
stop: Union[Optional[str], SequenceNotStr[str], None] | Omit = omit,
198203
stream_options: Optional[completion_create_params.StreamOptions] | Omit = omit,
199204
temperature: Optional[float] | Omit = omit,
@@ -261,6 +266,9 @@ def create(
261266
whether they appear in the text so far, increasing the model's likelihood to
262267
talk about new topics.
263268
269+
reasoning_effort: Constrains effort on reasoning for reasoning models. Reducing reasoning effort
270+
can result in faster responses and fewer tokens used on reasoning in a response.
271+
264272
stop: Up to 4 sequences where the API will stop generating further tokens. The
265273
returned text will not contain the stop sequence.
266274
@@ -322,6 +330,7 @@ def create(
322330
metadata: Optional[Dict[str, str]] | Omit = omit,
323331
n: Optional[int] | Omit = omit,
324332
presence_penalty: Optional[float] | Omit = omit,
333+
reasoning_effort: Optional[Literal["none", "minimal", "low", "medium", "high", "xhigh"]] | Omit = omit,
325334
stop: Union[Optional[str], SequenceNotStr[str], None] | Omit = omit,
326335
stream_options: Optional[completion_create_params.StreamOptions] | Omit = omit,
327336
temperature: Optional[float] | Omit = omit,
@@ -389,6 +398,9 @@ def create(
389398
whether they appear in the text so far, increasing the model's likelihood to
390399
talk about new topics.
391400
401+
reasoning_effort: Constrains effort on reasoning for reasoning models. Reducing reasoning effort
402+
can result in faster responses and fewer tokens used on reasoning in a response.
403+
392404
stop: Up to 4 sequences where the API will stop generating further tokens. The
393405
returned text will not contain the stop sequence.
394406
@@ -449,6 +461,7 @@ def create(
449461
metadata: Optional[Dict[str, str]] | Omit = omit,
450462
n: Optional[int] | Omit = omit,
451463
presence_penalty: Optional[float] | Omit = omit,
464+
reasoning_effort: Optional[Literal["none", "minimal", "low", "medium", "high", "xhigh"]] | Omit = omit,
452465
stop: Union[Optional[str], SequenceNotStr[str], None] | Omit = omit,
453466
stream: Optional[Literal[False]] | Literal[True] | Omit = omit,
454467
stream_options: Optional[completion_create_params.StreamOptions] | Omit = omit,
@@ -492,6 +505,7 @@ def create(
492505
"metadata": metadata,
493506
"n": n,
494507
"presence_penalty": presence_penalty,
508+
"reasoning_effort": reasoning_effort,
495509
"stop": stop,
496510
"stream": stream,
497511
"stream_options": stream_options,
@@ -558,6 +572,7 @@ async def create(
558572
metadata: Optional[Dict[str, str]] | Omit = omit,
559573
n: Optional[int] | Omit = omit,
560574
presence_penalty: Optional[float] | Omit = omit,
575+
reasoning_effort: Optional[Literal["none", "minimal", "low", "medium", "high", "xhigh"]] | Omit = omit,
561576
stop: Union[Optional[str], SequenceNotStr[str], None] | Omit = omit,
562577
stream: Optional[Literal[False]] | Omit = omit,
563578
stream_options: Optional[completion_create_params.StreamOptions] | Omit = omit,
@@ -623,6 +638,9 @@ async def create(
623638
whether they appear in the text so far, increasing the model's likelihood to
624639
talk about new topics.
625640
641+
reasoning_effort: Constrains effort on reasoning for reasoning models. Reducing reasoning effort
642+
can result in faster responses and fewer tokens used on reasoning in a response.
643+
626644
stop: Up to 4 sequences where the API will stop generating further tokens. The
627645
returned text will not contain the stop sequence.
628646
@@ -688,6 +706,7 @@ async def create(
688706
metadata: Optional[Dict[str, str]] | Omit = omit,
689707
n: Optional[int] | Omit = omit,
690708
presence_penalty: Optional[float] | Omit = omit,
709+
reasoning_effort: Optional[Literal["none", "minimal", "low", "medium", "high", "xhigh"]] | Omit = omit,
691710
stop: Union[Optional[str], SequenceNotStr[str], None] | Omit = omit,
692711
stream_options: Optional[completion_create_params.StreamOptions] | Omit = omit,
693712
temperature: Optional[float] | Omit = omit,
@@ -755,6 +774,9 @@ async def create(
755774
whether they appear in the text so far, increasing the model's likelihood to
756775
talk about new topics.
757776
777+
reasoning_effort: Constrains effort on reasoning for reasoning models. Reducing reasoning effort
778+
can result in faster responses and fewer tokens used on reasoning in a response.
779+
758780
stop: Up to 4 sequences where the API will stop generating further tokens. The
759781
returned text will not contain the stop sequence.
760782
@@ -816,6 +838,7 @@ async def create(
816838
metadata: Optional[Dict[str, str]] | Omit = omit,
817839
n: Optional[int] | Omit = omit,
818840
presence_penalty: Optional[float] | Omit = omit,
841+
reasoning_effort: Optional[Literal["none", "minimal", "low", "medium", "high", "xhigh"]] | Omit = omit,
819842
stop: Union[Optional[str], SequenceNotStr[str], None] | Omit = omit,
820843
stream_options: Optional[completion_create_params.StreamOptions] | Omit = omit,
821844
temperature: Optional[float] | Omit = omit,
@@ -883,6 +906,9 @@ async def create(
883906
whether they appear in the text so far, increasing the model's likelihood to
884907
talk about new topics.
885908
909+
reasoning_effort: Constrains effort on reasoning for reasoning models. Reducing reasoning effort
910+
can result in faster responses and fewer tokens used on reasoning in a response.
911+
886912
stop: Up to 4 sequences where the API will stop generating further tokens. The
887913
returned text will not contain the stop sequence.
888914
@@ -943,6 +969,7 @@ async def create(
943969
metadata: Optional[Dict[str, str]] | Omit = omit,
944970
n: Optional[int] | Omit = omit,
945971
presence_penalty: Optional[float] | Omit = omit,
972+
reasoning_effort: Optional[Literal["none", "minimal", "low", "medium", "high", "xhigh"]] | Omit = omit,
946973
stop: Union[Optional[str], SequenceNotStr[str], None] | Omit = omit,
947974
stream: Optional[Literal[False]] | Literal[True] | Omit = omit,
948975
stream_options: Optional[completion_create_params.StreamOptions] | Omit = omit,
@@ -986,6 +1013,7 @@ async def create(
9861013
"metadata": metadata,
9871014
"n": n,
9881015
"presence_penalty": presence_penalty,
1016+
"reasoning_effort": reasoning_effort,
9891017
"stop": stop,
9901018
"stream": stream,
9911019
"stream_options": stream_options,

0 commit comments

Comments
 (0)