Skip to content

Commit d1662ee

Browse files
authored
Update deepseek-r1-distill-llama-70b.mdx
1 parent 0fa843c commit d1662ee

File tree

1 file changed

+5
-7
lines changed

1 file changed

+5
-7
lines changed

pages/managed-inference/reference-content/deepseek-r1-distill-llama-70b.mdx

Lines changed: 5 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -19,22 +19,20 @@ categories:
1919
|-----------------|------------------------------------|
2020
| Provider | [Deepseek](https://huggingface.co/deepseek-ai/DeepSeek-R1-Distill-Llama-70B) |
2121
| License | [MIT](https://huggingface.co/datasets/choosealicense/licenses/blob/main/markdown/mit.md) |
22-
| Compatible Instances | H100 (FP8), H100-2 (FP8, BF16) |
23-
| Context Length | up to 32k tokens |
22+
| Compatible Instances | H100-2 (BF16) |
23+
| Context Length | up to 56k tokens |
2424

2525
## Model names
2626

2727
```bash
28-
meta/deepseek-r1-distill-llama-70b:fp8
29-
meta/deepseek-r1-distill-llama-70b:bf16
28+
deepseek/deepseek-r1-distill-llama-70b:bf16
3029
```
3130

3231
## Compatible Instances
3332

3433
| Instance type | Max context length |
3534
| ------------- |-------------|
36-
| H100 | 32k (FP8) |
37-
| H100-2 | 32k (FP8, BF16) |
35+
| H100-2 | 56k (BF16) |
3836

3937
## Model introduction
4038

@@ -46,7 +44,7 @@ DeepSeek R1 Distill Llama 70B is designed to improve performance of Llama models
4644
It is great to see Deepseek improving open(weight) models, and we are excited to fully support their mission with integration in the Scaleway ecosystem.
4745

4846
- DeepSeek-R1-Distill-Llama was optimized to reach accuracy close to Deepseek-R1 in tasks like mathematics and coding, while keeping inference costs limited and tokens speed efficient.
49-
- DeepSeek-R1-Distill-Llama supports a context window up to 32K tokens and tool calling, keeping interaction with other components possible.
47+
- DeepSeek-R1-Distill-Llama supports a context window up to 56K tokens and tool calling, keeping interaction with other components possible.
5048

5149
## How to use it
5250

0 commit comments

Comments
 (0)