You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
| Agent / Plan | Qwen 3 Coder (480B), Qwen 3 Coder (30B), Devstral (24B), GLM 4.5 (355B), GLM 4.5 Air (106B), Kimi K2 (1T), gpt-oss (120B), gpt-oss (20B) |[Claude Opus 4.1](https://hub.continue.dev/anthropic/claude-4-1-opus), [Claude Sonnet 4](https://hub.continue.dev/anthropic/claude-4-sonnet), [GPT-5](https://hub.continue.dev/openai/gpt-5), [Gemini 2.5 Pro](https://hub.continue.dev/google/gemini-2.5-pro)| Closed models are slightly better than open models |
22
-
| Chat / Edit | Qwen 3 Coder (480B), Qwen 3 Coder (30B), gpt-oss (120B), gpt-oss (20B) |[Claude Opus 4.1](https://hub.continue.dev/anthropic/claude-4-1-opus), [Claude Sonnet 4](https://hub.continue.dev/anthropic/claude-4-sonnet), [GPT-5](https://hub.continue.dev/openai/gpt-5), [Gemini 2.5 Pro](https://hub.continue.dev/google/gemini-2.5-pro)| Closed and open models have pretty similar performance |
23
-
| Autocomplete |[QwenCoder2.5 (1.5B)](https://hub.continue.dev/ollama/qwen2.5-coder-1.5b), QwenCoder2.5 (7B) |[Codestral](https://hub.continue.dev/mistral/codestral), Mercury Coder | Closed models are slightly better than open models |
21
+
| Agent / Plan | Qwen 3 Coder (480B), Qwen 3 Coder (30B), Qwen2.5-Coder (32B), Devstral (27B), Devstral (24B), GLM 4.5 (355B), GLM 4.5 Air (106B), Kimi K2 (1T), gpt-oss (120B), gpt-oss (20B) |[Claude Opus 4.1](https://hub.continue.dev/anthropic/claude-4-1-opus), [Claude Sonnet 4](https://hub.continue.dev/anthropic/claude-4-sonnet), GPT-4, [GPT-5](https://hub.continue.dev/openai/gpt-5), [Gemini 2.5 Pro](https://hub.continue.dev/google/gemini-2.5-pro), DeepSeek models| Closed models are slightly better than open models |
22
+
| Chat / Edit | Qwen 3 Coder (480B), Qwen 3 Coder (30B), gpt-oss (120B), gpt-oss (20B), DeepSeek Chat|[Claude Opus 4.1](https://hub.continue.dev/anthropic/claude-4-1-opus), [Claude Sonnet 4](https://hub.continue.dev/anthropic/claude-4-sonnet), [GPT-5](https://hub.continue.dev/openai/gpt-5), [Gemini 2.5 Pro](https://hub.continue.dev/google/gemini-2.5-pro)| Closed and open models have pretty similar performance |
23
+
| Autocomplete |[QwenCoder2.5 (1.5B)](https://hub.continue.dev/ollama/qwen2.5-coder-1.5b), QwenCoder2.5 (7B) |[Codestral](https://hub.continue.dev/mistral/codestral), Mercury Coder, Mercury Coder Small, DeepSeek Coder| Closed models are slightly better than open models |
24
24
| Apply | N/A |[Relace Instant Apply](https://hub.continue.dev/relace/instant-apply), [Morph Fast Apply](https://hub.continue.dev/morphllm/morph-v2)| Open models are basically non-existent / not good enough for this model role |
25
-
| Embed |N/A |[Voyage Code 3](https://hub.continue.dev/voyageai/voyage-code-3), [Morph Embeddings](https://hub.continue.dev/morphllm/morph-embedding-v2), Codestral Embed| Open models are basically non-existent / not good enough for this model role|
25
+
| Embed |Nomic Embed Text |[Voyage Code 3](https://hub.continue.dev/voyageai/voyage-code-3), [Morph Embeddings](https://hub.continue.dev/morphllm/morph-embedding-v2), Codestral Embed, text-embedding-3-large, text-embedding-004 | Open embeddings models are emerging but closed models still perform better|
26
26
| Rerank | zerank-1, zerank-1-small | rerank-2.5, Relace Code Rerank, [Morph Rerank](https://hub.continue.dev/morphllm/morph-rerank-v2)| Open models are beginning to emerge for this model role |
27
27
| Next Edit | Zeta | Mercury Coder | Closed models are significantly better than open models |
Copy file name to clipboardExpand all lines: docs/features/autocomplete/model-setup.mdx
+1-39Lines changed: 1 addition & 39 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -10,45 +10,7 @@ Setting up the right model for autocomplete is crucial for a smooth coding exper
10
10
For a complete comparison of all autocomplete models, see our [comprehensive model recommendations](/customization/models#recommended-models).
11
11
</Info>
12
12
13
-
## Recommended Models for Autocomplete in Continue
14
-
15
-
### Hosted (Best Performance)
16
-
17
-
For the highest quality autocomplete suggestions, we recommend **[Codestral](https://hub.continue.dev/mistral/codestral)** from Mistral.
18
-
19
-
This model is specifically designed for code completion and offers excellent performance across multiple programming languages.
20
-
21
-
**Codestral Quick Setup:**
22
-
23
-
1. Get your API key from [Mistral AI](https://console.mistral.ai)
24
-
2. Add [Codestral](https://hub.continue.dev/mistral/codestral) to your assistant on Continue Hub
25
-
3. Add `MISTRAL_API_KEY` as a [User Secret](https://docs.continue.dev/hub/secrets/secret-types#user-secrets) on Continue Hub [here](https://hub.continue.dev/settings/secrets)
26
-
4. Click `Reload config` in the assistant selector in the Continue IDE extension
27
-
28
-
### Hosted (Best Speed/Quality Tradeoff)
29
-
30
-
For fast, quality autocomplete suggestions, we recommend **[Mercury Coder Small](https://hub.continue.dev/inceptionlabs/mercury-coder-small)** from Inception.
31
-
32
-
This model is specifically designed for code completion and is particularly fast because it is a diffusion model.
33
-
34
-
**Mercury Coder Small Quick Setup:**
35
-
36
-
1. Get your API key from [Inception](https://platform.inceptionlabs.ai/)
37
-
2. Add [Mercury Coder Small](https://hub.continue.dev/inceptionlabs/mercury-coder-small) to your assistant on Continue Hub
38
-
3. Add `INCEPTION_API_KEY` as a [User Secret](https://docs.continue.dev/hub/secrets/secret-types#user-secrets) on Continue Hub [here](https://hub.continue.dev/settings/secrets)
39
-
4. Click `Reload config` in the assistant selector in the Continue IDE extension
40
-
41
-
### Local (Offline / Privacy First)
42
-
43
-
For a fully local autocomplete experience, we recommend **[Qwen 2.5 Coder 1.5B](https://hub.continue.dev/ollama/qwen2.5-coder-1.5b)**.
44
-
45
-
This model provides good suggestions while keeping your code completely private.
46
-
47
-
**Quick Setup:**
48
-
49
-
1. Install [Ollama](https://ollama.ai/)
50
-
2. Add [Qwen 2.5 Coder 1.5B](https://hub.continue.dev/ollama/qwen2.5-coder-1.5b) to your assistant on Continue Hub
51
-
3. Click `Reload config` in the assistant selector in the Continue IDE extension
13
+
For model recommendations, please refer to our [Model Recommendations page](/customization/models).
Copy file name to clipboardExpand all lines: docs/features/chat/model-setup.mdx
+1-10Lines changed: 1 addition & 10 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -7,16 +7,7 @@ The model you use for for Chat mode will be
7
7
For a comprehensive comparison of all available models by role, see our [model recommendations table](/customization/models#recommended-models).
8
8
</Info>
9
9
10
-
## What Models Are Recommended for Chat?
11
-
12
-
Our strong recommendation is to use [Claude Sonnet 4](https://hub.continue.dev/anthropic/claude-4-sonnet) from Anthropic.
13
-
14
-
Its strong tool calling and reasoning capabilities make it the best model for Agent mode.
15
-
16
-
1. Get your API key from [Anthropic](https://console.anthropic.com/)
17
-
2. Add [Claude Sonnet 4](https://hub.continue.dev/anthropic/claude-4-sonnet) to your assistant on Continue Hub
18
-
3. Add `ANTHROPIC_API_KEY` as a [User Secret](https://docs.continue.dev/hub/secrets/secret-types#user-secrets) on Continue Hub [here](https://hub.continue.dev/settings/secrets)
19
-
4. Click `Reload config` in the assistant selector in the Continue IDE extension
10
+
For model recommendations, please refer to our [Model Recommendations page](/customization/models).
0 commit comments