Skip to content

Commit

Permalink
feat(ai): byom
Browse files Browse the repository at this point in the history
  • Loading branch information
bene2k1 committed Feb 10, 2025
1 parent 7c7bd99 commit 1bbfef6
Show file tree
Hide file tree
Showing 2 changed files with 48 additions and 0 deletions.
4 changes: 4 additions & 0 deletions menu/navigation.json
Original file line number Diff line number Diff line change
Expand Up @@ -745,6 +745,10 @@
"label": "Deploy a model",
"slug": "create-deployment"
},
{
"label": "Import a custom model",
"slug": "import-custom-model"
},
{
"label": "Monitor a deployment",
"slug": "monitor-deployment"
Expand Down
44 changes: 44 additions & 0 deletions pages/managed-inference/how-to/import-custom-model.mdx
Original file line number Diff line number Diff line change
@@ -0,0 +1,44 @@
meta:
title: How to import custom models into Managed Inference
description: Learn how to import your custom models into Scaleway's Managed Inference platform.
content:
h1: How to import custom models into Managed Inference
paragraph: Learn how to import your custom models into Scaleway's Managed Inference platform.
tags: managed-inference ai-data import custom model
dates:
validation: 2025-01-10
posted: 2025-01-10
categories:
- ai-data

---

Scaleway provides a selection of common models for deployment from the Scaleway console. If you need a specific model, you can import it into Managed Inference directly from Hugging Face or a Scaleway Object Storage bucket.

<Macro id="requirements" />
- A Scaleway account logged into the [console](https://console.scaleway.com).
- [Owner](/identity-and-access-management/iam/concepts/#owner) status or [IAM permissions](/identity-and-access-management/iam/concepts/#permission) to perform actions in your Organization.

1. Click **Managed Inference** in the **AI & Data** section of the side menu in the [Scaleway console](https://console.scaleway.com/) to access the dashboard.
2. Click **Deploy a model** to launch the model deployment wizard.
3. In the **Choose a model** section, select **Custom model**. If you have no model yet, click **Import a model** to start the model import wizard.
4. Choose an upload source:
- **Hugging Face**: Pull the model from Hugging Face.
- **Object Storage**: This feature is coming soon.
5. Enter your Hugging Face access token, which must have READ access to the repository.
<Message type="note">
Learn how to generate a Hugging Face access token [here](https://huggingface.co/docs/hub/security-tokens).
</Message>
6. Enter the name of the Hugging Face repository to pull the model from.
<Message type="note">
Ensure you have access to gated models if applicable. Refer to the [Hugging Face documentation](https://huggingface.co/docs/hub/en/models-gated) for details.
</Message>
7. Choose a name for your model. The name must be unique within your Organization and Project and cannot be changed later.
8. Click **Verify import** to check your Hugging Face credentials and ensure model compatibility.
9. Review the summary of your import, which includes:
- Context size by node type.
- Quantization options.
- Estimated cost.
Once checked, click **Begin import** to finalize the process.

Your imported model will now appear in the model library. You can proceed to [deploy your model on Managed Inference](/ai-data/managed-inference/how-to/create-deployment/).

0 comments on commit 1bbfef6

Please sign in to comment.