-
Notifications
You must be signed in to change notification settings - Fork 507
Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
- Loading branch information
1 parent
daf5546
commit 29f4c10
Showing
3 changed files
with
117 additions
and
0 deletions.
There are no files selected for viewing
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,70 @@ | ||
# SOME DESCRIPTIVE TITLE. | ||
# Copyright (C) 2023, Xorbits Inc. | ||
# This file is distributed under the same license as the Xinference package. | ||
# FIRST AUTHOR <EMAIL@ADDRESS>, 2024. | ||
# | ||
#, fuzzy | ||
msgid "" | ||
msgstr "" | ||
"Project-Id-Version: Xinference \n" | ||
"Report-Msgid-Bugs-To: \n" | ||
"POT-Creation-Date: 2024-03-08 12:40+0800\n" | ||
"PO-Revision-Date: YEAR-MO-DA HO:MI+ZONE\n" | ||
"Last-Translator: FULL NAME <EMAIL@ADDRESS>\n" | ||
"Language-Team: LANGUAGE <[email protected]>\n" | ||
"MIME-Version: 1.0\n" | ||
"Content-Type: text/plain; charset=utf-8\n" | ||
"Content-Transfer-Encoding: 8bit\n" | ||
"Generated-By: Babel 2.11.0\n" | ||
|
||
#: ../../source/models/lora.rst:5 | ||
msgid "LoRA Integration" | ||
msgstr "集成LoRA" | ||
|
||
#: ../../source/models/lora.rst:7 | ||
msgid "" | ||
"Currently, Xinference supports launching ``LLM`` and ``image`` models " | ||
"with an attached LoRA fine-tuned model." | ||
msgstr "当前,Xinference 可以在启动 ``LLM`` 和 ``image`` 模型时连带一个 LoRA 微调模型用以辅助基础模型。" | ||
|
||
#: ../../source/models/lora.rst:10 | ||
msgid "Usage" | ||
msgstr "使用方式" | ||
|
||
#: ../../source/models/lora.rst:11 | ||
msgid "" | ||
"Different from built-in models, xinference currently does not involve " | ||
"managing LoRA models. Users need to first download the LoRA model " | ||
"themselves and then provide the storage path of the model files to " | ||
"xinference." | ||
msgstr "" | ||
"不同于内置模型,Xinference 目前不会涉及管理 LoRA 模型。用户需要首先下载对应的 LoRA 模型," | ||
"然后将模型存储路径提供给 Xinference 。" | ||
|
||
#: ../../source/models/lora.rst:38 | ||
msgid "Note" | ||
msgstr "注意事项" | ||
|
||
#: ../../source/models/lora.rst:40 | ||
msgid "" | ||
"The options ``image_lora_load_kwargs`` and ``image_lora_fuse_kwargs`` are" | ||
" only applicable to models with model_type ``image``. They correspond to " | ||
"the parameters in the ``load_lora_weights`` and ``fuse_lora`` interfaces " | ||
"of the ``diffusers`` library. If launching an LLM model, these parameters" | ||
" are not required." | ||
msgstr "" | ||
"上述 ``image_lora_load_kwargs`` 和 ``image_lora_fuse_kwargs`` 选项只应用于 ``image`` 模型。它们对应" | ||
"于 ``diffusers`` 库的 ``load_lora_weights`` 和 ``fuse_lora`` 接口中的额外参数。" | ||
"如果启动的是 ``LLM`` 模型,则无需设置这些选项。" | ||
|
||
#: ../../source/models/lora.rst:44 | ||
msgid "" | ||
"For LLM chat models, currently only LoRA models are supported that do not" | ||
" change the prompt style." | ||
msgstr "" | ||
"对于 ``LLM`` 聊天模型,当前仅支持那些微调后不更改原始基础模型的提示词模版的 LoRA 模型。" | ||
|
||
#: ../../source/models/lora.rst:46 | ||
msgid "When using GPU, both LoRA and its base model occupy the same devices." | ||
msgstr "" | ||
"使用 GPU 时,LoRA 模型与其基础模型在同样的设备上,不会对其他模型造成影响。" |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
|
@@ -214,3 +214,4 @@ Model Usage | |
builtin/index | ||
custom | ||
sources/sources | ||
lora |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,46 @@ | ||
.. _lora: | ||
|
||
================ | ||
LoRA Integration | ||
================ | ||
|
||
Currently, Xinference supports launching ``LLM`` and ``image`` models with an attached LoRA fine-tuned model. | ||
|
||
Usage | ||
^^^^^ | ||
Different from built-in models, xinference currently does not involve managing LoRA models. | ||
Users need to first download the LoRA model themselves and then provide the storage path of the model files to xinference. | ||
|
||
.. tabs:: | ||
|
||
.. code-tab:: bash shell | ||
|
||
xinference launch <options> --peft-model-path <lora_model_path> | ||
--image-lora-load-kwargs <load_params1> <load_value1> | ||
--image-lora-load-kwargs <load_params2> <load_value2> | ||
--image-lora-fuse-kwargs <fuse_params1> <fuse_value1> | ||
--image-lora-fuse-kwargs <fuse_params2> <fuse_value2> | ||
|
||
.. code-tab:: python | ||
|
||
from xinference.client import Client | ||
|
||
client = Client("http://<XINFERENCE_HOST>:<XINFERENCE_PORT>") | ||
client.launch_model( | ||
<other_options>, | ||
peft_model_path='<lora_model_path>', | ||
image_lora_load_kwargs={'<load_params1>': <load_value1>, '<load_params2>': <load_value2>}, | ||
image_lora_fuse_kwargs={'<fuse_params1>': <fuse_value1>, '<fuse_params2>': <fuse_value2>} | ||
) | ||
|
||
|
||
Note | ||
^^^^ | ||
|
||
* The options ``image_lora_load_kwargs`` and ``image_lora_fuse_kwargs`` are only applicable to models with model_type ``image``. | ||
They correspond to the parameters in the ``load_lora_weights`` and ``fuse_lora`` interfaces of the ``diffusers`` library. | ||
If launching an LLM model, these parameters are not required. | ||
|
||
* For LLM chat models, currently only LoRA models are supported that do not change the prompt style. | ||
|
||
* When using GPU, both LoRA and its base model occupy the same devices. |