Skip to content

Commit 966f933

Browse files
jeejeeleesimon-mo
authored andcommitted
[Bugfix] Fix LoRA extra vocab size (#15047)
Signed-off-by: Jee Jee Li <[email protected]>
1 parent 1a504af commit 966f933

File tree

5 files changed

+1
-5
lines changed

5 files changed

+1
-5
lines changed

examples/offline_inference/audio_language.py

-1
Original file line numberDiff line numberDiff line change
@@ -93,7 +93,6 @@ def run_phi4mm(question: str, audio_count: int) -> ModelRequestData:
9393
max_num_seqs=2,
9494
enable_lora=True,
9595
max_lora_rank=320,
96-
lora_extra_vocab_size=0,
9796
limit_mm_per_prompt={"audio": audio_count},
9897
)
9998

examples/offline_inference/vision_language.py

-1
Original file line numberDiff line numberDiff line change
@@ -682,7 +682,6 @@ def run_phi4mm(questions: list[str], modality: str) -> ModelRequestData:
682682
max_num_seqs=2,
683683
enable_lora=True,
684684
max_lora_rank=320,
685-
lora_extra_vocab_size=0,
686685
)
687686

688687
return ModelRequestData(

examples/offline_inference/vision_language_multi_image.py

-1
Original file line numberDiff line numberDiff line change
@@ -342,7 +342,6 @@ def load_phi4mm(question: str, image_urls: list[str]) -> ModelRequestData:
342342
limit_mm_per_prompt={"image": len(image_urls)},
343343
enable_lora=True,
344344
max_lora_rank=320,
345-
lora_extra_vocab_size=0,
346345
)
347346

348347
placeholders = "".join(f"<|image_{i}|>"

tests/models/decoder_only/vision_language/test_phi4mm.py

-1
Original file line numberDiff line numberDiff line change
@@ -100,7 +100,6 @@ def run_test(
100100
distributed_executor_backend=distributed_executor_backend,
101101
enable_lora=True,
102102
max_lora_rank=320,
103-
lora_extra_vocab_size=0,
104103
gpu_memory_utilization=0.8, # set to 0.8 to avoid OOM in CI
105104
enforce_eager=True,
106105
) as vllm_model:

vllm/config.py

+1-1
Original file line numberDiff line numberDiff line change
@@ -2324,7 +2324,7 @@ def __post_init__(self):
23242324
# Setting the maximum rank to 512 should be able to satisfy the vast
23252325
# majority of applications.
23262326
possible_max_ranks = (8, 16, 32, 64, 128, 256, 320, 512)
2327-
possible_lora_extra_vocab_size = (0, 256, 512)
2327+
possible_lora_extra_vocab_size = (256, 512)
23282328
if self.max_lora_rank not in possible_max_ranks:
23292329
raise ValueError(
23302330
f"max_lora_rank ({self.max_lora_rank}) must be one of "

0 commit comments

Comments
 (0)