Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

couldn't find it in the cached files and it looks like Elron/bleurt-tiny-512, how to set the model path? #545

Open
bannima opened this issue Feb 8, 2025 · 1 comment

Comments

@bannima
Copy link

bannima commented Feb 8, 2025

How to set the eval model path?

Eval

when I use the script to eval model with MATH-500

`NUM_GPUS=8 # Set to 8 for 32B and 70B models
MODEL=Deepseek_R1_distill/Qwen2.5-32B-Open-R1-Distill/
MODEL_ARGS="pretrained=$MODEL,dtype=bfloat16,max_model_length=32768,gpu_memory_utilisation=0.8,tensor_parallel_size=$NUM_GPUS"
OUTPUT_DIR=data/evals/Qwen2.5-32B-Open-R1-Distill

lighteval vllm $MODEL_ARGS "custom|math_500|0|0"
--custom-tasks src/open_r1/evaluate.py
--use-chat-template
--output-dir $OUTPUT_DIR
`

Error

Error: We couldn't connect to 'https://huggingface.co' to load this file, couldn't find it in the cached files and it looks like Elron/bleurt-tiny-512 is not the path to a directory containing a file named
config.json.

Where to set the eval model path in the script?

@hynky1999
Copy link
Collaborator

You shouldn't need the model, it should be lazy loaded now #536

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants