Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Segmentation fault #127

Open
YueZhan721 opened this issue Oct 4, 2024 · 6 comments
Open

Segmentation fault #127

YueZhan721 opened this issue Oct 4, 2024 · 6 comments

Comments

@YueZhan721
Copy link

image
Hello, thank you very much for your work. I am having a problem as shown in the picture. My steps are 1. fine-tune the llama2 initial model with unsloth, save the weights in .safetensors format; 2. use the convert-hf.py script to convert to .m format and .t format; 3. run it on a single Raspberry Pi. Hope hear from you, thanks again.

@b4rtaz
Copy link
Owner

b4rtaz commented Oct 4, 2024

Hello @YueZhan721!

Could you run this model on a more powerful computer? How much RAM does your Raspberry Pi have?

@YueZhan721
Copy link
Author

Hello @b4rtaz I can run it on the google colab. The Raspberry Pi I used is 5 with 8GB RAM, the .m weight file is about 2.3 GB. I'm wondering if I'm doing something wrong with the quantization or formatting process.

@YueZhan721
Copy link
Author

By the way, I found that many llama3 models have no tokenizer.model file, how can I convert it from .safetensors to .m. Thanks for your reply very much.

@b4rtaz
Copy link
Owner

b4rtaz commented Oct 4, 2024

Llama 2 7B should be approximately 3.95GB after quantization to Q40. If it's only 2.3 GB, something might be wrong.

Could you run any model on your Raspberry Pi (for example python launch.py tinyllama_1_1b_3t_q40)?

@b4rtaz
Copy link
Owner

b4rtaz commented Oct 4, 2024

how can I convert it from .safetensors to .m. Thanks for your reply very much.

Check convert-tokenizer-hf.py. It can convert some HF models.

@YueZhan721
Copy link
Author

Llama 2 7B should be approximately 3.95GB after quantization to Q40. If it's only 2.3 GB, something might be wrong.

Could you run any model on your Raspberry Pi (for example python launch.py tinyllama_1_1b_3t_q40)?

Yes, I can run those models(.m file) you provided. There may be something wrong when I save and convert the weight files. Thanks sincerely! Good work, and help to me.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants