-
Notifications
You must be signed in to change notification settings - Fork 119
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
OOM error of distributed training on 80GB GPUs with Mistral-7b #59
Comments
TracyPlus
changed the title
OOM
OOM error of distributed training on 80GB GPUs with Mistral-7b
Apr 8, 2024
I also encountered this problem, have you solved it now? @TracyPlus |
I also encountered this problem when i use Yarn by Llama2 |
1 similar comment
I also encountered this problem when i use Yarn by Llama2 |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
I run the following train.sh on Mistral-7b:
with accelerate config as:
but I encountered OutOfMemory problem on my 80G A800s:
I don't know if there's something wrong with my distributed training configuration、、🥺
Hope someone help me、、、🙏🙏
The text was updated successfully, but these errors were encountered: