Skip to content

[Feature]Use Megatron-core dist_checkpointing to load checkpoint with different parallel strategies #169

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Open
SeaOfOcean opened this issue Dec 5, 2024 · 0 comments

Comments

@SeaOfOcean
Copy link
Collaborator

Is your feature request related to a problem? Please describe.
Currently ChatLearn use a tool to convert checkpoint when different parallel strategies are detected, https://github.com/alibaba/ChatLearn/blob/main/chatlearn/utils/megatron_utils.py#L164

The online conversion has been addressed in Megatron core dist_checkpointing. https://docs.nvidia.com/megatron-core/developer-guide/latest/api-guide/dist_checkpointing.html

Describe the solution you'd like
Use Megatron-core dist checkpointing to save and load checkpoint.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant