We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
作者你好,
我在尝试使用change_mp.py将checkpoint拆分之后使用模型并行,但是在加载模型时提示word embedding大小不匹配。读过代码之后发现代码会在加载模型时将词表大小pad到某个数的整数倍(以提高计算效率),这个数是args.make_vocab_size_divisible_by * mpu.get_model_parallel_world_size(),因此MP改变时词表大小也会改变,导致无法正常加载模型参数。
args.make_vocab_size_divisible_by * mpu.get_model_parallel_world_size()
Chinese-Transformer-XL/pretrain_gpt2.py
Lines 669 to 677 in 0e702e4
一个temporary fix是将这里671行的multiple变量固定成args.make_vocab_size_divisible_by。
multiple
args.make_vocab_size_divisible_by
The text was updated successfully, but these errors were encountered:
请问你成功地进行finetune或者pretrain了吗?@Billijk 我这边通过下载链接得到的checkpoint在加载的时候会有runtime error
Sorry, something went wrong.
No branches or pull requests
作者你好,
我在尝试使用change_mp.py将checkpoint拆分之后使用模型并行,但是在加载模型时提示word embedding大小不匹配。读过代码之后发现代码会在加载模型时将词表大小pad到某个数的整数倍(以提高计算效率),这个数是
args.make_vocab_size_divisible_by * mpu.get_model_parallel_world_size()
,因此MP改变时词表大小也会改变,导致无法正常加载模型参数。Chinese-Transformer-XL/pretrain_gpt2.py
Lines 669 to 677 in 0e702e4
一个temporary fix是将这里671行的
multiple
变量固定成args.make_vocab_size_divisible_by
。The text was updated successfully, but these errors were encountered: