Fix qwen3_vl mix precision dtype #41701
Open
+8
−4
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
fast_pos_embed_interpolatereturnspos_embedsin the same dtype as the master weights.Therefore, when the master weights are in FP32 but the forward pass runs in BF16,
hidden_stateswill be upcast to FP32, causing dtype mismatches with other activations.CC @yonigozlan @molbap @zucchini-nlp