You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hi, by reading through the documentation, I am quite confused on the score_mod and block_mask argument of the flex attention api. They seem to be callables. I am wondering, is there a way I can provide a customized attention mask to the flex attention API just like attn_mask argument in torch.nn.functional.scaled_dot_product_attention? The reason I am asking is because my attention mask is very irregular (like in the attached image, white squares are masked positions) which is used in the encoder self-attention in Llama3.2 vision models (aka, aspect ratio mask).
Thank you!
The text was updated successfully, but these errors were encountered:
Hi, by reading through the documentation, I am quite confused on the score_mod and block_mask argument of the flex attention api. They seem to be callables. I am wondering, is there a way I can provide a customized attention mask to the flex attention API just like attn_mask argument in torch.nn.functional.scaled_dot_product_attention? The reason I am asking is because my attention mask is very irregular (like in the attached image, white squares are masked positions) which is used in the encoder self-attention in Llama3.2 vision models (aka, aspect ratio mask).
Thank you!
The text was updated successfully, but these errors were encountered: