You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Yes. There's also a cross attention version where q is separate by k, v are stacked into 1 tensor. Or you can call flash_attn_unpadded_func which takes in separate q, k, v tensors.
Hi, i‘m trying to use flashattn on cases where q,k ,v are of different values.
could someone confirm whether flashSelfAttn forward inputs (q, k, v)are a same tensor/values are same?
as the world self attn usually implies q,k,v are a same tensor.
code is here:
zoo/src/models/modules/mha.py
Lines 24 to 60 in 1417543
The text was updated successfully, but these errors were encountered: