You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hello, I'm a newcomer in Diffusion generation. I'd like to ask why in the SpatialLinearAttention, the context is obtained first using 'k' and 'v', which seems different from the typical self-attention mechanism (where attention coefficients are computed using 'q' and 'k'). Is there a specific reason for this approach or other paper mention its use? I hope to receive your explanation. Thank you!
The text was updated successfully, but these errors were encountered:
Hello, I'm a newcomer in Diffusion generation. I'd like to ask why in the SpatialLinearAttention, the context is obtained first using 'k' and 'v', which seems different from the typical self-attention mechanism (where attention coefficients are computed using 'q' and 'k'). Is there a specific reason for this approach or other paper mention its use? I hope to receive your explanation. Thank you!
The text was updated successfully, but these errors were encountered: