You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Standardizes parameter naming across attention functions
Changes `softmax_scale` to `scale` parameter name for consistency across CUDA, Triton, and Flex attention implementations.
Updates Flex attention to use keyword arguments and adds tensor transposes to match expected input format.
Removes unused return value from Flex attention call to align with other implementations.
0 commit comments