Skip to content

KPK101/FlashAttention

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

7 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

FlashAttention

Implementation of FlashAttention in pycuda

Status

  • Simple attention mechanism implementation in python using numpy

To-Do (CPU): Target (11/05)

  • Include multi-headed attention
  • More modular and checks
  • Make a PyTorch attention module

To-Do (GPU): Target (11/05)

  • Implement naive attention computation
  • Add tiling to blocks for compute
  • Fused kernels (matmul, softmax, linear layer)

About

Implementation of FlashAttention in pycuda

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages