Skip to main content

Blog

  1. Hacking "vanilla" FlashAttention for variable-length inputs
  2. Visualizing equivariances in transformer neural networks
  3. Mapping travels with Folium