pollen
by
josh
Search
Search
Dark mode
Light mode
Reader mode
Explorer
Home
❯
Readwise
❯
Podcasts
❯
FlashAttention 2 — Making Transformers 800% Faster WO Approximation With Tri Dao of Together AI
FlashAttention 2 — Making Transformers 800% Faster WO Approximation - With Tri Dao of Together AI
Sep 16, 2023
1 min read
1min Snip
Time 0:07:32
Author
Latent Space: The AI Engineer Podcast — CodeGen, Agents, Computer Vision, Data Science, AI UX and al…
Type
Podcast
Listen to episode
(share.snipd.com)
Graph View
Table of Contents
FlashAttention 2 — Making Transformers 800% Faster W/O Approximation - With Tri Dao of Together AI
Highlights