Dao AI Lab

Affiliations. Address. Contacts. Motto. Etc.

prof_pic.jpg

555 your office number

123 your address street

Your City, State 12345

Write your biography here. Tell the world about yourself. Link to your favorite subreddit. You can put a picture in, too. The code is already in, just name your picture prof_pic.jpg and put it in the img/ folder.

Put your address / P.O. box / other info right below your picture. You can also disable any of these elements by editing profile property of the YAML header of your _pages/about.md. Edit _bibliography/papers.bib and Jekyll will render your publications page automatically.

Link to your social media connections, too. This theme is set up to use Font Awesome icons and Academicons, like the ones below. Add your Facebook, Twitter, LinkedIn, Google Scholar, or just disable all of them.

news

Jan 15, 2016 A simple inline announcement with Markdown emoji! :sparkles: :smile:
Nov 07, 2015 A long announcement with details
Oct 22, 2015 A simple inline announcement.

latest posts

selected publications

  1. Marconi: Prefix Caching for the Era of Hybrid LLMs
    Rui Pan, Zhuang Wang, Zhen Jia, Can Karakus, Luca Zancato, Tri Dao, Ravi Netravali, and Yida Wang
    In Machine Learning and Systems (MLSys), 2025
  2. FlashAttention-3: Fast and Accurate Attention with Asynchrony and Low-precision
    Jay Shah*, Ganesh Bikshandi*, Ying Zhang, Vijay Thakkar, Pradeep Ramani, and Tri Dao
    In Advances in Neural Information Processing Systems (NeurIPS), 2024
  3. Transformers are SSMs: Generalized Models and Efficient Algorithms Through Structured State Space Duality
    Tri Dao* and Albert Gu*
    In International Conference on Machine Learning (ICML), 2024
  4. Mamba: Linear-Time Sequence Modeling with Selective State Spaces
    Albert Gu* and Tri Dao*
    Conference on Language Modeling (COLM), 2023
  5. FlashAttention: Fast and Memory-Efficient Exact Attention with IO-Awareness
    Tri Dao, Daniel Y. Fu, Stefano Ermon, Atri Rudra, and Christopher Ré
    In Advances in Neural Information Processing Systems, 2022
  6. Monarch: Expressive Structured Matrices for Efficient and Accurate Training
    Tri Dao, Beidi Chen, Nimit Sohoni, Arjun Desai, Michael Poli, Jessica Grogan, Alexander Liu, Aniruddh Rao, Atri Rudra, and Christopher Ré
    In International Conference on Machine Learning (ICML), 2022