Posts by Bowen Zeng
25 January 2026 - Optimize Sparse Attention in FFA
19 October 2025 - Long-Context Attention Benchmark
21 April 2025 - MagiAttention
Posts by Hanwen Sun
25 January 2026 - Optimize Sparse Attention in FFA
19 October 2025 - Long-Context Attention Benchmark
21 April 2025 - MagiAttention
Posts by Jerry Chen
07 February 2026 - Support Blackwell with FFA_FA4 Backend
Posts by Jin Li
04 February 2026 - Support Muon QK-Clip
25 January 2026 - Optimize Sparse Attention in FFA
21 January 2026 - Dynamic Attention Solver
21 April 2025 - MagiAttention
Posts by Kunlun Li
07 February 2026 - Support Blackwell with FFA_FA4 Backend
Posts by Qiangang Wang
07 February 2026 - Support Blackwell with FFA_FA4 Backend
21 January 2026 - Dynamic Attention Solver
19 October 2025 - Long-Context Attention Benchmark
21 April 2025 - MagiAttention
Posts by Yufeng Yang
07 February 2026 - Support Blackwell with FFA_FA4 Backend
Posts by Yujia Liu
07 February 2026 - Support Blackwell with FFA_FA4 Backend
Posts by Yunpeng Huang
15 February 2026 - How to Ensure Kernels Actually Overlapped
14 February 2026 - Distributed-Native FFA
07 February 2026 - Support Blackwell with FFA_FA4 Backend
04 February 2026 - Support Muon QK-Clip
25 January 2026 - Optimize Sparse Attention in FFA
24 January 2026 - Support Native Group Collective Based on DeepEP
21 January 2026 - Dynamic Attention Solver
22 December 2025 - Flash Attention 2 Math Derivation
17 November 2025 - Support Learnable Attention Sink
19 October 2025 - Long-Context Attention Benchmark
21 April 2025 - MagiAttention
Posts by Zewei Tao
08 February 2026 - Attention Engine for Inference
07 February 2026 - Support Blackwell with FFA_FA4 Backend
25 January 2026 - Optimize Sparse Attention in FFA
24 January 2026 - Support Native Group Collective Based on DeepEP
21 January 2026 - Dynamic Attention Solver
19 October 2025 - Long-Context Attention Benchmark
21 April 2025 - MagiAttention