yifan_zhang_ Profile Banner
Yifan Zhang Profile
Yifan Zhang

@yifan_zhang_

Followers
256
Following
71
Statuses
42

Graduate student at IIIS, Tsinghua, visiting graduate student at UCLA.

Los Angeles, CA
Joined October 2022
Don't wanna be here? Send us removal request.
@yifan_zhang_
Yifan Zhang
1 month
1/ Introducing “Tensor Product Attention Is All You Need” (TPA) and Tensor ProducT ATTenTion Transformer (T6)! 🚀 Ever wondered if there’s a more memory-efficient way to handle long contexts in LLMs? Homepage:
Tweet media one
5
66
314
@yifan_zhang_
Yifan Zhang
26 days
RT @KMohan2006: 1/n 'Tensor Product Attention is all you need' paper Key Points -> 1. KV size reduction by using contextual tensor decom…
0
15
0
@yifan_zhang_
Yifan Zhang
27 days
RT @thomasahle: Tensor Product Attention illustrated with Tensor Diagrams.
Tweet media one
0
30
0
@yifan_zhang_
Yifan Zhang
28 days
@Grad62304977 @QuanquanGu See this Flash_TPA.pdf
0
0
5
@yifan_zhang_
Yifan Zhang
28 days
1
1
8
@yifan_zhang_
Yifan Zhang
28 days
RT @QuanquanGu: MHA-->GQA-->MLA--->TPA🚀🚀🚀 Introducing Tensor Product Attention (TPA). To reduce KV cache size, various Multi-Head Attenti…
0
55
0
@yifan_zhang_
Yifan Zhang
30 days
RT @iScienceLuvr: Tensor Product Attention Is All You Need Proposes Tensor Product Attention (TPA), a mechanism that factorizes Q, K, and…
0
88
0
@yifan_zhang_
Yifan Zhang
30 days
RT @gm8xx8: Tensor Product Attention Is All You Need Tensor Product Attention reduces memory overhead by compressing KV cache using tensor…
0
39
0
@yifan_zhang_
Yifan Zhang
30 days
@Grad62304977 @QuanquanGu @YIFENGLIU_AI @HuizhuoY Forthcoming works will discuss about it (Flash TPA).
0
0
6
@yifan_zhang_
Yifan Zhang
30 days
RT @QuanquanGu: We're the architects now. 🏗️📐.
0
9
0
@yifan_zhang_
Yifan Zhang
30 days
RT @yifan_zhang_: 1/ Introducing “Tensor Product Attention Is All You Need” (TPA) and Tensor ProducT ATTenTion Transformer (T6)! 🚀 Ever wo…
0
66
0
@yifan_zhang_
Yifan Zhang
1 month
12/ Joint work with @yifan_zhang_, @YIFENGLIU_AI, @HuizhuoY, Zhen Qin, Yang Yuan, @QuanquanGu, and Andrew Chi-Chih Yao. Incredible work by an outstanding team!
0
1
12