[abstract] This approach significantly reduces the KV cache size relative to traditional multi-head attention
[3.3] For saving the KV cache, only the intermediate
latent representations need to be stored: [latex] where r is much smaller than nh · dh [n-sub-h, d-sub-h]
[background] In traditional multi-head attention you must cache full key and value matrices of size T x (nh · dh) where T is the token length, nh is the number of attention heads, dh is the dimensionality of each individual head
sounds like a big win for memory constrained environments like local inference
octocop · 1h ago
These titles need to stop, we've seen that in fact it is not all you need.
seeknotfind · 1h ago
All you need titles stopping is all you need.
Etheryte · 36m ago
All you need is love, and for these titles to stop. (But they won't do that.)
EGreg · 52m ago
We need more than that, and all you need to stop saying that!!
tankenmate · 1h ago
The title of this paper is a reference to a previous paper titled "Attention Is All You Need"[0][1]. This seminal work described the transformer model that is the basis for almost all LLMs, and is almost certainly the most cited paper on AI even though it was only published in 2017.
Right, it's an 8 year old reference that's been made hundreds of times.
People seem to love going to the references graveyard, digging up tired and dead ones and drag them around town hoping everyone thinks they're clever.
Also this was from 3 months ago.
nihzm · 17m ago
It has definitely been overused by too many authors.
This reminds me a passage of Orwell's essay "Politics and the English Language":
> A newly−invented metaphor assists thought by evoking a visual image, while on the other hand a metaphor which is technically "dead" (e.g., iron resolution) has in effect reverted to being an ordinary word and can generally be used without loss of vividness. But in between these two classes there is a huge dump of worn−out metaphors which have lost all evocative power and are merely used because they save people the trouble of inventing phrases for themselves
tankenmate · 25m ago
By that argument you must also hate anything that mentions the term "considered harmful", or makes any form of derivative cultural reference (like just about every episode of the Simpsons). Why do you let it get to you?
netdevphoenix · 27m ago
Why is this the most cited paper in AI and not the original 1943 paper who started it all?
zaptrem · 25m ago
Transformers are what made ML infinitely scalable and caused a huge amount of progress in very few years since everyone could just go scale things. However, idk how many of those papers actually even cite the transformer paper?
tankenmate · 22m ago
Probably because of the modern "publish or perish" mantra led to an exponential growth in publications, and "newer is better" means that newer impactful papers get cited more than older impactful publications. But that thesis is probably a paper in itself (of the meta analysis navel gazing variety).
kristel100 · 1h ago
Still wrapping my head around this architecture, but the idea of reducing headcount while maintaining performance is compelling. Would love to see a benchmark against something like FlashAttention.
wiz21c · 1h ago
Not quite related, but do the mamba models gain ground ?
Very cool idea. Can't wait for converted models on HF.
kavalg · 3h ago
My (possibly wrong) TLDR: TransMLA is a method to "compress" an already trained GQA model, with the additional option to further fine tune it. Shall make inference faster.
yorwba · 3h ago
It is not a method to compress a Grouped-Query Attention model, but to expand it into an equivalent Multi-head Latent Attention model with the same key-value cache size but larger effective key/value vectors and a correspondingly larger number of trainable parameters. With additional training, you can then obtain a better model that only uses a little bit more memory.
[3.3] For saving the KV cache, only the intermediate latent representations need to be stored: [latex] where r is much smaller than nh · dh [n-sub-h, d-sub-h]
[background] In traditional multi-head attention you must cache full key and value matrices of size T x (nh · dh) where T is the token length, nh is the number of attention heads, dh is the dimensionality of each individual head
sounds like a big win for memory constrained environments like local inference
[0] https://arxiv.org/abs/1706.03762 [1] https://en.wikipedia.org/wiki/Attention_Is_All_You_Need
People seem to love going to the references graveyard, digging up tired and dead ones and drag them around town hoping everyone thinks they're clever.
Also this was from 3 months ago.
> A newly−invented metaphor assists thought by evoking a visual image, while on the other hand a metaphor which is technically "dead" (e.g., iron resolution) has in effect reverted to being an ordinary word and can generally be used without loss of vividness. But in between these two classes there is a huge dump of worn−out metaphors which have lost all evocative power and are merely used because they save people the trouble of inventing phrases for themselves
Answering my own question: https://www.reddit.com/r/MachineLearning/comments/1hpg91o/d_...