JerryYin777 / Cross-Layer-Attention
View external linksLinks

Self Reproduction Code of Paper "Reducing Transformer Key-Value Cache Size with Cross-Layer Attention (MIT CSAIL)
17May 24, 2024Updated last year

Alternatives and similar repositories for Cross-Layer-Attention

Users that are interested in Cross-Layer-Attention are comparing it to the libraries listed below

Sorting:

Are these results useful?