tomaarsen / attention_sinksView on GitHub
Extend existing LLMs way beyond the original training length with constant memory usage, without retraining
736Apr 10, 2024Updated last year

Alternatives and similar repositories for attention_sinks

Users that are interested in attention_sinks are comparing it to the libraries listed below

Sorting:

Are these results useful?