tomaarsen / attention_sinks
View external linksLinks

Extend existing LLMs way beyond the original training length with constant memory usage, without retraining
737Apr 10, 2024Updated last year

Alternatives and similar repositories for attention_sinks

Users that are interested in attention_sinks are comparing it to the libraries listed below

Sorting:

Are these results useful?