tomaarsen / attention_sinks

Extend existing LLMs way beyond the original training length with constant memory usage, without retraining
687Updated 10 months ago

Alternatives and similar repositories for attention_sinks:

Users that are interested in attention_sinks are comparing it to the libraries listed below