ChaseLab-PKU / InstAttentionView on GitHub
InstAttention: In-Storage Attention Offloading for Cost-Effective Long-Context LLM Inference
16Mar 30, 2025Updated last year

Alternatives and similar repositories for InstAttention

Users that are interested in InstAttention are comparing it to the libraries listed below. We may earn a commission when you buy through links labeled 'Ad' on this page.

Sorting:

Are these results useful?