The repo for In-context Autoencoder
☆166May 11, 2024Updated last year
Alternatives and similar repositories for icae
Users that are interested in icae are comparing it to the libraries listed below
Sorting:
- [EMNLP 2023] Adapting Language Models to Compress Long Contexts☆333Sep 9, 2024Updated last year
- ☆21Jan 16, 2025Updated last year
- Pytorch implementation for "Compressed Context Memory For Online Language Model Interaction" (ICLR'24)☆63Apr 18, 2024Updated last year
- ☆18Dec 2, 2024Updated last year
- The official repo for "LLoCo: Learning Long Contexts Offline"☆118Jun 15, 2024Updated last year
- [ACL 2025 Main] Repository for the paper: 500xCompressor: Generalized Prompt Compression for Large Language Models☆56Mar 9, 2026Updated last week
- Code for "In-Context Former: Lightning-fast Compressing Context for Large Language Model" (Findings of EMNLP 2024)☆21Nov 21, 2024Updated last year
- Codebase for Hyperdecoders https://arxiv.org/abs/2203.08304☆14Oct 11, 2022Updated 3 years ago
- [NAACL 2025 Findings] Code for "Perception Compressor: A Training-Free Prompt Compression Framework in Long Context Scenarios"☆24Mar 5, 2025Updated last year
- Code for "Retaining Key Information under High Compression Rates: Query-Guided Compressor for LLMs" (ACL 2024)☆18Jun 12, 2024Updated last year
- Cramming 1568 Tokens into a Single Vector and Back Again: Exploring the Limits of Embedding Space Capacity (ACL 2025, oral)☆32Jun 14, 2025Updated 9 months ago
- ☆306Jul 10, 2025Updated 8 months ago
- [ACL 2024] Long-Context Language Modeling with Parallel Encodings☆169Jun 13, 2024Updated last year
- [Neurips2024] Source code for xRAG: Extreme Context Compression for Retrieval-augmented Generation with One Token☆173Jul 4, 2024Updated last year
- Learning to Compress Prompts with Gist Tokens - https://arxiv.org/abs/2304.08467☆311Feb 14, 2025Updated last year
- LongBench v2 and LongBench (ACL 25'&24')☆1,122Jan 15, 2025Updated last year
- "FiD-ICL: A Fusion-in-Decoder Approach for Efficient In-Context Learning" (ACL 2023)☆15Jul 24, 2023Updated 2 years ago
- [EMNLP 2025] TokenSkip: Controllable Chain-of-Thought Compression in LLMs☆206Nov 30, 2025Updated 3 months ago
- Leveraging passage embeddings for efficient listwise reranking with large language models.☆50Dec 7, 2024Updated last year
- 📰 Must-read papers and blogs on LLM based Long Context Modeling 🔥☆1,938Feb 27, 2026Updated 3 weeks ago
- [ACL2025 Oral🔥]Turning Trash into Treasure: Accelerating Inference of Large Language Models with Token Recycling☆24Nov 11, 2025Updated 4 months ago
- ☆14Oct 3, 2024Updated last year
- [NAACL 2025 Main Selected Oral] Repository for the paper: Prompt Compression for Large Language Models: A Survey☆36May 18, 2025Updated 10 months ago
- ☆84Nov 10, 2025Updated 4 months ago
- Official implementation of "MMNeuron: Discovering Neuron-Level Domain-Specific Interpretation in Multimodal Large Language Model". Our co…☆25Dec 20, 2024Updated last year
- [EMNLP 2024] CompAct: Compressing Retrieved Documents Actively for Question Answering☆38Sep 20, 2024Updated last year
- Code for the EMNLP24 paper "A simple and effective L2 norm based method for KV Cache compression."☆18Dec 13, 2024Updated last year
- code for EMNLP 2024 paper: Interpreting Arithmetic Mechanism in Large Language Models through Comparative Neuron Analysis☆12Nov 17, 2024Updated last year
- ☆21Apr 16, 2024Updated last year
- [NAACL 2025] ETHIC: Evaluating Large Language Models on Long-Context Tasks with High Information Coverage☆16Sep 2, 2025Updated 6 months ago
- ☆39Oct 16, 2025Updated 5 months ago
- This repository contains the code for the paper: SirLLM: Streaming Infinite Retentive LLM☆60May 28, 2024Updated last year
- Code for paper: Long cOntext aliGnment via efficient preference Optimization☆24Oct 10, 2025Updated 5 months ago
- 📰 Must-read papers on KV Cache Compression (constantly updating 🤗).☆674Feb 24, 2026Updated 3 weeks ago
- ACL 2024 | LooGLE: Long Context Evaluation for Long-Context Language Models☆195Oct 8, 2024Updated last year
- Large Language Models Can Be Contextual Privacy Protection Learners☆15Oct 28, 2024Updated last year
- The official implementation of paper: SimLayerKV: A Simple Framework for Layer-Level KV Cache Reduction.☆51Oct 18, 2024Updated last year
- Fusion-in-Decoder☆592Oct 4, 2023Updated 2 years ago
- The Official Implementation of Ada-KV [NeurIPS 2025]☆128Nov 26, 2025Updated 3 months ago