krafton-ai / lexicoLinks
KV cache compression via sparse coding
☆11Updated 2 months ago
Alternatives and similar repositories for lexico
Users that are interested in lexico are comparing it to the libraries listed below
Sorting:
- [ICLR 2025] ELICIT: LLM Augmentation Via External In-context Capability☆11Updated 4 months ago
- 2D-TPE: Two-Dimensional Positional Encoding Enhances Table Understanding for Large Language Models (WWW 2025)☆10Updated 3 months ago
- ☆10Updated this week
- This is the official pytorch implementation for paper: Filter, Correlate, Compress: Training-Free Token Reduction for MLLM Acceleration☆15Updated 4 months ago
- Official implementation for "Pruning Large Language Models with Semi-Structural Adaptive Sparse Training" (AAAI 2025)☆12Updated 2 weeks ago
- ☆12Updated 5 months ago
- ☆11Updated 9 months ago
- (ICML 2025) Rethinking Chain-of-Thought from the Perspective of Self-Training☆9Updated 5 months ago
- ☆14Updated 7 months ago
- ☆10Updated this week
- [ICLR 2025] Causal Graphical Models for Vision-Language Compositional Understanding☆9Updated 3 months ago
- Mobile-Bench: An Evaluation Benchmark for LLM-based Mobile Agents☆13Updated 7 months ago
- Official Pytorch Implementation of "Cross-Attention Head Position Patterns Can Align with Human Visual Concepts in Text-to-Image Generati…☆9Updated 7 months ago
- ☆16Updated 2 months ago
- ☆21Updated 5 months ago
- Official code for the paper "HEXA-MoE: Efficient and Heterogeneous-Aware MoE Acceleration with Zero Computation Redundancy"☆13Updated 4 months ago
- Confidence Regulation Neurons in Language Models (NeurIPS 2024)☆10Updated 5 months ago
- [⭐️ WACV 2025 Oral ⭐️] PETALface: Parameter Efficient Transfer Learning for Low-resolution Face Recognition☆13Updated last month
- Code Repository for the NeurIPS 2024 Paper "Toward Efficient Inference for Mixture of Experts".☆19Updated 8 months ago
- ☆22Updated 2 weeks ago
- LLaVA-MR: Large Language-and-Vision Assistant for Video Moment Retrieval☆8Updated 7 months ago
- Official Implementation of FastKV: KV Cache Compression for Fast Long-Context Processing with Token-Selective Propagation☆21Updated 2 months ago
- ☆12Updated 6 months ago
- Adapt MLLMs to Domains via Post-Training☆9Updated 6 months ago
- ☆29Updated last week
- [WIP🚧] 2025 up-to-date list of resources on visual tokenizers (primarily for visual generation). Give it a star 🌟 if you find it useful…☆14Updated 6 months ago
- ☆12Updated 3 months ago
- ☆8Updated 5 months ago
- [ICLR 2025] SWIFT: On-the-Fly Self-Speculative Decoding for LLM Inference Acceleration☆52Updated 4 months ago
- ☆10Updated 2 weeks ago