Storing long contexts in tiny caches with self-study
☆243Dec 5, 2025Updated 2 months ago
Alternatives and similar repositories for cartridges
Users that are interested in cartridges are comparing it to the libraries listed below
Sorting:
- ☆468Nov 25, 2025Updated 3 months ago
- A method for evaluating the high-level coherence of machine-generated texts. Identifies high-level coherence issues in transformer-based …☆11Mar 18, 2023Updated 2 years ago
- Official repo for BWLer: Barycentric Weight Layer☆29Sep 26, 2025Updated 5 months ago
- LLMProc: Unix-inspired runtime that treats LLMs as processes.☆34Jul 17, 2025Updated 7 months ago
- Weird autoencoder experiments☆24Jan 26, 2026Updated last month
- Efficient Long-context Language Model Training by Core Attention Disaggregation☆91Updated this week
- LLM training in simple, raw C/CUDA☆15Dec 5, 2024Updated last year
- ☆14Nov 20, 2022Updated 3 years ago
- Test-Time Memory Framework: Control Hallucinations in Foundation Models☆11Nov 4, 2025Updated 3 months ago
- Official Project Page for HLA: Higher-order Linear Attention (https://arxiv.org/abs/2510.27258)☆45Jan 6, 2026Updated last month
- Async RL Training at Scale☆1,096Updated this week
- ☆27Jun 12, 2023Updated 2 years ago
- Simple GRPO scripts and configurations.☆59Feb 6, 2025Updated last year
- Official code repo for paper "Great Memory, Shallow Reasoning: Limits of kNN-LMs"☆23Apr 30, 2025Updated 10 months ago
- My submission for the GPUMODE/AMD fp8 mm challenge☆29Jun 4, 2025Updated 8 months ago
- Code for the EMNLP24 paper "A simple and effective L2 norm based method for KV Cache compression."☆18Dec 13, 2024Updated last year
- Research work aimed at addressing the problem of modeling infinite-length context☆46Dec 18, 2025Updated 2 months ago
- The evaluation framework for training-free sparse attention in LLMs☆121Jan 27, 2026Updated last month
- ☆67Mar 21, 2025Updated 11 months ago
- A framework for optimizing DSPy programs with RL☆318Jan 12, 2026Updated last month
- Approximating the joint distribution of language models via MCTS☆22Nov 3, 2024Updated last year
- Xmixers: A collection of SOTA efficient token/channel mixers☆28Sep 4, 2025Updated 5 months ago
- AdaSplash: Adaptive Sparse Flash Attention (aka Flash Entmax Attention)☆33Sep 30, 2025Updated 5 months ago
- [NeurIPS 2025 Spotlight] Reasoning Environments for Reinforcement Learning with Verifiable Rewards☆1,352Jan 16, 2026Updated last month
- [ACL 2025] Squeezed Attention: Accelerating Long Prompt LLM Inference☆57Nov 20, 2024Updated last year
- Checkpoint-engine is a simple middleware to update model weights in LLM inference engines☆912Updated this week
- The repository for the code of the UltraFastBERT paper☆519Mar 24, 2024Updated last year
- Official repo for Learning to Reason for Long-Form Story Generation☆74Apr 19, 2025Updated 10 months ago
- Our library for RL environments + evals☆3,869Updated this week
- Repository for "I am a Strange Dataset: Metalinguistic Tests for Language Models"☆45Jan 11, 2024Updated 2 years ago
- seqax = sequence modeling + JAX☆171Jul 23, 2025Updated 7 months ago
- A tree-based prefix cache library that allows rapid creation of looms: hierarchal branching pathways of LLM generations.☆78Feb 11, 2025Updated last year
- Code for UzLiB (Uzbek Linguistic Benchmark) for LLMs☆19Feb 20, 2026Updated last week
- ☆23Jul 11, 2025Updated 7 months ago
- The official implementation of the paper "Self-Updatable Large Language Models by Integrating Context into Model Parameters"☆15May 18, 2025Updated 9 months ago
- ☆13Jan 7, 2025Updated last year
- ☆14Jun 24, 2024Updated last year
- ☆11Oct 11, 2023Updated 2 years ago
- 🚀 LLM inference optimization simulator, modeling compute-bound prefill and memory-bound decode phases.☆13Jul 12, 2025Updated 7 months ago