imagination-research / sot
[ICLR 2024] Skeleton-of-Thought: Prompting LLMs for Efficient Parallel Generation
☆150Updated 10 months ago
Alternatives and similar repositories for sot:
Users that are interested in sot are comparing it to the libraries listed below
- The official repo for "LLoCo: Learning Long Contexts Offline"☆114Updated 7 months ago
- Parameter-Efficient Sparsity Crafting From Dense to Mixture-of-Experts for Instruction Tuning on General Tasks☆139Updated 4 months ago
- ☆192Updated last month
- [ICLR2025] Breaking Throughput-Latency Trade-off for Long Sequences with Speculative Decoding☆105Updated last month
- ☆125Updated last year
- Unofficial implementation for the paper "Mixture-of-Depths: Dynamically allocating compute in transformer-based language models"☆145Updated 7 months ago
- [ICLR'25] Fast Inference of MoE Models with CPU-GPU Orchestration☆180Updated 2 months ago
- Experiments on speculative sampling with Llama models☆123Updated last year
- A framework to study AI models in Reasoning, Alignment, and use of Memory (RAM).☆157Updated 3 weeks ago
- Q-GaLore: Quantized GaLore with INT4 Projection and Layer-Adaptive Low-Rank Gradients.☆190Updated 6 months ago
- Explorations into some recent techniques surrounding speculative decoding☆233Updated last month
- The code for the paper ROUTERBENCH: A Benchmark for Multi-LLM Routing System☆101Updated 7 months ago
- Positional Skip-wise Training for Efficient Context Window Extension of LLMs to Extremely Length (ICLR 2024)☆204Updated 8 months ago
- ☆120Updated 7 months ago
- scalable and robust tree-based speculative decoding algorithm☆331Updated this week
- ☆214Updated 7 months ago
- GEAR: An Efficient KV Cache Compression Recipefor Near-Lossless Generative Inference of LLM☆152Updated 6 months ago
- Code for the paper "Rethinking Benchmark and Contamination for Language Models with Rephrased Samples"☆296Updated last year
- REST: Retrieval-Based Speculative Decoding, NAACL 2024☆190Updated last month
- [NeurIPS 2024] KVQuant: Towards 10 Million Context Length LLM Inference with KV Cache Quantization☆327Updated 5 months ago
- ☆80Updated 3 months ago
- Layer-Condensed KV cache w/ 10 times larger batch size, fewer params and less computation. Dramatic speed up with better task performance…☆147Updated last week
- Repo for Rho-1: Token-level Data Selection & Selective Pretraining of LLMs.☆388Updated 9 months ago
- ☆59Updated 9 months ago
- Code for the paper "QMoE: Practical Sub-1-Bit Compression of Trillion-Parameter Models".☆263Updated last year
- Repo hosting codes and materials related to speeding LLMs' inference using token merging.☆34Updated 9 months ago
- FuseAI Project☆80Updated this week
- Benchmark baseline for retrieval qa applications☆96Updated 9 months ago
- Open Implementations of LLM Analyses☆98Updated 3 months ago
- ☆72Updated 2 weeks ago