epfl-dlab / transformers-CFG
π€ A specialized library for integrating context-free grammars (CFG) in EBNF with the Hugging Face Transformers
β89Updated 2 weeks ago
Related projects β
Alternatives and complementary repositories for transformers-CFG
- RepoQA: Evaluating Long-Context Code Understandingβ99Updated last week
- A simple unified framework for evaluating LLMsβ138Updated this week
- Code for paper "LEVER: Learning to Verifiy Language-to-Code Generation with Execution" (ICML'23)β79Updated last year
- Positional Skip-wise Training for Efficient Context Window Extension of LLMs to Extremely Length (ICLR 2024)β200Updated 5 months ago
- r2e: turn any github repository into a programming agent environmentβ87Updated last week
- Code repository for the c-BTM paperβ105Updated last year
- The source code of our work "Prepacking: A Simple Method for Fast Prefilling and Increased Throughput in Large Language Models"β56Updated last month
- Manage scalable open LLM inference endpoints in Slurm clustersβ237Updated 3 months ago
- Archon provides a modular framework for combining different inference-time techniques and LMs with just a JSON config file.β119Updated 2 weeks ago
- A pipeline for LLM knowledge distillationβ77Updated 3 months ago
- The official repo for "LLoCo: Learning Long Contexts Offline"β110Updated 4 months ago
- REST: Retrieval-Based Speculative Decoding, NAACL 2024β174Updated last month
- Benchmarking LLMs with Challenging Tasks from Real Usersβ194Updated last week
- Official repository for "Scaling Retrieval-Based Langauge Models with a Trillion-Token Datastore".β128Updated this week
- Simple and efficient pytorch-native transformer training and inference (batched)β61Updated 7 months ago
- Efficient and general syntactical decoding for Large Language Modelsβ196Updated this week
- open-source code for paper: Retrieval Head Mechanistically Explains Long-Context Factualityβ156Updated 3 months ago
- π Repository for "AppWorld: A Controllable World of Apps and People for Benchmarking Interactive Coding Agent", ACL'24 Best Resource Papβ¦β106Updated 2 weeks ago
- Code for the paper "Rethinking Benchmark and Contamination for Language Models with Rephrased Samples"β292Updated 10 months ago
- Can Language Models Solve Olympiad Programming?β100Updated 3 months ago
- Tree Attention: Topology-aware Decoding for Long-Context Attention on GPU clustersβ104Updated last month
- Multipack distributed sampler for fast padding-free training of LLMsβ175Updated 3 months ago
- RuLES: a benchmark for evaluating rule-following in language modelsβ210Updated last month
- Layer-Condensed KV cache w/ 10 times larger batch size, fewer params and less computation. Dramatic speed up with better task performanceβ¦β137Updated this week
- Evaluating LLMs with Dynamic Dataβ68Updated this week
- ModuleFormer is a MoE-based architecture that includes two different types of experts: stick-breaking attention heads and feedforward expβ¦β216Updated 7 months ago
- β105Updated 3 months ago
- LOFT: A 1 Million+ Token Long-Context Benchmarkβ142Updated last week
- A framework to study AI models in Reasoning, Alignment, and use of Memory (RAM).β142Updated this week
- (ICML 2024) Alphazero-like Tree-Search can guide large language model decoding and trainingβ215Updated 5 months ago