xlang-ai / batch-prompting
[EMNLP 2023 Industry Track] A simple prompting approach that enables the LLMs to run inference in batches.
☆72Updated 11 months ago
Alternatives and similar repositories for batch-prompting:
Users that are interested in batch-prompting are comparing it to the libraries listed below
- Implementation of the paper: "Leave No Context Behind: Efficient Infinite Context Transformers with Infini-attention" from Google in pyTO…☆53Updated last week
- Astraios: Parameter-Efficient Instruction Tuning Code Language Models☆57Updated 10 months ago
- A dataset of LLM-generated chain-of-thought steps annotated with mistake location.☆77Updated 6 months ago
- Long Context Extension and Generalization in LLMs☆48Updated 4 months ago
- Official implementation for 'Extending LLMs’ Context Window with 100 Samples'☆76Updated last year
- ☆64Updated 10 months ago
- ☆38Updated 10 months ago
- ☆82Updated 4 months ago
- Lightweight tool to identify Data Contamination in LLMs evaluation☆46Updated 11 months ago
- ☆72Updated 9 months ago
- Codebase for Instruction Following without Instruction Tuning☆33Updated 4 months ago
- Scalable Meta-Evaluation of LLMs as Evaluators☆43Updated last year
- Official code repo for paper "Great Memory, Shallow Reasoning: Limits of kNN-LMs"☆21Updated 5 months ago
- The source code of our work "Prepacking: A Simple Method for Fast Prefilling and Increased Throughput in Large Language Models"☆59Updated 4 months ago
- ☆34Updated last year
- Transformers at any scale☆41Updated last year
- PyTorch building blocks for the OLMo ecosystem☆54Updated this week
- Code and Data for "Long-context LLMs Struggle with Long In-context Learning"☆100Updated 7 months ago
- Fast and Robust Early-Exiting Framework for Autoregressive Language Models with Synchronized Parallel Decoding (EMNLP 2023 Long)☆56Updated 4 months ago
- ☆39Updated 6 months ago
- Implementation of "LM-Infinite: Simple On-the-Fly Length Generalization for Large Language Models"☆42Updated 3 months ago
- Reference implementation for Reward-Augmented Decoding: Efficient Controlled Text Generation With a Unidirectional Reward Model☆42Updated last year
- Retrieval as Attention