nickrosh / evol-teacher
Open Source WizardCoder Dataset
☆156Updated last year
Alternatives and similar repositories for evol-teacher:
Users that are interested in evol-teacher are comparing it to the libraries listed below
- evol augment any dataset online☆57Updated last year
- ☆268Updated last year
- Generative Judge for Evaluating Alignment☆228Updated last year
- [EMNLP 2024] LongAlign: A Recipe for Long Context Alignment of LLMs☆241Updated 2 months ago
- Official code for "MAmmoTH2: Scaling Instructions from the Web" [NeurIPS 2024]☆135Updated 3 months ago
- Codes for the paper "∞Bench: Extending Long Context Evaluation Beyond 100K Tokens": https://arxiv.org/abs/2402.13718☆307Updated 4 months ago
- [ICML 2023] Data and code release for the paper "DS-1000: A Natural and Reliable Benchmark for Data Science Code Generation".☆233Updated 3 months ago
- ☆257Updated 6 months ago
- Implementation of paper Data Engineering for Scaling Language Models to 128K Context☆451Updated 11 months ago
- Unofficial implementation of AlpaGasus☆90Updated last year
- ACL 2024 | LooGLE: Long Context Evaluation for Long-Context Language Models☆175Updated 4 months ago
- [ACL'24 Outstanding] Data and code for L-Eval, a comprehensive long context language models evaluation benchmark☆369Updated 7 months ago
- [ACL'24] Superfiltering: Weak-to-Strong Data Filtering for Fast Instruction-Tuning☆141Updated 5 months ago
- 🐋 An unofficial implementation of Self-Alignment with Instruction Backtranslation.☆136Updated 7 months ago
- FireAct: Toward Language Agent Fine-tuning☆265Updated last year
- Accepted by Transactions on Machine Learning Research (TMLR)☆124Updated 4 months ago
- ☆305Updated 8 months ago
- ☆84Updated last year
- Codes and Data for Scaling Relationship on Learning Mathematical Reasoning with Large Language Models☆245Updated 5 months ago
- Run evaluation on LLMs using human-eval benchmark☆394Updated last year
- Reformatted Alignment☆114Updated 4 months ago
- InsTag: A Tool for Data Analysis in LLM Supervised Fine-tuning☆240Updated last year
- CRUXEval: Code Reasoning, Understanding, and Execution Evaluation☆125Updated 4 months ago
- ToolQA, a new dataset to evaluate the capabilities of LLMs in answering challenging questions with external tools. It offers two levels …☆251Updated last year
- Positional Skip-wise Training for Efficient Context Window Extension of LLMs to Extremely Length (ICLR 2024)☆204Updated 9 months ago
- ✨ RepoBench: Benchmarking Repository-Level Code Auto-Completion Systems - ICLR 2024☆144Updated 6 months ago
- Implementation of the LongRoPE: Extending LLM Context Window Beyond 2 Million Tokens Paper☆125Updated 7 months ago
- A large-scale, fine-grained, diverse preference dataset (and models).☆329Updated last year
- ☆43Updated 8 months ago
- 🐙 OctoPack: Instruction Tuning Code Large Language Models☆451Updated 2 weeks ago