loubnabnl / santacoder-finetuning
Fine-tune SantaCoder for Code/Text Generation.
☆191Updated 2 years ago
Alternatives and similar repositories for santacoder-finetuning:
Users that are interested in santacoder-finetuning are comparing it to the libraries listed below
- 🐙 OctoPack: Instruction Tuning Code Large Language Models☆462Updated 2 months ago
- CodeGen2 models for program synthesis☆274Updated last year
- Run evaluation on LLMs using human-eval benchmark☆407Updated last year
- ☆269Updated 2 years ago
- ✨ RepoBench: Benchmarking Repository-Level Code Auto-Completion Systems - ICLR 2024☆156Updated 8 months ago
- Open Source WizardCoder Dataset☆157Updated last year
- Code for the paper "Rethinking Benchmark and Contamination for Language Models with Rephrased Samples"☆299Updated last year
- [NeurIPS'24] SelfCodeAlign: Self-Alignment for Code Generation☆304Updated 2 months ago
- Generate textbook-quality synthetic LLM pretraining data☆498Updated last year
- Open Instruction Generalist is an assistant trained on massive synthetic instructions to perform many millions of tasks☆208Updated last year
- Repository for analysis and experiments in the BigCode project.☆118Updated last year
- evol augment any dataset online☆59Updated last year
- A bagel, with everything.☆319Updated last year
- Data and code for "DocPrompting: Generating Code by Retrieving the Docs" @ICLR 2023☆243Updated last year
- Official repository for LongChat and LongEval☆519Updated 11 months ago
- Manage scalable open LLM inference endpoints in Slurm clusters☆254Updated 9 months ago
- batched loras☆341Updated last year
- ☆84Updated last year
- [ICML 2023] Data and code release for the paper "DS-1000: A Natural and Reliable Benchmark for Data Science Code Generation".☆241Updated 5 months ago
- ☆172Updated last year
- ☆94Updated last year
- Merge Transformers language models by use of gradient parameters.☆206Updated 8 months ago
- CrossCodeEval: A Diverse and Multilingual Benchmark for Cross-File Code Completion (NeurIPS 2023)☆139Updated 8 months ago
- CRUXEval: Code Reasoning, Understanding, and Execution Evaluation☆136Updated 6 months ago
- A multi-programming language benchmark for LLMs☆241Updated 3 months ago
- ☆73Updated last year
- Fast & more realistic evaluation of chat language models. Includes leaderboard.☆186Updated last year
- [ICLR 2024] Lemur: Open Foundation Models for Language Agents☆545Updated last year
- Accepted by Transactions on Machine Learning Research (TMLR)☆126Updated 6 months ago
- Experiments on speculative sampling with Llama models☆125Updated last year