FSoft-AI4Code / CodeCapybara
Open-source Self-Instruction Tuning Code LLM
☆168Updated last year
Related projects ⓘ
Alternatives and complementary repositories for CodeCapybara
- [EMNLP 2023] The Vault: A Comprehensive Multilingual Dataset for Advancing Code Understanding and Generation☆84Updated 3 months ago
- ☆72Updated last year
- ☆263Updated last year
- evol augment any dataset online☆55Updated last year
- ☆86Updated last year
- Run evaluation on LLMs using human-eval benchmark☆380Updated last year
- Fine-tune SantaCoder for Code/Text Generation.☆186Updated last year
- Graph-based method for end-to-end code completion with context awareness on repository☆47Updated 2 months ago
- Small and Efficient Mathematical Reasoning LLMs☆71Updated 9 months ago
- Spherical Merge Pytorch/HF format Language Models with minimal feature loss.☆112Updated last year
- Convenient wrapper for fine-tuning and inference of Large Language Models (LLMs) with several quantization techniques (GTPQ, bitsandbytes…☆145Updated last year
- [NeurIPS'24] SelfCodeAlign: Self-Alignment for Code Generation☆270Updated 3 weeks ago
- [NeurIPS 2023 D&B] Code repository for InterCode benchmark https://arxiv.org/abs/2306.14898☆194Updated 6 months ago
- Pre-training code for CrystalCoder 7B LLM☆53Updated 6 months ago
- A hard gym for programming☆140Updated 4 months ago
- ☆39Updated 5 months ago
- Open Source WizardCoder Dataset☆153Updated last year
- Data preparation code for CrystalCoder 7B LLM☆42Updated 6 months ago
- QLoRA: Efficient Finetuning of Quantized LLMs☆77Updated 7 months ago
- CRUXEval: Code Reasoning, Understanding, and Execution Evaluation☆115Updated last month
- Open Implementations of LLM Analyses☆94Updated last month
- [EMNLP 2023] The CoT Collection: Improving Zero-shot and Few-shot Learning of Language Models via Chain-of-Thought Fine-Tuning☆214Updated last year
- RepoQA: Evaluating Long-Context Code Understanding☆100Updated 3 weeks ago
- Enhancing AI Software Engineering with Repository-level Code Graph☆96Updated 2 months ago
- ☆171Updated last year
- A set of utilities for running few-shot prompting experiments on large-language models☆113Updated last year
- ☆75Updated last year
- xCodeEval: A Large Scale Multilingual Multitask Benchmark for Code Understanding, Generation, Translation and Retrieval☆74Updated 2 months ago
- Evaluating tool-augmented LLMs in conversation settings☆72Updated 5 months ago
- This is the repo for the paper Shepherd -- A Critic for Language Model Generation☆213Updated last year