juyongjiang / CodeUp
CodeUp: A Multilingual Code Generation Llama2 Model with Parameter-Efficient Instruction-Tuning on a Single RTX 3090
☆117Updated last year
Related projects ⓘ
Alternatives and complementary repositories for CodeUp
- Meta-CoT: Generalizable Chain-of-Thought Prompting in Mixed-task Scenarios with Large Language Models☆87Updated last year
- Open-source Self-Instruction Tuning Code LLM☆168Updated last year
- ☆72Updated last year
- Official repo of Respond-and-Respond: data, code, and evaluation☆98Updated 3 months ago
- ☆86Updated last year
- Client Code Examples, Use Cases and Benchmarks for Enterprise h2oGPTe RAG-Based GenAI Platform☆81Updated this week
- QLoRA: Efficient Finetuning of Quantized LLMs☆77Updated 7 months ago
- Open Source WizardCoder Dataset☆153Updated last year
- Merge Transformers language models by use of gradient parameters.☆201Updated 3 months ago
- Load multiple LoRA modules simultaneously and automatically switch the appropriate combination of LoRA modules to generate the best answe…☆144Updated 9 months ago
- Spherical Merge Pytorch/HF format Language Models with minimal feature loss.☆112Updated last year
- My implementation of "Algorithm of Thoughts: Enhancing Exploration of Ideas in Large Language Models"☆92Updated last year
- MiniCheck: Efficient Fact-Checking of LLMs on Grounding Documents [EMNLP 2024]☆104Updated last month
- Convenient wrapper for fine-tuning and inference of Large Language Models (LLMs) with several quantization techniques (GTPQ, bitsandbytes…☆145Updated last year
- ☆83Updated last year
- evol augment any dataset online☆55Updated last year
- Small and Efficient Mathematical Reasoning LLMs☆71Updated 9 months ago
- ☆171Updated last year
- Instruct-tune Open LLaMA / RedPajama / StableLM models on consumer hardware using QLoRA☆80Updated 11 months ago
- This is work done by the Oxen.ai Community, trying to reproduce the Self-Rewarding Language Model paper from MetaAI.☆110Updated last week
- ☆263Updated last year
- Notus is a collection of fine-tuned LLMs using SFT, DPO, SFT+DPO, and/or any other RLHF techniques, while always keeping a data-first app…☆161Updated 10 months ago
- ☆103Updated 3 months ago
- ☆37Updated 11 months ago
- Code repo for "Agent Instructs Large Language Models to be General Zero-Shot Reasoners"☆82Updated 2 months ago
- Run evaluation on LLMs using human-eval benchmark☆379Updated last year
- Camel-Coder: Collaborative task completion with multiple agents. Role-based prompts, intervention mechanism, and thoughtful suggestions☆33Updated last year
- ☆33Updated 6 months ago
- ☆152Updated 2 months ago
- Benchmarking LLMs with Challenging Tasks from Real Users☆198Updated 2 weeks ago