victorsungo / WizardLMLinks
Family of instruction-following LLMs powered by Evol-Instruct: WizardLM, WizardCoder
☆44Updated last year
Alternatives and similar repositories for WizardLM
Users that are interested in WizardLM are comparing it to the libraries listed below
Sorting:
- FuseAI Project☆87Updated last year
- Data preparation code for CrystalCoder 7B LLM☆45Updated last year
- Langchain implementation of HuggingGPT☆134Updated 2 years ago
- ☆96Updated last year
- Mixture-of-Experts (MoE) Language Model☆194Updated last year
- An Implementation of "Orca: Progressive Learning from Complex Explanation Traces of GPT-4"☆43Updated last year
- My implementation of "Algorithm of Thoughts: Enhancing Exploration of Ideas in Large Language Models"☆99Updated 2 years ago
- XVERSE-MoE-A36B: A multilingual large language model developed by XVERSE Technology Inc.☆38Updated last year
- The data processing pipeline for the Koala chatbot language model☆118Updated 2 years ago
- Pre-training code for CrystalCoder 7B LLM☆57Updated last year
- Code and data for CoachLM, an automatic instruction revision approach LLM instruction tuning.☆60Updated last year
- ☆278Updated 2 years ago
- Open Implementations of LLM Analyses☆107Updated last year
- a Fine-tuned LLaMA that is Good at Arithmetic Tasks☆178Updated 2 years ago
- ☆83Updated last year
- [ICLR 2025] A trinity of environments, tools, and benchmarks for general virtual agents☆221Updated 7 months ago
- [ACL 2024] LLM2LLM: Boosting LLMs with Novel Iterative Data Enhancement☆193Updated last year
- The Next Generation Multi-Modality Superintelligence☆70Updated last year
- LongQLoRA: Extent Context Length of LLMs Efficiently☆168Updated 2 years ago
- Implementation of Toolformer: Language Models Can Teach Themselves to Use Tools☆144Updated 2 years ago
- ☆320Updated last year
- ☆123Updated last year
- Code for paper titled "Towards the Law of Capacity Gap in Distilling Language Models"☆102Updated last year
- Data preparation code for Amber 7B LLM☆94Updated last year
- XVERSE-65B: A multilingual large language model developed by XVERSE Technology Inc.☆141Updated last year
- Official repo of Respond-and-Respond: data, code, and evaluation☆103Updated last year
- Fast LLM Training CodeBase With dynamic strategy choosing [Deepspeed+Megatron+FlashAttention+CudaFusionKernel+Compiler];☆40Updated 2 years ago
- Flacuna was developed by fine-tuning Vicuna on Flan-mini, a comprehensive instruction collection encompassing various tasks. Vicuna is al…☆111Updated 2 years ago
- ☆35Updated 2 years ago
- code for Scaling Laws of RoPE-based Extrapolation☆73Updated 2 years ago