WENGSYX / LMTunerLinks
LMTuner: Make the LLM Better for Everyone
☆35Updated last year
Alternatives and similar repositories for LMTuner
Users that are interested in LMTuner are comparing it to the libraries listed below
Sorting:
- Data preparation code for CrystalCoder 7B LLM☆45Updated last year
- Official implementation for 'Extending LLMs’ Context Window with 100 Samples'☆78Updated last year
- Code for preprint "Metadata Conditioning Accelerates Language Model Pre-training (MeCo)"☆39Updated last month
- The official code repo and data hub of top_nsigma sampling strategy for LLMs.☆26Updated 4 months ago
- This repository contains the joint use of CPO and SimPO method for better reference-free preference learning methods.☆53Updated 10 months ago
- ☆47Updated 2 weeks ago
- LongHeads: Multi-Head Attention is Secretly a Long Context Processor☆29Updated last year
- Source code of "Reasons to Reject? Aligning Language Models with Judgments"☆58Updated last year
- FuseAI Project☆87Updated 5 months ago
- Astraios: Parameter-Efficient Instruction Tuning Code Language Models☆58Updated last year
- [ICLR'24 spotlight] Tool-Augmented Reward Modeling☆50Updated 2 weeks ago
- ☆68Updated 2 years ago
- ☆27Updated this week
- ☆36Updated 9 months ago
- Minimal implementation of the Self-Play Fine-Tuning Converts Weak Language Models to Strong Language Models paper (ArXiv 20232401.01335)☆28Updated last year
- ☆45Updated last year
- Code for the arXiv preprint "The Unreasonable Effectiveness of Easy Training Data"☆48Updated last year
- Codebase for Instruction Following without Instruction Tuning☆34Updated 9 months ago
- Reformatted Alignment☆113Updated 9 months ago
- ☆35Updated last year
- Official repository for ACL 2025 paper "Model Extrapolation Expedites Alignment"☆73Updated last month
- Implementations of online merging optimizers proposed by Online Merging Optimizers for Boosting Rewards and Mitigating Tax in Alignment☆75Updated last year
- ☆47Updated last year
- Unofficial implementation of AlpaGasus☆91Updated last year
- DuoGuard: A Two-Player RL-Driven Framework for Multilingual LLM Guardrails☆24Updated 4 months ago
- ☆35Updated last year
- Code for RL4F: Generating Natural Language Feedback with Reinforcement Learning for Repairing Model Outputs. ACL 2023.☆63Updated 6 months ago
- Implementation of "LM-Infinite: Simple On-the-Fly Length Generalization for Large Language Models"☆41Updated 7 months ago
- A simple GPT-based evaluation tool for multi-aspect, interpretable assessment of LLMs.☆85Updated last year
- ☆32Updated 7 months ago