Infini-AI-Lab / S2FT
☆12Updated last month
Alternatives and similar repositories for S2FT:
Users that are interested in S2FT are comparing it to the libraries listed below
- Official repository of "LiNeS: Post-training Layer Scaling Prevents Forgetting and Enhances Model Merging"☆24Updated 3 months ago
- Codebase for Instruction Following without Instruction Tuning☆33Updated 4 months ago
- ☆31Updated 8 months ago
- This repo is based on https://github.com/jiaweizzhao/GaLore☆24Updated 5 months ago
- ☆71Updated 6 months ago
- A repository for research on medium sized language models.☆76Updated 8 months ago
- Code for the arXiv preprint "The Unreasonable Effectiveness of Easy Training Data"☆46Updated last year
- This is the oficial repository for "Safer-Instruct: Aligning Language Models with Automated Preference Data"☆17Updated 11 months ago
- The official code repo and data hub of top_nsigma sampling strategy for LLMs.☆20Updated last week
- ☆15Updated 6 months ago
- ☆75Updated last month
- ☆32Updated 3 weeks ago
- From GaLore to WeLore: How Low-Rank Weights Non-uniformly Emerge from Low-Rank Gradients. Ajay Jaiswal, Lu Yin, Zhenyu Zhang, Shiwei Liu,…☆42Updated 7 months ago
- DPO, but faster 🚀☆33Updated 2 months ago
- SLED: Self Logits Evolution Decoding for Improving Factuality in Large Language Model https://arxiv.org/pdf/2411.02433☆21Updated 2 months ago
- ☆26Updated last month
- Official Repository of Are Your LLMs Capable of Stable Reasoning?☆18Updated this week
- [NeurIPS-2024] 📈 Scaling Laws with Vocabulary: Larger Models Deserve Larger Vocabularies https://arxiv.org/abs/2407.13623☆76Updated 4 months ago
- Official repository for ICML 2024 paper "MoRe Fine-Tuning with 10x Fewer Parameters"☆17Updated this week
- Official implementation of the ICML 2024 paper RoSA (Robust Adaptation)☆38Updated last year
- ☆23Updated 5 months ago
- [NeurIPS 2024 Spotlight] Code and data for the paper "Finding Transformer Circuits with Edge Pruning".☆47Updated 2 months ago
- official implementation of paper "Process Reward Model with Q-value Rankings"☆48Updated 2 weeks ago
- Using FlexAttention to compute attention with different masking patterns☆40Updated 4 months ago
- ☆13Updated 2 months ago