formll / resolving-scaling-law-discrepancies
☆18Updated 8 months ago
Alternatives and similar repositories for resolving-scaling-law-discrepancies:
Users that are interested in resolving-scaling-law-discrepancies are comparing it to the libraries listed below
- ☆30Updated 2 months ago
- The repository contains code for Adaptive Data Optimization☆20Updated 3 months ago
- Efficient Scaling laws and collaborative pretraining.☆15Updated last month
- ☆12Updated last year
- Official repository of "LiNeS: Post-training Layer Scaling Prevents Forgetting and Enhances Model Merging"☆24Updated 4 months ago
- Offcial Repo of Paper "Eliminating Position Bias of Language Models: A Mechanistic Approach""☆11Updated 7 months ago
- ☆30Updated last year
- Is In-Context Learning Sufficient for Instruction Following in LLMs? [ICLR 2025]☆29Updated 2 months ago
- Official code for the paper "Attention as a Hypernetwork"☆25Updated 9 months ago
- Official code repo for paper "Great Memory, Shallow Reasoning: Limits of kNN-LMs"☆22Updated 6 months ago
- Official repository of paper "RNNs Are Not Transformers (Yet): The Key Bottleneck on In-context Retrieval"☆26Updated 11 months ago
- Codebase for Instruction Following without Instruction Tuning☆33Updated 6 months ago
- ☆18Updated 4 months ago
- Self-Supervised Alignment with Mutual Information☆16Updated 9 months ago
- Code for ICLR 2025 Paper "What is Wrong with Perplexity for Long-context Language Modeling?"☆44Updated last month
- Code for paper: "LASeR: Learning to Adaptively Select Reward Models with Multi-Arm Bandits"☆13Updated 5 months ago
- [ICML 2024] Junk DNA Hypothesis: A Task-Centric Angle of LLM Pre-trained Weights through Sparsity; Lu Yin*, Ajay Jaiswal*, Shiwei Liu, So…☆16Updated 9 months ago
- This is the oficial repository for "Safer-Instruct: Aligning Language Models with Automated Preference Data"☆17Updated last year
- Code for the paper "Data Feedback Loops: Model-driven Amplification of Dataset Biases"☆15Updated 2 years ago
- ☆12Updated 3 months ago
- ☆15Updated 8 months ago
- ☆27Updated last year
- Stick-breaking attention☆48Updated last week
- ☆11Updated 9 months ago
- This is an official implementation of the Reward rAnked Fine-Tuning Algorithm (RAFT), also known as iterative best-of-n fine-tuning or re…☆26Updated 6 months ago
- ☆26Updated 3 weeks ago