SakanaAI / CycleQDLinks
CycleQD is a framework for parameter space model merging.
☆48Updated 11 months ago
Alternatives and similar repositories for CycleQD
Users that are interested in CycleQD are comparing it to the libraries listed below
Sorting:
- Code for Discovering Preference Optimization Algorithms with and for Large Language Models☆65Updated last year
- Official implementation of "TAID: Temporally Adaptive Interpolated Distillation for Efficient Knowledge Transfer in Language Models"☆120Updated 3 months ago
- ☆16Updated last year
- ☆22Updated 2 years ago
- List of papers on Self-Correction of LLMs.☆80Updated last year
- Mamba training library developed by kotoba technologies☆69Updated last year
- Ongoing Research Project for continaual pre-training LLM(dense mode)☆44Updated 10 months ago
- ☆20Updated last year
- Lottery Ticket Adaptation☆39Updated last year
- Checkpointable dataset utilities for foundation model training☆32Updated last year
- Code for the "Cultural evolution in populations of Large Language Models" paper☆34Updated last year
- [ICLR 2025] SDTT: a simple and effective distillation method for discrete diffusion models☆46Updated 4 months ago
- A repository for research on medium sized language models.☆77Updated last year
- Plug in & Play Pytorch Implementation of the paper: "Evolutionary Optimization of Model Merging Recipes" by Sakana AI☆31Updated last year
- Memory Mosaics are networks of associative memories working in concert to achieve a prediction task.☆57Updated 11 months ago
- The official repository of ALE-Bench☆152Updated 3 weeks ago
- Train, tune, and infer Bamba model☆138Updated 7 months ago
- Anchored Preference Optimization and Contrastive Revisions: Addressing Underspecification in Alignment☆61Updated last year
- Swallowプロジェクト 事後学習済み大規模言語モデル 評価フレームワーク☆24Updated 3 months ago
- Bayes-Adaptive RL for LLM Reasoning☆43Updated 8 months ago
- Multi-Agent Verification: Scaling Test-Time Compute with Multiple Verifiers☆26Updated 10 months ago
- A testbed for agents and environments that can automatically improve models through data generation.☆28Updated 10 months ago
- ☆62Updated last year
- ☆12Updated 10 months ago
- Support Continual pre-training & Instruction Tuning forked from llama-recipes☆34Updated last year
- Unofficial Implementation of Evolutionary Model Merging☆41Updated last year
- An AI benchmark for creative, human-like problem solving using Sudoku variants☆156Updated last month
- Advantage Leftover Lunch Reinforcement Learning (A-LoL RL): Improving Language Models with Advantage-based Offline Policy Gradients☆26Updated last year
- Q-Probe: A Lightweight Approach to Reward Maximization for Language Models☆41Updated last year
- Code of "Regularized Best-of-N Sampling with Minimum Bayes Risk Objective for Language Model Alignment" (2025).☆14Updated 9 months ago