unslothai / unsloth-zoo
Utils for Unsloth
☆73Updated last week
Alternatives and similar repositories for unsloth-zoo:
Users that are interested in unsloth-zoo are comparing it to the libraries listed below
- ☆113Updated 2 weeks ago
- Train, tune, and infer Bamba model☆88Updated this week
- ☆53Updated 10 months ago
- A toolkit for fine-tuning, inferencing, and evaluating GreenBitAI's LLMs.☆82Updated last month
- Easy to use, High Performant Knowledge Distillation for LLMs☆60Updated this week
- Q-GaLore: Quantized GaLore with INT4 Projection and Layer-Adaptive Low-Rank Gradients.☆198Updated 9 months ago
- EvolKit is an innovative framework designed to automatically enhance the complexity of instructions used for fine-tuning Large Language M…☆213Updated 5 months ago
- An efficent implementation of the method proposed in "The Era of 1-bit LLMs"☆154Updated 6 months ago
- Set of scripts to finetune LLMs☆37Updated last year
- Collection of autoregressive model implementation☆85Updated this week
- A pipeline for LLM knowledge distillation☆100Updated 3 weeks ago
- Unsloth Studio☆79Updated 3 weeks ago
- PyTorch implementation of models from the Zamba2 series.☆179Updated 3 months ago
- ☆129Updated 8 months ago
- Fine-tunes a student LLM using teacher feedback for improved reasoning and answer quality. Implements GRPO with teacher-provided evaluati…☆41Updated last month
- ☆48Updated 5 months ago
- Train your own SOTA deductive reasoning model☆88Updated last month
- Google TPU optimizations for transformers models☆108Updated 3 months ago
- ☆117Updated 8 months ago
- Repo hosting codes and materials related to speeding LLMs' inference using token merging.☆36Updated 11 months ago
- Block Transformer: Global-to-Local Language Modeling for Fast Inference (NeurIPS 2024)☆152Updated 2 weeks ago
- ☆75Updated last year
- Data preparation code for Amber 7B LLM☆88Updated 11 months ago
- My fork os allen AI's OLMo for educational purposes.☆30Updated 4 months ago
- LLM-Training-API: Including Embeddings & ReRankers, mergekit, LaserRMT☆27Updated last year
- Training-free Post-training Efficient Sub-quadratic Complexity Attention. Implemented with OpenAI Triton.☆127Updated last week
- Lightweight demos for finetuning LLMs. Powered by 🤗 transformers and open-source datasets.☆76Updated 6 months ago
- A high-throughput and memory-efficient inference and serving engine for LLMs☆262Updated 6 months ago
- Simple GRPO scripts and configurations.☆58Updated 2 months ago
- minimal GRPO implementation from scratch☆85Updated last month