evintunador / minGemma
a simplified version of Google's Gemma model to be used for learning
☆23Updated 8 months ago
Related projects ⓘ
Alternatives and complementary repositories for minGemma
- Video+code lecture on building nanoGPT from scratch☆64Updated 5 months ago
- Q-GaLore: Quantized GaLore with INT4 Projection and Layer-Adaptive Low-Rank Gradients.☆173Updated 4 months ago
- The simplest, fastest repository for training/finetuning medium-sized xLSTMs.☆38Updated 5 months ago
- An efficent implementation of the method proposed in "The Era of 1-bit LLMs"☆154Updated last month
- ☆118Updated 3 months ago
- Collection of autoregressive model implementation☆67Updated this week
- ☆93Updated last month
- RWKV in nanoGPT style☆177Updated 5 months ago
- Scripts to create your own moe models using mlx☆86Updated 8 months ago
- EfficientQAT: Efficient Quantization-Aware Training for Large Language Models☆224Updated last month
- GPTQLoRA: Efficient Finetuning of Quantized LLMs with GPTQ☆97Updated last year
- A pipeline for LLM knowledge distillation☆78Updated 3 months ago
- PyTorch implementation of models from the Zamba2 series.☆158Updated this week
- Fast parallel LLM inference for MLX☆149Updated 4 months ago
- ☆104Updated 8 months ago
- ☆87Updated 9 months ago
- Train your own small bitnet model☆56Updated last month
- look how they massacred my boy☆58Updated last month
- Experimenting with small language models☆47Updated 10 months ago
- ☆64Updated 5 months ago
- Spherical Merge Pytorch/HF format Language Models with minimal feature loss.☆112Updated last year
- Set of scripts to finetune LLMs☆36Updated 7 months ago
- A single repo with all scripts and utils to train / fine-tune the Mamba model with or without FIM☆50Updated 7 months ago
- ☆29Updated 5 months ago
- Small and Efficient Mathematical Reasoning LLMs☆71Updated 9 months ago
- Training small GPT-2 style models using Kolmogorov-Arnold networks.☆108Updated 5 months ago
- PB-LLM: Partially Binarized Large Language Models☆148Updated last year
- An easy-to-understand framework for LLM samplers that rewind and revise generated tokens☆113Updated 3 weeks ago
- Fully fine-tune large models like Mistral, Llama-2-13B, or Qwen-14B completely for free☆221Updated 3 weeks ago
- PyTorch implementation of Infini-Transformer from "Leave No Context Behind: Efficient Infinite Context Transformers with Infini-attention…☆280Updated 6 months ago