evintunador / minGemma
a simplified version of Google's Gemma model to be used for learning
☆24Updated 11 months ago
Alternatives and similar repositories for minGemma:
Users that are interested in minGemma are comparing it to the libraries listed below
- ☆123Updated 6 months ago
- Video+code lecture on building nanoGPT from scratch☆65Updated 8 months ago
- Q-GaLore: Quantized GaLore with INT4 Projection and Layer-Adaptive Low-Rank Gradients.☆191Updated 7 months ago
- Collection of autoregressive model implementation☆81Updated last week
- The simplest, fastest repository for training/finetuning medium-sized xLSTMs.☆39Updated 8 months ago
- An open source replication of the stawberry method that leverages Monte Carlo Search with PPO and or DPO☆28Updated last week
- A single repo with all scripts and utils to train / fine-tune the Mamba model with or without FIM☆51Updated 10 months ago
- Parameter-Efficient Sparsity Crafting From Dense to Mixture-of-Experts for Instruction Tuning on General Tasks☆31Updated 9 months ago
- inference code for mixtral-8x7b-32kseqlen☆99Updated last year
- ☆65Updated 8 months ago
- Set of scripts to finetune LLMs☆36Updated 10 months ago
- RWKV in nanoGPT style☆187Updated 8 months ago
- RWKV-7: Surpassing GPT☆79Updated 3 months ago
- Fully fine-tune large models like Mistral, Llama-2-13B, or Qwen-14B completely for free☆230Updated 3 months ago
- Fast approximate inference on a single GPU with sparsity aware offloading☆38Updated last year
- ☆53Updated 8 months ago
- Fast parallel LLM inference for MLX☆163Updated 7 months ago
- Zeus LLM Trainer is a rewrite of Stanford Alpaca aiming to be the trainer for all Large Language Models☆69Updated last year
- Block Transformer: Global-to-Local Language Modeling for Fast Inference (NeurIPS 2024)☆149Updated 2 months ago
- look how they massacred my boy☆63Updated 4 months ago
- ☆111Updated 2 months ago
- GPTQLoRA: Efficient Finetuning of Quantized LLMs with GPTQ☆99Updated last year
- ☆74Updated last year
- Code for the paper "QMoE: Practical Sub-1-Bit Compression of Trillion-Parameter Models".☆266Updated last year
- ☆84Updated last month
- Implementation of mamba with rust☆77Updated 11 months ago
- working implimention of deepseek MLA☆30Updated last month
- NanoGPT (124M) quality in 2.67B tokens☆27Updated this week
- Inference of Mamba models in pure C☆183Updated 11 months ago