evintunador / minGemma
a simplified version of Google's Gemma model to be used for learning
☆22Updated 6 months ago
Related projects: ⓘ
- The simplest, fastest repository for training/finetuning medium-sized xLSTMs.☆38Updated 3 months ago
- Video+code lecture on building nanoGPT from scratch☆64Updated 3 months ago
- Parameter-Efficient Sparsity Crafting From Dense to Mixture-of-Experts for Instruction Tuning on General Tasks☆31Updated 3 months ago
- 1.58 Bit LLM on Apple Silicon using MLX☆97Updated 4 months ago
- Q-GaLore: Quantized GaLore with INT4 Projection and Layer-Adaptive Low-Rank Gradients.☆158Updated 2 months ago
- Zeus LLM Trainer is a rewrite of Stanford Alpaca aiming to be the trainer for all Large Language Models☆67Updated last year
- A pipeline for LLM knowledge distillation☆68Updated last month
- ☆109Updated last month
- An efficent implementation of the method proposed in "The Era of 1-bit LLMs"☆155Updated 2 months ago
- Tune MPTs☆84Updated last year
- ☆64Updated 3 months ago
- Just a bunch of benchmark logs for different LLMs☆112Updated last month
- ☆75Updated 3 weeks ago
- Low-Rank adapter extraction for fine-tuned transformers model☆154Updated 4 months ago
- Comprehensive analysis of difference in performance of QLora, Lora, and Full Finetunes.☆81Updated last year
- Fully fine-tune large models like Mistral, Llama-2-13B, or Qwen-14B completely for free☆217Updated 6 months ago
- Some simple scripts that I use day-to-day when working with LLMs and Huggingface Hub☆154Updated 11 months ago
- Llama3.1 learns to Listen☆134Updated last week
- Fast parallel LLM inference for MLX☆118Updated 2 months ago
- ☆71Updated last year
- Spherical Merge Pytorch/HF format Language Models with minimal feature loss.☆107Updated last year
- ☆82Updated 3 weeks ago
- Parameter-Efficient Sparsity Crafting From Dense to Mixture-of-Experts for Instruction Tuning on General Tasks☆123Updated 6 months ago
- Set of scripts to finetune LLMs☆36Updated 5 months ago
- Full finetuning of large language models without large memory requirements☆94Updated 8 months ago
- GPTQLoRA: Efficient Finetuning of Quantized LLMs with GPTQ☆96Updated last year
- One click templates for inferencing Language Models☆97Updated last week
- Official Repository of The Mamba in the Llama: Distilling and Accelerating Hybrid Models☆130Updated this week
- A single repo with all scripts and utils to train / fine-tune the Mamba model with or without FIM☆46Updated 5 months ago
- A GPT with self-similar nested properties☆18Updated 6 months ago