evintunador / templateGPT
customizable template GPT code designed for easy novel architecture experimentation
☆23Updated this week
Related projects: ⓘ
- ☆109Updated last month
- Fully fine-tune large models like Mistral, Llama-2-13B, or Qwen-14B completely for free☆217Updated 6 months ago
- Q-GaLore: Quantized GaLore with INT4 Projection and Layer-Adaptive Low-Rank Gradients.☆158Updated 2 months ago
- Set of scripts to finetune LLMs☆36Updated 5 months ago
- An Open Source Toolkit For LLM Distillation☆284Updated last month
- A comprehensive repository of reasoning tasks for LLMs (and beyond)☆260Updated last month
- ☆75Updated 3 weeks ago
- Toolkit for attaching, training, saving and loading of new heads for transformer models☆237Updated last week
- Just a bunch of benchmark logs for different LLMs☆112Updated last month
- An efficent implementation of the method proposed in "The Era of 1-bit LLMs"☆155Updated 2 months ago
- Manage scalable open LLM inference endpoints in Slurm clusters☆217Updated 2 months ago
- 1.58-bit LLaMa model☆77Updated 5 months ago
- Comprehensive analysis of difference in performance of QLora, Lora, and Full Finetunes.☆81Updated last year
- Low-Rank adapter extraction for fine-tuned transformers model☆154Updated 4 months ago
- Generate Synthetic Data Using OpenAI, MistralAI or AnthropicAI☆223Updated 4 months ago
- ☆92Updated last year
- Training small GPT-2 style models using Kolmogorov-Arnold networks.☆105Updated 3 months ago
- GPT-2 (124M) quality in 5B tokens☆227Updated last week
- ☆48Updated 6 months ago
- ☆85Updated 7 months ago
- ☆82Updated 3 weeks ago
- Modeling code for a BitNet b1.58 Llama-style model.☆22Updated 4 months ago
- A compact LLM pretrained in 9 days by using high quality data☆225Updated 3 weeks ago
- Embed arbitrary modalities (images, audio, documents, etc) into large language models.☆170Updated 5 months ago
- Video+code lecture on building nanoGPT from scratch☆64Updated 3 months ago
- ☆89Updated 11 months ago
- This is our own implementation of 'Layer Selective Rank Reduction'☆229Updated 3 months ago
- RAFT, or Retrieval-Augmented Fine-Tuning, is a method comprising of a fine-tuning and a RAG-based retrieval phase. It is particularly sui…☆60Updated 3 weeks ago
- Micro Llama is a small Llama based model with 300M parameters trained from scratch with $500 budget☆115Updated 5 months ago
- One click templates for inferencing Language Models☆97Updated last week