eric-prog / GPU-GrantsLinks
GPUGrants - a list of GPU grants that I can think of
☆32Updated 5 months ago
Alternatives and similar repositories for GPU-Grants
Users that are interested in GPU-Grants are comparing it to the libraries listed below
Sorting:
- ☆43Updated 3 months ago
- $100K or 100 Days: Trade-offs when Pre-Training with Academic Resources☆143Updated 3 months ago
- Collection of autoregressive model implementation☆86Updated 4 months ago
- ⏰ AI conference deadline countdowns☆280Updated last week
- ☆28Updated 10 months ago
- LoRA and DoRA from Scratch Implementations☆210Updated last year
- Implementation of the Llama architecture with RLHF + Q-learning☆166Updated 6 months ago
- Mixture-of-Transformers: A Sparse and Scalable Architecture for Multi-Modal Foundation Models. TMLR 2025.☆93Updated 3 months ago
- Minimal (400 LOC) implementation Maximum (multi-node, FSDP) GPT training☆131Updated last year
- Fast bare-bones BPE for modern tokenizer training☆164Updated 2 months ago
- ☆46Updated 5 months ago
- Notebooks for fine tuning pali gemma☆112Updated 4 months ago
- ☆124Updated 10 months ago
- An introduction to LLM Sampling☆79Updated 8 months ago
- Notebook and Scripts that showcase running quantized diffusion models on consumer GPUs☆38Updated 10 months ago
- Large scale 4D parallelism pre-training for 🤗 transformers in Mixture of Experts *(still work in progress)*☆87Updated last year
- minimal GRPO implementation from scratch☆96Updated 5 months ago
- An extension of the nanoGPT repository for training small MOE models.☆181Updated 5 months ago
- Repository for ACM India Summer School on Generative AI for Text☆13Updated last year
- Arrakis is a library to conduct, track and visualize mechanistic interpretability experiments.☆31Updated 4 months ago
- JEEBench, EMNLP 2023☆40Updated last year
- Implementation of CALM from the paper "LLM Augmented LLMs: Expanding Capabilities through Composition", out of Google Deepmind☆177Updated 11 months ago
- The simplest, fastest repository for training/finetuning medium-sized GPTs.☆153Updated 2 months ago
- OpenCoconut implements a latent reasoning paradigm where we generate thoughts before decoding.☆173Updated 7 months ago
- Code to reproduce "Transformers Can Do Arithmetic with the Right Embeddings", McLeish et al (NeurIPS 2024)☆192Updated last year
- ☆141Updated 2 weeks ago
- A basic pure pytorch implementation of flash attention☆16Updated 10 months ago
- A puzzle to learn about prompting☆132Updated 2 years ago
- Our solution for the arc challenge 2024☆174Updated 2 months ago
- Resources from the EleutherAI Math Reading Group☆54Updated 6 months ago