eric-prog / GPU-GrantsLinks
GPUGrants - a list of GPU grants that I can think of
☆49Updated 2 months ago
Alternatives and similar repositories for GPU-Grants
Users that are interested in GPU-Grants are comparing it to the libraries listed below
Sorting:
- ☆29Updated last year
- ⏰ AI conference deadline countdowns☆288Updated 2 weeks ago
- $100K or 100 Days: Trade-offs when Pre-Training with Academic Resources☆147Updated last month
- The official implementation of the paper "What Matters in Transformers? Not All Attention is Needed".☆181Updated 2 weeks ago
- ☆45Updated 6 months ago
- An extension of the nanoGPT repository for training small MOE models.☆215Updated 8 months ago
- LoRA and DoRA from Scratch Implementations☆215Updated last year
- Code to reproduce "Transformers Can Do Arithmetic with the Right Embeddings", McLeish et al (NeurIPS 2024)☆194Updated last year
- The simplest, fastest repository for training/finetuning medium-sized GPTs.☆173Updated 5 months ago
- Open source interpretability artefacts for R1.☆163Updated 7 months ago
- A curated reading list of research in Adaptive Computation, Inference-Time Computation & Mixture of Experts (MoE).☆160Updated 10 months ago
- Conference schedule, top papers, and analysis of the data for NeurIPS 2023!☆121Updated last year
- Research projects built on top of Transformers☆101Updated 8 months ago
- List of AI Internships☆129Updated 2 years ago
- LLM-Merging: Building LLMs Efficiently through Merging☆205Updated last year
- ☆225Updated last month
- Survey: A collection of AWESOME papers and resources on the latest research in Mixture of Experts.☆139Updated last year
- Landing repository for the paper "Softpick: No Attention Sink, No Massive Activations with Rectified Softmax"☆85Updated 2 months ago
- Implementation of the Llama architecture with RLHF + Q-learning☆168Updated 9 months ago
- FlexAttention based, minimal vllm-style inference engine for fast Gemma 2 inference.☆305Updated 3 weeks ago
- Prune transformer layers☆74Updated last year
- Tutorials for Triton, a language for writing gpu kernels☆57Updated 2 years ago
- Pytorch implementation of the PEER block from the paper, Mixture of A Million Experts, by Xu Owen He at Deepmind☆131Updated 3 weeks ago
- Code for "LayerSkip: Enabling Early Exit Inference and Self-Speculative Decoding", ACL 2024☆347Updated 6 months ago
- nanoGPT-like codebase for LLM training☆110Updated 3 weeks ago
- OpenCoconut implements a latent reasoning paradigm where we generate thoughts before decoding.☆173Updated 10 months ago
- Understand and test language model architectures on synthetic tasks.☆240Updated 2 months ago
- Minimal (400 LOC) implementation Maximum (multi-node, FSDP) GPT training☆132Updated last year
- This repository collects all relevant resources about interpretability in LLMs☆384Updated last year
- Notes on Direct Preference Optimization☆23Updated last year