HackerCupAI / starter-kitsLinks
☆64Updated 7 months ago
Alternatives and similar repositories for starter-kits
Users that are interested in starter-kits are comparing it to the libraries listed below
Sorting:
- A competition to get you started on the NeurIPS AI Hackercup☆28Updated 8 months ago
- ☆24Updated 7 months ago
- A set of scripts and notebooks on LLM finetunning and dataset creation☆111Updated 8 months ago
- A repository to unravel the language of GPUs, making their kernel conversations easy to understand☆184Updated last week
- An introduction to LLM Sampling☆78Updated 5 months ago
- Starter pack for NeurIPS LLM Efficiency Challenge 2023.☆122Updated last year
- ML/DL Math and Method notes☆61Updated last year
- NeurIPS Large Language Model Efficiency Challenge: 1 LLM + 1GPU + 1Day☆254Updated last year
- Code for NeurIPS LLM Efficiency Challenge☆58Updated last year
- Simple repository for training small reasoning models☆31Updated 3 months ago
- Highly commented implementations of Transformers in PyTorch☆136Updated last year
- ☆35Updated last week
- Arrakis is a library to conduct, track and visualize mechanistic interpretability experiments.☆29Updated last month
- ☆20Updated last year
- Resources from the EleutherAI Math Reading Group☆53Updated 3 months ago
- A puzzle to learn about prompting☆127Updated 2 years ago
- Comprehensive analysis of difference in performance of QLora, Lora, and Full Finetunes.☆81Updated last year
- Code to reproduce "Transformers Can Do Arithmetic with the Right Embeddings", McLeish et al (NeurIPS 2024)☆190Updated last year
- Synthetic data generation and benchmark implementation for "Episodic Memories Generation and Evaluation Benchmark for Large Language Mode…☆45Updated last month
- ☆190Updated 3 months ago
- ☆123Updated 7 months ago
- Write a fast kernel and run it on Discord. See how you compare against the best!☆44Updated this week
- ☆78Updated 11 months ago
- Large scale 4D parallelism pre-training for 🤗 transformers in Mixture of Experts *(still work in progress)*☆82Updated last year
- an open source reproduction of NVIDIA's nGPT (Normalized Transformer with Representation Learning on the Hypersphere)☆100Updated 2 months ago
- Mixture-of-Transformers: A Sparse and Scalable Architecture for Multi-Modal Foundation Models. TMLR 2025.☆61Updated 3 weeks ago
- Compiling useful links, papers, benchmarks, ideas, etc.☆46Updated 2 months ago
- $100K or 100 Days: Trade-offs when Pre-Training with Academic Resources☆139Updated 2 weeks ago
- LLM-Merging: Building LLMs Efficiently through Merging☆197Updated 8 months ago
- Collection of autoregressive model implementation☆85Updated last month