HackerCupAI / starter-kitsLinks
☆68Updated last year
Alternatives and similar repositories for starter-kits
Users that are interested in starter-kits are comparing it to the libraries listed below
Sorting:
- A set of scripts and notebooks on LLM finetunning and dataset creation☆116Updated last year
- A competition to get you started on the NeurIPS AI Hackercup☆29Updated last year
- A repository to unravel the language of GPUs, making their kernel conversations easy to understand☆198Updated 8 months ago
- ☆29Updated last year
- A zero-to-one guide on scaling modern transformers with n-dimensional parallelism.☆115Updated last month
- ML/DL Math and Method notes☆66Updated 2 years ago
- An introduction to LLM Sampling☆79Updated last year
- Fine-tune an LLM to perform batch inference and online serving.☆120Updated 8 months ago
- Seemless interface of using PyTOrch distributed with Jupyter notebooks☆57Updated 4 months ago
- LLM training in simple, raw C/CUDA☆15Updated last year
- NeurIPS Large Language Model Efficiency Challenge: 1 LLM + 1GPU + 1Day☆260Updated 2 years ago
- Large scale 4D parallelism pre-training for 🤗 transformers in Mixture of Experts *(still work in progress)*☆86Updated 2 years ago
- ☆125Updated last year
- A collection of lightweight interpretability scripts to understand how LLMs think☆89Updated 2 weeks ago
- Code for NeurIPS LLM Efficiency Challenge☆60Updated last year
- This code repository contains the code used for my "Optimizing Memory Usage for Training LLMs and Vision Transformers in PyTorch" blog po…☆92Updated 2 years ago
- Write a fast kernel and run it on Discord. See how you compare against the best!☆68Updated last week
- ☆22Updated last year
- A puzzle to learn about prompting☆135Updated 2 years ago
- ☆31Updated last year
- MoE training for Me and You and maybe other people☆335Updated last month
- Minimal example scripts of the Hugging Face Trainer, focused on staying under 150 lines☆196Updated last year
- Fast bare-bones BPE for modern tokenizer training☆175Updated 7 months ago
- ☆92Updated last year
- A Jax-based library for building transformers, includes implementations of GPT, Gemma, LlaMa, Mixtral, Whisper, SWin, ViT and more.☆298Updated last year
- ☆259Updated 2 months ago
- Proof-of-concept of global switching between numpy/jax/pytorch in a library.☆18Updated last year
- Simple repository for training small reasoning models☆49Updated last year
- Project 2 (Building Large Language Models) for Stanford CS324: Understanding and Developing Large Language Models (Winter 2022)☆105Updated 2 years ago
- Notebooks for fine tuning pali gemma☆117Updated 9 months ago