suneeta-mall / deep_learning_at_scaleLinks
Contains hands-on example code for [O'reilly book "Deep Learning At Scale"](https://www.oreilly.com/library/view/deep-learning-at/9781098145279/).
☆29Updated last year
Alternatives and similar repositories for deep_learning_at_scale
Users that are interested in deep_learning_at_scale are comparing it to the libraries listed below
Sorting:
- ☆584Updated this week
- Slides, notes, and materials for the workshop☆334Updated last year
- A repository to unravel the language of GPUs, making their kernel conversations easy to understand☆196Updated 5 months ago
- Fine-tune an LLM to perform batch inference and online serving.☆113Updated 6 months ago
- A set of scripts and notebooks on LLM finetunning and dataset creation☆111Updated last year
- Accelerate Model Training with PyTorch 2.X, published by Packt☆48Updated 2 weeks ago
- Minimal example scripts of the Hugging Face Trainer, focused on staying under 150 lines☆196Updated last year
- Complete implementation of Llama2 with/without KV cache & inference 🚀☆48Updated last year
- ☆219Updated 10 months ago
- ☆225Updated last month
- ☆177Updated last year
- Notes on quantization in neural networks☆109Updated last year
- Obsolete version of CUDA-mode repo -- use cuda-mode/lectures instead☆26Updated last year
- Where GPUs get cooked 👩🍳🔥☆319Updated 2 months ago
- GPU Kernels☆209Updated 7 months ago
- A curated list of resources for learning and exploring Triton, OpenAI's programming language for writing efficient GPU code.☆437Updated 8 months ago
- Recreating PyTorch from scratch (C/C++, CUDA, NCCL and Python, with multi-GPU support and automatic differentiation!)☆159Updated this week
- LORA: Low-Rank Adaptation of Large Language Models implemented using PyTorch☆117Updated 2 years ago
- An extension of the nanoGPT repository for training small MOE models.☆215Updated 8 months ago
- FlexAttention based, minimal vllm-style inference engine for fast Gemma 2 inference.☆305Updated 3 weeks ago
- 100 days of building GPU kernels!☆540Updated 7 months ago
- A collection of LogitsProcessors to customize and enhance LLM behavior for specific tasks.☆374Updated 4 months ago
- Starter pack for NeurIPS LLM Efficiency Challenge 2023.☆128Updated 2 years ago
- ☆45Updated 6 months ago
- Distributed training (multi-node) of a Transformer model☆87Updated last year
- RAGs: Simple implementations of Retrieval Augmented Generation (RAG) Systems☆140Updated 10 months ago
- ML/DL Math and Method notes☆64Updated last year
- My own repository containing the codes I wrote to practice CUDA programming.☆63Updated 2 years ago
- Best practices & guides on how to write distributed pytorch training code☆543Updated last month
- Simple MPI implementation for prototyping or learning☆289Updated 3 months ago