sar-mo / CS2051-HonorsDiscreteMathLinks
A collection of resources for CS 2051, an undergraduate Honors Discrete Mathematics course at Georgia Tech.
☆10Updated 2 years ago
Alternatives and similar repositories for CS2051-HonorsDiscreteMath
Users that are interested in CS2051-HonorsDiscreteMath are comparing it to the libraries listed below
Sorting:
- Just large language models. Hackable, with as little abstraction as possible. Done for my own purposes, feel free to rip.☆44Updated 2 years ago
- A blog where I write about research papers and blog posts I read.☆12Updated last year
- A puzzle to learn about prompting☆135Updated 2 years ago
- Extract full next-token probabilities via language model APIs☆248Updated last year
- Small scale distributed training of sequential deep learning models, built on Numpy and MPI.☆154Updated 2 years ago
- Puzzles for exploring transformers☆382Updated 2 years ago
- Minimalistic, extremely fast, and hackable researcher's toolbench for GPT models in 307 lines of code. Reaches <3.8 validation loss on wi…☆355Updated last year
- Simple Transformer in Jax☆140Updated last year
- batched loras☆347Updated 2 years ago
- seqax = sequence modeling + JAX☆169Updated 5 months ago
- A zero-to-one guide on scaling modern transformers with n-dimensional parallelism.☆112Updated last week
- ☆94Updated 2 years ago
- Helpers and such for working with Lambda Cloud☆51Updated 2 years ago
- ☆461Updated last year
- Solve puzzles. Learn CUDA.☆63Updated 2 years ago
- Fast bare-bones BPE for modern tokenizer training☆174Updated 6 months ago
- Resources from the EleutherAI Math Reading Group☆54Updated 10 months ago
- Alex Krizhevsky's original code from Google Code☆198Updated 9 years ago
- Stream of my favorite papers and links☆44Updated last week
- Large scale 4D parallelism pre-training for 🤗 transformers in Mixture of Experts *(still work in progress)*☆86Updated 2 years ago
- Inference code for LLaMA models in JAX☆120Updated last year
- Memory optimized Mixture of Experts☆72Updated 5 months ago
- inference code for mixtral-8x7b-32kseqlen☆104Updated 2 years ago
- 🦾💻🌐 distributed training & serverless inference at scale on RunPod☆19Updated last year
- Following Karpathy with GPT-2 implementation and training, writing lots of comments cause I have memory of a goldfish☆172Updated last year
- An implementation of the transformer architecture onto an Nvidia CUDA kernel☆202Updated 2 years ago
- Building blocks for foundation models.☆587Updated 2 years ago
- ☆233Updated this week
- PyTorch/XLA integration with JetStream (https://github.com/google/JetStream) for LLM inference"☆79Updated 3 weeks ago
- The simplest, fastest repository for training/finetuning medium-sized GPTs.☆181Updated 6 months ago