sar-mo / CS2051-HonorsDiscreteMathLinks
A collection of resources for CS 2051, an undergraduate Honors Discrete Mathematics course at Georgia Tech.
☆10Updated 2 years ago
Alternatives and similar repositories for CS2051-HonorsDiscreteMath
Users that are interested in CS2051-HonorsDiscreteMath are comparing it to the libraries listed below
Sorting:
- Just large language models. Hackable, with as little abstraction as possible. Done for my own purposes, feel free to rip.☆44Updated 2 years ago
- Small scale distributed training of sequential deep learning models, built on Numpy and MPI.☆155Updated 2 years ago
- Simple Transformer in Jax☆142Updated last year
- Extract full next-token probabilities via language model APIs☆248Updated last year
- seqax = sequence modeling + JAX☆170Updated 6 months ago
- Fast bare-bones BPE for modern tokenizer training☆174Updated 7 months ago
- Puzzles for exploring transformers☆384Updated 2 years ago
- Resources from the EleutherAI Math Reading Group☆54Updated 11 months ago
- Helpers and such for working with Lambda Cloud☆51Updated 2 years ago
- FastAsk is a Python package that installs an easy to use command to your terminal to get a quick answer to a question, using either OpenA…☆53Updated last year
- Solve puzzles. Learn CUDA.☆63Updated 2 years ago
- A set of Python scripts that makes your experience on TPU better☆56Updated 4 months ago
- ☆237Updated 3 weeks ago
- Following Karpathy with GPT-2 implementation and training, writing lots of comments cause I have memory of a goldfish☆172Updated last year
- A puzzle to learn about prompting☆135Updated 2 years ago
- parallelized hyperdimensional tictactoe☆126Updated last year
- SIMD quantization kernels☆94Updated 4 months ago
- batched loras☆349Updated 2 years ago
- A single notebook for fine-tuning GPT-3.5 turbo☆31Updated last year
- Project 2 (Building Large Language Models) for Stanford CS324: Understanding and Developing Large Language Models (Winter 2022)☆105Updated 2 years ago
- A really tiny autograd engine☆99Updated 8 months ago
- Large scale 4D parallelism pre-training for 🤗 transformers in Mixture of Experts *(still work in progress)*☆86Updated 2 years ago
- Resources for skilling up in AI alignment research engineering. Covers basics of deep learning, mechanistic interpretability, and RL.☆238Updated 5 months ago
- Stream of my favorite papers and links☆44Updated 3 weeks ago
- could we make an ml stack in 100,000 lines of code?☆46Updated last year
- Memory optimized Mixture of Experts☆72Updated 6 months ago
- Inference code for LLaMA models in JAX☆120Updated last year
- Long context evaluation for large language models☆225Updated 10 months ago
- Minimalistic, extremely fast, and hackable researcher's toolbench for GPT models in 307 lines of code. Reaches <3.8 validation loss on wi…☆355Updated last year
- An implementation of the transformer architecture onto an Nvidia CUDA kernel☆202Updated 2 years ago