sar-mo / CS2051-HonorsDiscreteMathLinks
A collection of resources for CS 2051, an undergraduate Honors Discrete Mathematics course at Georgia Tech.
☆10Updated 2 years ago
Alternatives and similar repositories for CS2051-HonorsDiscreteMath
Users that are interested in CS2051-HonorsDiscreteMath are comparing it to the libraries listed below
Sorting:
- Fast bare-bones BPE for modern tokenizer training☆175Updated 7 months ago
- seqax = sequence modeling + JAX☆170Updated 6 months ago
- Simple Transformer in Jax☆142Updated last year
- Extract full next-token probabilities via language model APIs☆248Updated last year
- Small scale distributed training of sequential deep learning models, built on Numpy and MPI.☆155Updated 2 years ago
- Just large language models. Hackable, with as little abstraction as possible. Done for my own purposes, feel free to rip.☆44Updated 2 years ago
- Minimalistic, extremely fast, and hackable researcher's toolbench for GPT models in 307 lines of code. Reaches <3.8 validation loss on wi…☆355Updated last year
- A blog where I write about research papers and blog posts I read.☆12Updated last year
- Accelerate, Optimize performance with streamlined training and serving options with JAX.☆336Updated last week
- ☆291Updated last year
- batched loras☆349Updated 2 years ago
- Project 2 (Building Large Language Models) for Stanford CS324: Understanding and Developing Large Language Models (Winter 2022)☆105Updated 2 years ago
- Google TPU optimizations for transformers models☆134Updated 2 weeks ago
- A set of Python scripts that makes your experience on TPU better☆56Updated 4 months ago
- Solve puzzles. Learn CUDA.☆63Updated 2 years ago
- Puzzles for exploring transformers☆386Updated 2 years ago
- Cost aware hyperparameter tuning algorithm☆179Updated last year
- Large scale 4D parallelism pre-training for 🤗 transformers in Mixture of Experts *(still work in progress)*☆86Updated 2 years ago
- ☆562Updated last year
- A zero-to-one guide on scaling modern transformers with n-dimensional parallelism.☆115Updated last month
- The simplest, fastest repository for training/finetuning medium-sized GPTs.☆186Updated 3 weeks ago
- ☆237Updated last month
- Minimal yet performant LLM examples in pure JAX☆240Updated 3 weeks ago
- Legible, Scalable, Reproducible Foundation Models with Named Tensors and Jax☆693Updated 2 weeks ago
- A puzzle to learn about prompting☆135Updated 2 years ago
- JAX implementation of the Llama 2 model☆216Updated 2 years ago
- PyTorch/XLA integration with JetStream (https://github.com/google/JetStream) for LLM inference"☆79Updated last month
- 🧱 Modula software package☆322Updated 5 months ago
- A FlashAttention implementation for JAX with support for efficient document mask computation and context parallelism.☆158Updated 3 months ago
- Long context evaluation for large language models☆226Updated 11 months ago