francoisfleuret / picogptLinks
Minimal GPT (~350 lines with a simple task to test it)
☆63Updated 11 months ago
Alternatives and similar repositories for picogpt
Users that are interested in picogpt are comparing it to the libraries listed below
Sorting:
- ☆150Updated last year
- The boundary of neural network trainability is fractal☆220Updated last year
- Getting crystal-like representations with harmonic loss☆192Updated 7 months ago
- ☆60Updated 3 years ago
- Jax like function transformation engine but micro, microjax☆33Updated last year
- ☆44Updated 3 weeks ago
- σ-GPT: A New Approach to Autoregressive Models☆69Updated last year
- Because we don't want a jupyter notebook mess...☆61Updated 5 months ago
- Diffusion models in PyTorch☆114Updated this week
- An implementation of PSGD Kron second-order optimizer for PyTorch☆97Updated 3 months ago
- Explorations into the proposal from the paper "Grokfast, Accelerated Grokking by Amplifying Slow Gradients"☆103Updated 11 months ago
- Exploration into the Firefly algorithm in Pytorch☆41Updated 9 months ago
- ☆53Updated last year
- Various handy scripts to quickly setup new Linux and Windows sandboxes, containers and WSL.☆40Updated last week
- Graph neural networks in JAX.☆68Updated last year
- Pytorch implementation of preconditioned stochastic gradient descent (Kron and affine preconditioner, low-rank approximation precondition…☆188Updated last month
- $100K or 100 Days: Trade-offs when Pre-Training with Academic Resources☆147Updated last month
- H-Net Dynamic Hierarchical Architecture☆80Updated 2 months ago
- A package for defining deep learning models using categorical algebraic expressions.☆61Updated last year
- ☆68Updated last year
- Implementation of the proposed Spline-Based Transformer from Disney Research☆105Updated last year
- ☆140Updated 3 weeks ago
- Induce brain-like topographic structure in your neural networks☆70Updated 3 months ago
- ☆61Updated last year
- A Python Library for Learning Non-Euclidean Representations☆67Updated 3 months ago
- Brain-Inspired Modular Training (BIMT), a method for making neural networks more modular and interpretable.☆173Updated 2 years ago
- LLMs represent numbers on a helix and manipulate that helix to do addition.☆27Updated 9 months ago
- ☆82Updated last year
- Deep Networks Grok All the Time and Here is Why☆37Updated last year
- An interactive exploration of Transformer programming.☆271Updated 2 years ago