srush / anynp
Proof-of-concept of global switching between numpy/jax/pytorch in a library.
☆18Updated 10 months ago
Alternatives and similar repositories for anynp:
Users that are interested in anynp are comparing it to the libraries listed below
- ☆20Updated last year
- Experiment of using Tangent to autodiff triton☆78Updated last year
- Code associated to papers on superposition (in ML interpretability)☆27Updated 2 years ago
- gzip Predicts Data-dependent Scaling Laws☆34Updated 11 months ago
- ☆60Updated 3 years ago
- JAX implementation of the Mistral 7b v0.2 model☆35Updated 10 months ago
- Official repository for the paper "Can You Learn an Algorithm? Generalizing from Easy to Hard Problems with Recurrent Networks"☆59Updated 3 years ago
- some common Huggingface transformers in maximal update parametrization (µP)☆80Updated 3 years ago
- Jax like function transformation engine but micro, microjax☆31Updated 6 months ago
- A place to store reusable transformer components of my own creation or found on the interwebs☆55Updated this week
- ☆21Updated 2 months ago
- A JAX implementation of stochastic addition.☆14Updated 2 years ago
- ☆52Updated 7 months ago
- ☆27Updated 9 months ago
- A simple library for scaling up JAX programs☆134Updated 6 months ago
- A case study of efficient training of large language models using commodity hardware.☆69Updated 2 years ago
- This is a port of Mistral-7B model in JAX☆32Updated 10 months ago
- ☆78Updated 10 months ago
- Triton Implementation of HyperAttention Algorithm☆47Updated last year
- Large scale 4D parallelism pre-training for 🤗 transformers in Mixture of Experts *(still work in progress)*☆82Updated last year
- A MAD laboratory to improve AI architecture designs 🧪☆113Updated 4 months ago
- Implementing RASP transformer programming language https://arxiv.org/pdf/2106.06981.pdf.☆52Updated 3 years ago
- nanoGPT-like codebase for LLM training☆94Updated last month
- Einsum-like high-level array sharding API for JAX☆34Updated 9 months ago
- Distributed pretraining of large language models (LLMs) on cloud TPU slices, with Jax and Equinox.☆24Updated 7 months ago
- train with kittens!☆57Updated 6 months ago
- ☆37Updated last year
- Resources from the EleutherAI Math Reading Group☆53Updated 2 months ago
- Automatically take good care of your preemptible TPUs☆36Updated last year
- Experiments for efforts to train a new and improved t5☆77Updated last year