EleutherAI / pyfra
Python Research Framework
☆107Updated 2 years ago
Related projects ⓘ
Alternatives and complementary repositories for pyfra
- Swarm training framework using Haiku + JAX + Ray for layer parallel transformer language models on unreliable, heterogeneous nodes☆237Updated last year
- ☆38Updated last year
- Implementation of the specific Transformer architecture from PaLM - Scaling Language Modeling with Pathways - in Jax (Equinox framework)☆185Updated 2 years ago
- A GPT, made only of MLPs, in Jax☆55Updated 3 years ago
- Contrastive Language-Image Pretraining☆143Updated 2 years ago
- A case study of efficient training of large language models using commodity hardware.☆68Updated 2 years ago
- ☆57Updated 2 years ago
- GPT, but made only out of MLPs☆86Updated 3 years ago
- ☆155Updated 4 years ago
- Learned Hyperparameter Optimizers☆58Updated 3 years ago
- See the issue board for the current status of active and prospective projects!☆65Updated 2 years ago
- One stop shop for all things carp☆59Updated 2 years ago
- The official repository for our paper "The Devil is in the Detail: Simple Tricks Improve Systematic Generalization of Transformers". We s…☆66Updated last year
- Babysit your preemptible TPUs☆84Updated last year
- A python library for highly configurable transformers - easing model architecture search and experimentation.☆49Updated 2 years ago
- ☆64Updated 2 years ago
- A collection of optimizers, some arcane others well known, for Flax.☆29Updated 3 years ago
- Re-implementation of 'Grokking: Generalization beyond overfitting on small algorithmic datasets'☆38Updated 2 years ago
- Submissions for AI and Efficiency SOTA's☆56Updated 4 years ago
- A Pytree Module system for Deep Learning in JAX☆214Updated last year
- gpu tester detects broken and slow gpus in a cluster☆67Updated last year
- A library to create and manage configuration files, especially for machine learning projects.☆77Updated 2 years ago
- Amos optimizer with JEstimator lib.☆81Updated 6 months ago
- A pure-functional implementation of a machine learning transformer model in Python/JAX☆175Updated 2 years ago
- A port of muP to JAX/Haiku☆25Updated 2 years ago
- Train very large language models in Jax.☆195Updated last year
- Implementation of a Transformer that Ponders, using the scheme from the PonderNet paper☆79Updated 3 years ago
- Official code for "Distributed Deep Learning in Open Collaborations" (NeurIPS 2021)☆116Updated 2 years ago
- Functional deep learning☆106Updated last year