bhavnicksm / vanilla-transformer-jax
JAX/Flax implimentation of 'Attention Is All You Need' by Vaswani et al. (https://arxiv.org/abs/1706.03762)
โ14Updated 3 years ago
Alternatives and similar repositories for vanilla-transformer-jax:
Users that are interested in vanilla-transformer-jax are comparing it to the libraries listed below
- HomebrewNLP in JAX flavour for maintable TPU-Trainingโ49Updated last year
- โ66Updated 2 years ago
- Large scale 4D parallelism pre-training for ๐ค transformers in Mixture of Experts *(still work in progress)*โ81Updated last year
- โ20Updated last year
- Unofficial but Efficient Implementation of "Mamba: Linear-Time Sequence Modeling with Selective State Spaces" in JAXโ83Updated last year
- Flexibly track outputs and grad-outputs of torch.nn.Module.โ13Updated last year
- data related codebase for polyglot projectโ19Updated 2 years ago
- โ60Updated 3 years ago
- LoRA for arbitrary JAX models and functionsโ135Updated last year
- Implementation of Token Shift GPT - An autoregressive model that solely relies on shifting the sequence space for mixingโ48Updated 3 years ago
- โ33Updated 6 months ago
- Implementation of numerous Vision Transformers in Google's JAX and Flax.โ22Updated 2 years ago
- A port of muP to JAX/Haikuโ25Updated 2 years ago
- A basic pure pytorch implementation of flash attentionโ16Updated 5 months ago
- Automatically take good care of your preemptible TPUsโ36Updated last year
- Latent Diffusion Language Modelsโ68Updated last year
- Machine Learning eXperiment Utilitiesโ46Updated 9 months ago
- QAmeleon introduces synthetic multilingual QA data using PaLM, a 540B large language model. This dataset was generated by prompt tuning Pโฆโ34Updated last year
- Contains my experiments with the `big_vision` repo to train ViTs on ImageNet-1k.โ22Updated 2 years ago
- Implementation of the specific Transformer architecture from PaLM - Scaling Language Modeling with Pathways - in Jax (Equinox framework)โ187Updated 2 years ago
- Various transformers for FSDP researchโ37Updated 2 years ago
- This is a port of Mistral-7B model in JAXโ32Updated 9 months ago
- Implementation of some personal helper functions for Einops, my most favorite tensor manipulation library โค๏ธโ54Updated 2 years ago
- JAX implementation of the Mistral 7b v0.2 modelโ35Updated 9 months ago
- Repo for training MLMs, CLMs, or T5-type models on the OLM pretraining data, but it should work with any hugging face text dataset.โ93Updated 2 years ago
- โ67Updated 2 years ago
- Code associated to papers on superposition (in ML interpretability)โ28Updated 2 years ago
- Embedding Recycling for Language modelsโ38Updated last year
- Pytorch/XLA SPMD Test code in Google TPUโ23Updated last year
- some common Huggingface transformers in maximal update parametrization (ยตP)โ80Updated 3 years ago