andyehrenberg / flaxlm
☆27Updated this week
Related projects: ⓘ
- Automatically take good care of your preemptible TPUs☆28Updated last year
- HomebrewNLP in JAX flavour for maintable TPU-Training☆46Updated 7 months ago
- ☆20Updated last year
- Experiment of using Tangent to autodiff triton☆66Updated 7 months ago
- Machine Learning eXperiment Utilities☆42Updated 3 months ago
- ☆35Updated 5 months ago
- ☆56Updated 2 years ago
- some common Huggingface transformers in maximal update parametrization (µP)☆76Updated 2 years ago
- A library to create and manage configuration files, especially for machine learning projects.☆77Updated 2 years ago
- ☆48Updated 3 months ago
- ☆67Updated 2 years ago
- Code Release for "Broken Neural Scaling Laws" (BNSL) paper☆57Updated 10 months ago
- A simple library for scaling up JAX programs☆116Updated last month
- ☆27Updated 4 months ago
- Engineering the state of RNN language models (Mamba, RWKV, etc.)☆31Updated 3 months ago
- A MAD laboratory to improve AI architecture designs 🧪☆84Updated 4 months ago
- QAmeleon introduces synthetic multilingual QA data using PaLM, a 540B large language model. This dataset was generated by prompt tuning P…☆33Updated last year
- Meta-learning inductive biases in the form of useful conserved quantities.☆37Updated last year
- Blog post☆16Updated 7 months ago
- LoRA for arbitrary JAX models and functions☆127Updated 6 months ago
- Unofficial but Efficient Implementation of "Mamba: Linear-Time Sequence Modeling with Selective State Spaces" in JAX☆74Updated 7 months ago
- Implementation of a Transformer that Ponders, using the scheme from the PonderNet paper☆78Updated 2 years ago
- ☆47Updated 3 months ago
- My explorations into editing the knowledge and memories of an attention network☆34Updated last year
- ☆68Updated 2 months ago
- Silly twitter torch implementations.☆46Updated last year
- If it quacks like a tensor...☆48Updated 7 months ago
- Transformer with Mu-Parameterization, implemented in Jax/Flax. Supports FSDP on TPU pods.☆29Updated 3 weeks ago
- A case study of efficient training of large language models using commodity hardware.☆68Updated 2 years ago
- ☆66Updated 3 months ago