google / qwixLinks
a Jax quantization library
☆84Updated this week
Alternatives and similar repositories for qwix
Users that are interested in qwix are comparing it to the libraries listed below
Sorting:
- Minimal yet performant LLM examples in pure JAX☆226Updated 2 weeks ago
- torchax is a PyTorch frontend for JAX. It gives JAX the ability to author JAX programs using familiar PyTorch syntax. It also provides JA…☆166Updated last week
- ☆296Updated this week
- JAX-Toolbox☆377Updated this week
- ☆192Updated this week
- jax-triton contains integrations between JAX and OpenAI Triton☆436Updated last month
- ☆151Updated last week
- A simple library for scaling up JAX programs☆144Updated 2 months ago
- A user-friendly tool chain that enables the seamless execution of ONNX models using JAX as the backend.☆130Updated 3 weeks ago
- A FlashAttention implementation for JAX with support for efficient document mask computation and context parallelism.☆156Updated 2 months ago
- ☆287Updated last year
- seqax = sequence modeling + JAX☆169Updated 5 months ago
- JAX bindings for Flash Attention v2☆102Updated 2 weeks ago
- 🧱 Modula software package☆322Updated 4 months ago
- Tokamax: A GPU and TPU kernel library.☆158Updated this week
- Modular, scalable library to train ML models☆187Updated this week
- FlashRNN - Fast RNN Kernels with I/O Awareness☆174Updated 2 months ago
- PyTorch/XLA integration with JetStream (https://github.com/google/JetStream) for LLM inference"☆79Updated last month
- ☆342Updated last week
- Orbax provides common checkpointing and persistence utilities for JAX users☆475Updated this week
- Implementation of Flash Attention in Jax☆223Updated last year
- Load compute kernels from the Hub☆359Updated last week
- ☆70Updated last year
- Named Tensors for Legible Deep Learning in JAX☆215Updated 2 months ago
- Accelerated First Order Parallel Associative Scan☆193Updated last week
- A set of Python scripts that makes your experience on TPU better☆55Updated 4 months ago
- Dion optimizer algorithm☆416Updated last week
- A zero-to-one guide on scaling modern transformers with n-dimensional parallelism.☆113Updated 2 weeks ago
- supporting pytorch FSDP for optimizers☆84Updated last year
- Inference code for LLaMA models in JAX☆120Updated last year