jrosseruk / Torch2Jax-DeepSeek-R1-Distill-Qwen-1.5BLinks
Flax (Jax) implementation of DeepSeek-R1-Distill-Qwen-1.5B with weights ported from Hugging Face.
☆22Updated 7 months ago
Alternatives and similar repositories for Torch2Jax-DeepSeek-R1-Distill-Qwen-1.5B
Users that are interested in Torch2Jax-DeepSeek-R1-Distill-Qwen-1.5B are comparing it to the libraries listed below
Sorting:
- Minimal but scalable implementation of large language models in JAX☆35Updated 2 weeks ago
- Latent Program Network (from the "Searching Latent Program Spaces" paper)☆96Updated 6 months ago
- Minimal yet performant LLM examples in pure JAX☆158Updated last week
- The simplest, fastest repository for training/finetuning medium-sized GPTs.☆162Updated 2 months ago
- seqax = sequence modeling + JAX☆167Updated last month
- ☆108Updated last week
- A flexible and efficient implementation of Flash Attention 2.0 for JAX, supporting multiple backends (GPU/TPU/CPU) and platforms (Triton/…☆27Updated 6 months ago
- A MAD laboratory to improve AI architecture designs 🧪☆129Updated 9 months ago
- ☆281Updated last year
- Accelerate, Optimize performance with streamlined training and serving options with JAX.☆310Updated this week
- A simple library for scaling up JAX programs☆143Updated 10 months ago
- A set of Python scripts that makes your experience on TPU better☆54Updated last year
- Cost aware hyperparameter tuning algorithm☆168Updated last year
- Custom triton kernels for training Karpathy's nanoGPT.☆19Updated 11 months ago
- Understand and test language model architectures on synthetic tasks.☆225Updated 2 months ago
- LoRA for arbitrary JAX models and functions☆142Updated last year
- 🧱 Modula software package☆237Updated last month
- nanoGPT-like codebase for LLM training☆107Updated 4 months ago
- Normalized Transformer (nGPT)☆188Updated 10 months ago
- A JAX-native LLM Post-Training Library☆144Updated this week
- ☆101Updated this week
- ☆34Updated 9 months ago
- Jax/Flax rewrite of Karpathy's nanoGPT☆60Updated 2 years ago
- Implementation of Diffusion Transformer (DiT) in JAX☆291Updated last year
- Einsum-like high-level array sharding API for JAX☆35Updated last year
- 📄Small Batch Size Training for Language Models☆62Updated 3 weeks ago
- Distributed pretraining of large language models (LLMs) on cloud TPU slices, with Jax and Equinox.☆24Updated 11 months ago
- Evaluating the Mamba architecture on the Othello game☆48Updated last year
- ☆142Updated last week
- supporting pytorch FSDP for optimizers☆84Updated 9 months ago