erfanzar / EasyDeL
Accelerate, Optimize performance with streamlined training and serving options with JAX.
☆257Updated last week
Alternatives and similar repositories for EasyDeL:
Users that are interested in EasyDeL are comparing it to the libraries listed below
- (EasyDel Former) is a utility library designed to simplify and enhance the development in JAX☆27Updated 2 weeks ago
- A flexible and efficient implementation of Flash Attention 2.0 for JAX, supporting multiple backends (GPU/TPU/CPU) and platforms (Triton/…☆23Updated 3 weeks ago
- ☆215Updated 8 months ago
- Inference code for LLaMA models in JAX☆116Updated 10 months ago
- JAX implementation of the Llama 2 model☆216Updated last year
- Multipack distributed sampler for fast padding-free training of LLMs☆186Updated 7 months ago
- Legible, Scalable, Reproducible Foundation Models with Named Tensors and Jax☆561Updated this week
- The simplest, fastest repository for training/finetuning medium-sized GPTs.☆103Updated 4 months ago
- Google TPU optimizations for transformers models☆104Updated 2 months ago
- seqax = sequence modeling + JAX☆150Updated last week
- Fast, Modern, Memory Efficient, and Low Precision PyTorch Optimizers☆87Updated 8 months ago
- ☆87Updated 2 weeks ago
- A set of Python scripts that makes your experience on TPU better☆50Updated 8 months ago
- LoRA for arbitrary JAX models and functions☆135Updated last year
- Website for hosting the Open Foundation Models Cheat Sheet.☆264Updated 2 weeks ago
- supporting pytorch FSDP for optimizers☆80Updated 3 months ago
- Understand and test language model architectures on synthetic tasks.☆185Updated 3 weeks ago
- ☆184Updated last month
- 🚀 Efficiently (pre)training foundation models with native PyTorch features, including FSDP for training and SDPA implementation of Flash…☆233Updated this week
- nanoGPT-like codebase for LLM training☆91Updated last week
- ☆158Updated last month
- Automated Identification of Redundant Layer Blocks for Pruning in Large Language Models☆226Updated 11 months ago
- JAX Synergistic Memory Inspector☆171Updated 8 months ago
- Deep learning library implemented from scratch in numpy. Mixtral, Mamba, LLaMA, GPT, ResNet, and other experiments.☆51Updated 11 months ago
- Code for exploring Based models from "Simple linear attention language models balance the recall-throughput tradeoff"☆224Updated last month
- Micro Llama is a small Llama based model with 300M parameters trained from scratch with $500 budget☆145Updated last year
- Minimal (400 LOC) implementation Maximum (multi-node, FSDP) GPT training☆123Updated 11 months ago
- Named Tensors for Legible Deep Learning in JAX☆167Updated last week
- A Jax-based library for designing and training transformer models from scratch.☆282Updated 7 months ago
- PyTorch building blocks for the OLMo ecosystem☆172Updated this week