huggingface / optimum-tpu
Google TPU optimizations for transformers models
☆107Updated 2 months ago
Alternatives and similar repositories for optimum-tpu:
Users that are interested in optimum-tpu are comparing it to the libraries listed below
- PyTorch/XLA integration with JetStream (https://github.com/google/JetStream) for LLM inference"☆59Updated 2 weeks ago
- Load compute kernels from the Hub☆113Updated this week
- Pytorch/XLA SPMD Test code in Google TPU☆23Updated last year
- Manage scalable open LLM inference endpoints in Slurm clusters☆254Updated 9 months ago
- A high-throughput and memory-efficient inference and serving engine for LLMs☆262Updated 6 months ago
- Repo hosting codes and materials related to speeding LLMs' inference using token merging.☆35Updated 11 months ago
- Inference server benchmarking tool☆48Updated last week
- A fast implementation of T5/UL2 in PyTorch using Flash Attention☆99Updated 3 weeks ago
- Q-GaLore: Quantized GaLore with INT4 Projection and Layer-Adaptive Low-Rank Gradients.☆195Updated 8 months ago
- 🚀 Efficiently (pre)training foundation models with native PyTorch features, including FSDP for training and SDPA implementation of Flash…☆238Updated this week
- ☆112Updated this week
- Fast, Modern, Memory Efficient, and Low Precision PyTorch Optimizers☆88Updated 8 months ago
- Easy and lightning fast training of 🤗 Transformers on Habana Gaudi processor (HPU)☆184Updated this week
- ☆67Updated 2 years ago
- PyTorch per step fault tolerance (actively under development)☆273Updated this week
- Collection of autoregressive model implementation☆85Updated last month
- ☆205Updated 2 months ago
- ☆49Updated last year
- an open source reproduction of NVIDIA's nGPT (Normalized Transformer with Representation Learning on the Hypersphere)☆93Updated last month
- ArcticTraining is a framework designed to simplify and accelerate the post-training process for large language models (LLMs)☆64Updated this week
- An efficent implementation of the method proposed in "The Era of 1-bit LLMs"☆153Updated 5 months ago
- Tree Attention: Topology-aware Decoding for Long-Context Attention on GPU clusters☆125Updated 4 months ago
- Code for the paper "QMoE: Practical Sub-1-Bit Compression of Trillion-Parameter Models".☆273Updated last year
- ☆100Updated 10 months ago
- Code for "LayerSkip: Enabling Early Exit Inference and Self-Speculative Decoding", ACL 2024☆284Updated last month
- ☆199Updated 2 weeks ago
- JetStream is a throughput and memory optimized engine for LLM inference on XLA devices, starting with TPUs (and GPUs in future -- PRs wel…☆313Updated this week
- Exploring finetuning public checkpoints on filter 8K sequences on Pile☆115Updated 2 years ago
- ☆125Updated last year
- Large scale 4D parallelism pre-training for 🤗 transformers in Mixture of Experts *(still work in progress)*☆81Updated last year