huggingface / optimum-habana
Easy and lightning fast training of ๐ค Transformers on Habana Gaudi processor (HPU)
โ153Updated this week
Related projects โ
Alternatives and complementary repositories for optimum-habana
- ๐๏ธ A unified multi-backend utility for benchmarking Transformers, Timm, PEFT, Diffusers and Sentence-Transformers with full support of Oโฆโ257Updated this week
- ๐ค Optimum Intel: Accelerate inference with Intel optimization toolsโ409Updated this week
- A high-throughput and memory-efficient inference and serving engine for LLMsโ253Updated last month
- GPTQ inference Triton kernelโ284Updated last year
- Intelยฎ Extension for DeepSpeed* is an extension to DeepSpeed that brings feature support with SYCL kernels on Intel GPU(XPU) device. Noteโฆโ57Updated 2 months ago
- Applied AI experiments and examples for PyTorchโ166Updated 3 weeks ago
- Google TPU optimizations for transformers modelsโ75Updated this week
- Easy and Efficient Quantization for Transformersโ180Updated 4 months ago
- A high-throughput and memory-efficient inference and serving engine for LLMsโ42Updated this week
- Large Language Model Text Generation Inference on Habana Gaudiโ26Updated last week
- ๐ Collection of components for development, training, tuning, and inference of foundation models leveraging PyTorch native components.โ165Updated this week
- QServe: W4A8KV4 Quantization and System Co-design for Efficient LLM Servingโ443Updated last week
- FP16xINT4 LLM inference kernel that can achieve near-ideal ~4x speedups up to medium batchsizes of 16-32 tokens.โ624Updated 2 months ago
- vLLM: A high-throughput and memory-efficient inference and serving engine for LLMsโ89Updated this week
- Blazing fast training of ๐ค Transformers on Graphcore IPUsโ82Updated 8 months ago
- This repository contains the experimental PyTorch native float8 training UXโ211Updated 3 months ago
- ๐ Efficiently (pre)training foundation models with native PyTorch features, including FSDP for training and SDPA implementation of Flashโฆโ193Updated this week
- [NeurIPS 2024] KVQuant: Towards 10 Million Context Length LLM Inference with KV Cache Quantizationโ305Updated 3 months ago
- โ111Updated 8 months ago
- Advanced Quantization Algorithm for LLMs. This is official implementation of "Optimize Weight Rounding via Signed Gradient Descent for tโฆโ248Updated this week
- Reference models for Intel(R) Gaudi(R) AI Acceleratorโ155Updated 2 weeks ago
- An efficient GPU support for LLM inference with x-bit quantization (e.g. FP6,FP5).โ208Updated 3 weeks ago
- Latency and Memory Analysis of Transformer Models for Training and Inferenceโ355Updated last week
- โ191Updated this week
- A performant, memory-efficient checkpointing library for PyTorch applications, designed with large, complex distributed workloads in mindโฆโ146Updated 2 weeks ago
- Dynamic batching library for Deep Learning inference. Tutorials for LLM, GPT scenarios.โ86Updated 3 months ago
- Fast Matrix Multiplications for Lookup Table-Quantized LLMsโ187Updated this week
- The Triton backend for the ONNX Runtime.โ132Updated this week
- [MLSys'24] Atom: Low-bit Quantization for Efficient and Accurate LLM Servingโ278Updated 4 months ago
- An innovative library for efficient LLM inference via low-bit quantizationโ348Updated 2 months ago