huggingface / optimum-furiosaLinks
Accelerated inference of 🤗 models using FuriosaAI NPU chips.
☆26Updated 2 months ago
Alternatives and similar repositories for optimum-furiosa
Users that are interested in optimum-furiosa are comparing it to the libraries listed below
Sorting:
- ☆171Updated 6 months ago
- Hugging Face's Zapier Integration 🤗⚡️☆47Updated 2 years ago
- **ARCHIVED** Filesystem interface to 🤗 Hub☆58Updated 2 years ago
- Blazing fast training of 🤗 Transformers on Graphcore IPUs☆86Updated last year
- [WIP] A 🔥 interface for running code in the cloud☆85Updated 2 years ago
- manage histories of LLM applied applications☆91Updated last year
- Tune MPTs☆84Updated 2 years ago
- Developing tools to automatically analyze datasets☆74Updated 10 months ago
- Google TPU optimizations for transformers models☆120Updated 7 months ago
- A more memory-efficient rewrite of the HF transformers implementation of Llama for use with quantized weights.☆64Updated last year
- ☆67Updated 3 years ago
- ☆18Updated last week
- Easy and lightning fast training of 🤗 Transformers on Habana Gaudi processor (HPU)☆193Updated last week
- OSLO: Open Source for Large-scale Optimization☆175Updated last year
- 🎨 Imagine what Picasso could have done with AI. Self-host your StableDiffusion API.☆50Updated 2 years ago
- 🏋️ A unified multi-backend utility for benchmarking Transformers, Timm, PEFT, Diffusers and Sentence-Transformers with full support of O…☆310Updated last week
- ☆199Updated last year
- O-GIA is an umbrella for research, infrastructure and projects ecosystem that should provide open source, reproducible datasets, models, …☆90Updated 2 years ago
- Command-line script for inferencing from models such as LLaMA, in a chat scenario, with LoRA adaptations☆33Updated 2 years ago
- DeepSpeed is a deep learning optimization library that makes distributed training easy, efficient, and effective.☆168Updated last month
- Python bindings for ggml☆146Updated 11 months ago
- ☆50Updated last year
- Easy, fast and very cheap training and inference on AWS Trainium and Inferentia chips.☆237Updated this week
- DiffusionWithAutoscaler☆29Updated last year
- experiments with inference on llama☆104Updated last year
- Advanced Ultra-Low Bitrate Compression Techniques for the LLaMA Family of LLMs☆110Updated last year
- ☆53Updated 2 years ago
- ☆251Updated this week
- ☆26Updated 2 years ago
- Drop in replacement for OpenAI, but with Open models.☆152Updated 2 years ago