huggingface / optimum-furiosa
Accelerated inference of π€ models using FuriosaAI NPU chips.
β26Updated 7 months ago
Alternatives and similar repositories for optimum-furiosa:
Users that are interested in optimum-furiosa are comparing it to the libraries listed below
- β16Updated this week
- Google TPU optimizations for transformers modelsβ90Updated last week
- **ARCHIVED** Filesystem interface to π€ Hubβ58Updated last year
- manage histories of LLM applied applicationsβ88Updated last year
- [WIP] A π₯ interface for running code in the cloudβ86Updated last year
- PyTorch/XLA integration with JetStream (https://github.com/google/JetStream) for LLM inference"β48Updated last week
- Hugging Face's Zapier Integration π€β‘οΈβ47Updated last year
- Pixel Parsing. A reproduction of OCR-free end-to-end document understanding models with open dataβ21Updated 6 months ago
- Repository for Sparse Finetuning of LLMs via modified version of the MosaicML llmfoundryβ40Updated last year
- QLoRA with Enhanced Multi GPU Supportβ36Updated last year
- β31Updated last year
- Local emulator for Hugging Face Inference Endpoints customer handlersβ25Updated last year
- Zeus LLM Trainer is a rewrite of Stanford Alpaca aiming to be the trainer for all Large Language Modelsβ69Updated last year
- β22Updated last year
- GPTQLoRA: Efficient Finetuning of Quantized LLMs with GPTQβ99Updated last year
- Low-Rank Adaptation of Large Language Models clean implementationβ9Updated last year
- β54Updated last year
- Visualize expert firing frequencies across sentences in the Mixtral MoE modelβ17Updated last year
- Blazing fast training of π€ Transformers on Graphcore IPUsβ85Updated 10 months ago
- π¨ Imagine what Picasso could have done with AI. Self-host your StableDiffusion API.β50Updated last year
- vLLM: A high-throughput and memory-efficient inference and serving engine for LLMsβ88Updated this week
- A library for squeakily cleaning and filtering language datasets.β45Updated last year
- A safetensors extension to efficiently store sparse quantized tensors on diskβ66Updated this week
- Code for the examples presented in the talk "Training a Llama in your backyard: fine-tuning very large models on consumer hardware" givenβ¦β14Updated last year
- A collection of all available inference solutions for the LLMsβ76Updated 4 months ago
- β165Updated last week
- ML/DL Math and Method notesβ58Updated last year
- Repository for CPU Kernel Generation for LLM Inferenceβ25Updated last year
- Sentence Embedding as a Serviceβ14Updated last year
- GitHub action that'll sync files from a GitHub Repo with the Hugging Face Hub π€β69Updated 3 months ago