arlo-phoenix / CTranslate2-rocmLinks
Fast inference engine for Transformer models
☆46Updated 10 months ago
Alternatives and similar repositories for CTranslate2-rocm
Users that are interested in CTranslate2-rocm are comparing it to the libraries listed below
Sorting:
- ☆391Updated 5 months ago
- An optimized quantization and inference library for running LLMs locally on modern consumer-class GPUs☆493Updated this week
- AI Inferencing at the Edge. A simple one-file way to run various GGML models with KoboldAI's UI with AMD ROCm offloading☆690Updated 3 weeks ago
- The official API server for Exllama. OAI compatible, lightweight, and fast.☆1,047Updated 3 weeks ago
- Linux based GDDR6/GDDR6X VRAM temperature reader for NVIDIA RTX 3000/4000 series GPUs.☆103Updated 4 months ago
- Core, Junction, and VRAM temperature reader for Linux + GDDR6/GDDR6X GPUs☆53Updated 4 months ago
- AMD (Radeon GPU) ROCm based setup for popular AI tools on Ubuntu 24.04.1☆210Updated 6 months ago
- ☆83Updated this week
- A daemon that automatically manages the performance states of NVIDIA GPUs.☆96Updated 2 weeks ago
- Simple monkeypatch to boost AMD Navi 3 GPUs☆45Updated 4 months ago
- Whisper command line client compatible with original OpenAI client based on CTranslate2.☆1,102Updated last month
- Input text from speech in any Linux window, the lean, fast and accurate way, using whisper.cpp OFFLINE. Speak with local LLMs via llama.c…☆138Updated last month
- Run LLMs on AMD Ryzen™ AI NPUs. Just like Ollama - but purpose-built and deeply optimized for the AMD NPUs.☆194Updated this week
- Croco.Cpp is fork of KoboldCPP infering GGML/GGUF models on CPU/Cuda with KoboldAI's UI. It's powered partly by IK_LLama.cpp, and compati…☆136Updated this week
- Prometheus exporter for Linux based GDDR6/GDDR6X VRAM and GPU Core Hot spot temperature reader for NVIDIA RTX 3000/4000 series GPUs.☆23Updated 11 months ago
- An OpenAI API compatible text to speech server using Coqui AI's xtts_v2 and/or piper tts as the backend.☆808Updated 7 months ago
- Lightweight Inference server for OpenVINO☆211Updated this week
- CUDA on AMD GPUs☆560Updated last week
- ROCm docker images with fixes/support for extra architectures, such as gfx803/gfx1010.☆31Updated 2 years ago
- Python bindings for whisper.cpp☆290Updated last week
- Web UI for ExLlamaV2☆513Updated 7 months ago
- Stable Diffusion and Flux in pure C/C++☆21Updated this week
- vLLM for AMD gfx906 GPUs, e.g. Radeon VII / MI50 / MI60☆247Updated this week
- RIFE, Real-Time Intermediate Flow Estimation for Video Frame Interpolation implemented with ncnn library☆55Updated 3 months ago
- 8-bit CUDA functions for PyTorch Rocm compatible☆41Updated last year
- Easy to use interface for the Whisper model optimized for all GPUs!☆300Updated last month
- A complete install script for the complete AMD ROCm ML stack allowing you to get the most out of your hardware with the least amount of s…☆35Updated this week
- ☆42Updated 2 years ago
- ROCm docker images with fixes/support for legecy architecture gfx803. eg.Radeon RX 590/RX 580/RX 570/RX 480☆75Updated 3 months ago
- General Site for the GFX803 ROCm Stuff☆112Updated 3 weeks ago