arlo-phoenix / CTranslate2-rocmLinks
Fast inference engine for Transformer models
☆37Updated 8 months ago
Alternatives and similar repositories for CTranslate2-rocm
Users that are interested in CTranslate2-rocm are comparing it to the libraries listed below
Sorting:
- ☆356Updated 3 months ago
- AI Inferencing at the Edge. A simple one-file way to run various GGML models with KoboldAI's UI with AMD ROCm offloading☆647Updated 2 weeks ago
- An optimized quantization and inference library for running LLMs locally on modern consumer-class GPUs☆436Updated this week
- Simple monkeypatch to boost AMD Navi 3 GPUs☆43Updated 2 months ago
- AMD (Radeon GPU) ROCm based setup for popular AI tools on Ubuntu 24.04.1☆209Updated 4 months ago
- llama.cpp fork with additional SOTA quants and improved performance☆652Updated this week
- ROCm docker images with fixes/support for extra architectures, such as gfx803/gfx1010.☆30Updated last year
- Core, Junction, and VRAM temperature reader for Linux + GDDR6/GDDR6X GPUs☆47Updated 2 months ago
- Model swapping for llama.cpp (or any local OpenAPI compatible server)☆1,035Updated 2 weeks ago
- A daemon that automatically manages the performance states of NVIDIA GPUs.☆89Updated last month
- ☆80Updated this week
- The official API server for Exllama. OAI compatible, lightweight, and fast.☆1,000Updated this week
- Linux based GDDR6/GDDR6X VRAM temperature reader for NVIDIA RTX 3000/4000 series GPUs.☆101Updated 2 months ago
- Stable Diffusion and Flux in pure C/C++☆20Updated this week
- Web UI for ExLlamaV2☆503Updated 5 months ago
- Input text from speech in any Linux window, the lean, fast and accurate way, using whisper.cpp OFFLINE. Speak with local LLMs via llama.c…☆114Updated this week
- ☆233Updated 2 years ago
- Croco.Cpp is fork of KoboldCPP infering GGML/GGUF models on CPU/Cuda with KoboldAI's UI. It's powered partly by IK_LLama.cpp, and compati…☆111Updated this week
- ROCm docker images with fixes/support for legecy architecture gfx803. eg.Radeon RX 590/RX 580/RX 570/RX 480☆70Updated last month
- LLM Frontend in a single html file☆517Updated 6 months ago
- Lightweight Inference server for OpenVINO☆188Updated this week
- CUDA on AMD GPUs☆527Updated 2 months ago
- ☆42Updated 2 years ago
- The HIP Environment and ROCm Kit - A lightweight open source build system for HIP and ROCm☆234Updated this week
- AMD APU compatible Ollama. Get up and running with Llama 3.3, DeepSeek-R1, Phi-4, Gemma 3, Mistral Small 3.1 and other large language mod…☆63Updated this week
- build scripts for ROCm☆186Updated last year
- vLLM for AMD gfx906 GPUs, e.g. Radeon VII / MI50 / MI60☆111Updated last week
- A complete install script for the complete AMD ROCm ML stack allowing you to get the most out of your hardware with the least amount of s…☆29Updated last month
- 8-bit CUDA functions for PyTorch Rocm compatible☆41Updated last year
- General Site for the GFX803 ROCm Stuff☆89Updated 3 weeks ago