scooter-lacroix / Stan-s-ML-StackLinks
A complete package that provides you with all the components needed to get started of dive deeper into Machine Learning Workloads on Consumer AMD cards, providing CUDA functionality through fully leveraging HIP and ROCm!
☆40Updated 3 weeks ago
Alternatives and similar repositories for Stan-s-ML-Stack
Users that are interested in Stan-s-ML-Stack are comparing it to the libraries listed below
Sorting:
- Prometheus exporter for Linux based GDDR6/GDDR6X VRAM and GPU Core Hot spot temperature reader for NVIDIA RTX 3000/4000 series GPUs.☆23Updated last year
- Inference engine for Intel devices. Serve LLMs, VLMs, Whisper, Kokoro-TTS, Embedding and Rerank models over OpenAI endpoints.☆241Updated last week
- A daemon that automatically manages the performance states of NVIDIA GPUs.☆97Updated 2 weeks ago
- Run LLMs on AMD Ryzen™ AI NPUs in minutes. Just like Ollama - but purpose-built and deeply optimized for the AMD NPUs.☆451Updated this week
- ☆48Updated 2 years ago
- ☆414Updated 7 months ago
- AMD (Radeon GPU) ROCm based setup for popular AI tools on Ubuntu 24.04.1☆216Updated 2 weeks ago
- Fresh builds of llama.cpp with AMD ROCm™ 7 acceleration☆103Updated this week
- General Site for the GFX803 ROCm Stuff☆126Updated 2 months ago
- Fast inference engine for Transformer models☆54Updated last year
- llama-swap + a minimal ollama compatible api☆33Updated 2 weeks ago
- An optimized quantization and inference library for running LLMs locally on modern consumer-class GPUs☆571Updated this week
- ☆524Updated this week
- A utility that uses Whisper to transcribe videos and various translation APIs to translate the transcribed text and save them as SRT (sub…☆73Updated last year
- GPU Power and Performance Manager☆61Updated last year
- ☆85Updated last week
- llama.cpp fork with additional SOTA quants and improved performance☆1,329Updated this week
- Comparison of the output quality of quantization methods, using Llama 3, transformers, GGUF, EXL2.☆165Updated last year
- ROCm docker images with fixes/support for legecy architecture gfx803. eg.Radeon RX 590/RX 580/RX 570/RX 480☆76Updated 5 months ago
- ☆49Updated last month
- Run multiple resource-heavy Large Models (LM) on the same machine with limited amount of VRAM/other resources by exposing them on differe…☆83Updated 3 weeks ago
- InferX: Inference as a Service Platform☆138Updated this week
- Web UI for ExLlamaV2☆511Updated 9 months ago
- Everything you need to setup on your AMD system for Machine Learning Stuff☆19Updated 3 months ago
- Reliable model swapping for any local OpenAI compatible server - llama.cpp, vllm, etc☆1,862Updated last week
- Open source LLM UI, compatible with all local LLM providers.☆176Updated last year
- The official API server for Exllama. OAI compatible, lightweight, and fast.☆1,083Updated last week
- A platform to self-host AI on easy mode☆176Updated this week
- A tool to determine whether or not your PC can run a given LLM☆164Updated 9 months ago
- LLM Frontend in a single html file☆663Updated last week