scooter-lacroix / Stan-s-ML-StackLinks
A complete package that provides you with all the components needed to get started of dive deeper into Machine Learning Workloads on Consumer AMD cards, providing CUDA functionality through fully leveraging HIP and ROCm!
☆44Updated 2 months ago
Alternatives and similar repositories for Stan-s-ML-Stack
Users that are interested in Stan-s-ML-Stack are comparing it to the libraries listed below
Sorting:
- Prometheus exporter for Linux based GDDR6/GDDR6X VRAM and GPU Core Hot spot temperature reader for NVIDIA RTX 3000/4000 series GPUs.☆24Updated last year
- A daemon that automatically manages the performance states of NVIDIA GPUs.☆103Updated last month
- Inference engine for Intel devices. Serve LLMs, VLMs, Whisper, Kokoro-TTS, Embedding and Rerank models over OpenAI endpoints.☆266Updated last week
- Fresh builds of llama.cpp with AMD ROCm™ 7 acceleration☆149Updated this week
- ☆420Updated 8 months ago
- GPU Power and Performance Manager☆64Updated last year
- An optimized quantization and inference library for running LLMs locally on modern consumer-class GPUs☆609Updated this week
- AMD (Radeon GPU) ROCm based setup for popular AI tools on Ubuntu 24.04.1☆216Updated last month
- Web UI for ExLlamaV2☆514Updated 10 months ago
- Core, Junction, and VRAM temperature reader for Linux + GDDR6/GDDR6X GPUs☆64Updated 2 months ago
- Comparison of the output quality of quantization methods, using Llama 3, transformers, GGUF, EXL2.☆165Updated last year
- ☆48Updated 2 years ago
- llama-swap + a minimal ollama compatible api☆38Updated this week
- ☆695Updated last week
- ☆87Updated 3 weeks ago
- A utility that uses Whisper to transcribe videos and various translation APIs to translate the transcribed text and save them as SRT (sub…☆74Updated last year
- Open source LLM UI, compatible with all local LLM providers.☆176Updated last year
- KoboldCpp Smart Launcher with GPU Layer and Tensor Override Tuning☆29Updated 7 months ago
- Easy to use interface for the Whisper model optimized for all GPUs!☆405Updated 4 months ago
- A tool to determine whether or not your PC can run a given LLM☆166Updated 11 months ago
- Input your VRAM and RAM and the toolchain will produce a GGUF model tuned to your system within seconds — flexible model sizing and lowes…☆71Updated this week
- llama.cpp-gfx906☆72Updated last week
- Your Trusty Memory-enabled AI Companion - Simple RAG chatbot optimized for local LLMs | 12 Languages Supported | OpenAI API Compatible☆344Updated 10 months ago
- Y'all thought the dead internet theory wasn't real, but HERE IT IS☆208Updated last year
- InferX: Inference as a Service Platform☆143Updated last week
- Run multiple resource-heavy Large Models (LM) on the same machine with limited amount of VRAM/other resources by exposing them on differe…☆85Updated this week
- ☆228Updated 7 months ago
- A local AI companion that uses a collection of free, open source AI models in order to create two virtual companions that will follow you…☆237Updated 2 months ago
- Code for Papeg.ai☆227Updated 11 months ago
- A library and CLI utilities for managing performance states of NVIDIA GPUs.☆31Updated last year