Talnz007 / VulkanIlmLinks
GPU-accelerated LLaMA inference wrapper for legacy Vulkan-capable systems a Pythonic way to run AI with knowledge (Ilm) on fire (Vulkan).
☆24Updated this week
Alternatives and similar repositories for VulkanIlm
Users that are interested in VulkanIlm are comparing it to the libraries listed below
Sorting:
- A tshark MCP server for packet capture and analysis☆23Updated 4 months ago
- A lightweight LLaMA.cpp HTTP server Docker image based on Alpine Linux.☆29Updated 2 weeks ago
- ☆92Updated last year
- Lightweight cli coding agent☆57Updated 5 months ago
- Editor with LLM generation tree exploration☆77Updated 8 months ago
- LocalScore is an open benchmark which helps you understand how well your computer can handle local AI tasks.☆64Updated last month
- ☆206Updated last month
- *NIX SHELL with Local AI/LLM integration☆23Updated 7 months ago
- ☆62Updated last year
- A web application that converts speech to speech 100% private☆76Updated 4 months ago
- Generate a wiki for your research topic, sourcing from the web and your docs.☆52Updated 7 months ago
- Run multiple resource-heavy Large Models (LM) on the same machine with limited amount of VRAM/other resources by exposing them on differe…☆81Updated last week
- A simple, easy-to-customize pipeline for local RAG evaluation. Starter prompts and metric definitions included.☆24Updated last month
- 🐝 Create powerful, collaborative AI applications.☆64Updated 11 months ago
- A simple tool to anonymize LLM prompts.☆65Updated 8 months ago
- RetroChat is a powerful command-line interface for interacting with various AI language models. It provides a seamless experience for eng…☆81Updated 3 months ago
- Inference engine for Intel devices. Serve LLMs, VLMs, Whisper, Kokoro-TTS over OpenAI endpoints.☆213Updated last week
- Code scanner to check for issues in prompts and LLM calls☆73Updated 6 months ago
- A platform to self-host AI on easy mode☆170Updated this week
- No-messing-around sh client for llama.cpp's server☆30Updated last year
- Chat WebUI is an easy-to-use user interface for interacting with AI, and it comes with multiple useful built-in tools such as web search …☆46Updated last month
- Let LLMs control embedded devices via the Model Context Protocol.☆146Updated 3 months ago
- Lightweight & fast AI inference proxy for self-hosted LLMs backends like Ollama, LM Studio and others. Designed for speed, simplicity and…☆101Updated last week
- George is an API leveraging AI to make it easy to control a computer with natural language.☆50Updated 9 months ago
- LexiCrawler is a powerful Go-based web crawling API meticulously designed to extract, clean, and transform web page content into a pristi…☆48Updated 7 months ago
- Generate a llama-quantize command to copy the quantization parameters of any GGUF☆24Updated 2 months ago
- Merliot Device Hub☆164Updated 4 months ago
- Lightweight C inference for Qwen3 GGUF. Multiturn prefix caching & batch processing.☆18Updated last month
- A document based RAG application☆130Updated 6 months ago
- Documentation site for fast-agent☆21Updated last week