aidatatools / ollama-benchmarkLinks
LLM Benchmark for Throughput via Ollama (Local LLMs)
☆297Updated last month
Alternatives and similar repositories for ollama-benchmark
Users that are interested in ollama-benchmark are comparing it to the libraries listed below
Sorting:
- A proxy server for multiple ollama instances with Key security☆499Updated 2 weeks ago
- Download models from the Ollama library, without Ollama☆100Updated 10 months ago
- Handy tool to measure the performance and efficiency of LLMs workloads.☆71Updated 5 months ago
- Dagger functions to import Hugging Face GGUF models into a local ollama instance and optionally push them to ollama.com.☆117Updated last year
- beep boop 🤖 (experimental)☆114Updated 9 months ago
- Code execution utilities for Open WebUI & Ollama☆298Updated 10 months ago
- An OpenAI API compatible API for chat with image input and questions about the images. aka Multimodal.☆259Updated 7 months ago
- Benchmark llm performance☆105Updated last year
- OpenAPI Tool Servers☆693Updated 2 weeks ago
- A simple to use Ollama autocompletion engine with options exposed and streaming functionality☆137Updated 6 months ago
- VSCode AI coding assistant powered by self-hosted llama.cpp endpoint.☆183Updated 8 months ago
- Tool to download models from Huggingface Hub and convert them to GGML/GGUF for llama.cpp☆159Updated 5 months ago
- Wraps any OpenAI API interface as Responses with MCPs support so it supports Codex. Adding any missing stateful features. Ollama and Vllm…☆110Updated 3 months ago
- Nginx proxy server in a Docker container to Authenticate & Proxy requests to Ollama from Public Internet via Cloudflare Tunnel☆141Updated last month
- A platform to self-host AI on easy mode☆171Updated last week
- 🚀 Retrieval Augmented Generation (RAG) with txtai. Combine search and LLMs to find insights with your own data.☆408Updated 5 months ago
- Link you Ollama models to LM-Studio☆143Updated last year
- A fast batching API to serve LLM models☆187Updated last year
- InferX: Inference as a Service Platform☆136Updated this week
- LLM plugin for models hosted by OpenRouter☆271Updated 2 weeks ago
- Inference engine for Intel devices. Serve LLMs, VLMs, Whisper, Kokoro-TTS over OpenAI endpoints.☆211Updated this week
- A open webui function for better R1 experience☆78Updated 7 months ago
- Open‑WebUI Tools is a modular toolkit designed to extend and enrich your Open WebUI instance, turning it into a powerful AI workstation. …☆369Updated 2 weeks ago
- Instantly calculate the maximum size of quantized language models that can fit in your available RAM, helping you optimize your models fo…☆241Updated 5 months ago
- llmbasedos — Local-First OS Where Your AI Agents Wake Up and Work☆279Updated last month
- LLM plugin providing access to models running on an Ollama server☆337Updated 3 weeks ago
- A simple, intuitive toolkit for quickly implementing LLM powered applications.☆265Updated 9 months ago
- Parse files (e.g. code repos) and websites to clipboard or a file for ingestions by AI / LLMs☆304Updated 2 months ago
- The RunPod worker template for serving our large language model endpoints. Powered by vLLM.☆371Updated 2 weeks ago
- ☆207Updated last month