aidatatools / ollama-benchmarkLinks
LLM Benchmark for Throughput via Ollama (Local LLMs)
β311Updated 3 months ago
Alternatives and similar repositories for ollama-benchmark
Users that are interested in ollama-benchmark are comparing it to the libraries listed below
Sorting:
- Code execution utilities for Open WebUI & Ollamaβ305Updated last year
- beep boop π€ (experimental)β117Updated 10 months ago
- Download models from the Ollama library, without Ollamaβ109Updated last year
- A proxy server for multiple ollama instances with Key securityβ527Updated last week
- VSCode AI coding assistant powered by self-hosted llama.cpp endpoint.β183Updated 9 months ago
- Benchmark llm performanceβ106Updated last year
- A simple to use Ollama autocompletion engine with options exposed and streaming functionalityβ139Updated 7 months ago
- Dagger functions to import Hugging Face GGUF models into a local ollama instance and optionally push them to ollama.com.β119Updated last year
- A platform to self-host AI on easy modeβ176Updated this week
- Handy tool to measure the performance and efficiency of LLMs workloads.β73Updated 6 months ago
- OpenAPI Tool Serversβ749Updated last month
- Link you Ollama models to LM-Studioβ145Updated last year
- π Retrieval Augmented Generation (RAG) with txtai. Combine search and LLMs to find insights with your own data.β414Updated last week
- Inference engine for Intel devices. Serve LLMs, VLMs, Whisper, Kokoro-TTS, Embedding and Rerank models over OpenAI endpoints.β241Updated last week
- An OpenAI API compatible API for chat with image input and questions about the images. aka Multimodal.β265Updated 8 months ago
- Wraps any OpenAI API interface as Responses with MCPs support so it supports Codex. Adding any missing stateful features. Ollama and Vllmβ¦β128Updated 2 weeks ago
- InferX: Inference as a Service Platformβ138Updated this week
- Web UI and API for managing MCP Orchestrator (mcpo) instances and configurationsβ125Updated 5 months ago
- Comparison of the output quality of quantization methods, using Llama 3, transformers, GGUF, EXL2.β165Updated last year
- Nginx proxy server in a Docker container to Authenticate & Proxy requests to Ollama from Public Internet via Cloudflare Tunnelβ149Updated 2 months ago
- β106Updated 2 months ago
- Parse files (e.g. code repos) and websites to clipboard or a file for ingestions by AI / LLMsβ313Updated 3 months ago
- LLMX; Easiest 3rd party Local LLM UI for the web!β280Updated 2 weeks ago
- A fast batching API to serve LLM modelsβ188Updated last year
- Dolphin System Messagesβ363Updated 9 months ago
- A open webui function for better R1 experienceβ77Updated 8 months ago
- High-performance lightweight proxy and load balancer for LLM infrastructure. Intelligent routing, automatic failover and unified model diβ¦β117Updated this week
- Fully-featured, beautiful web interface for vLLM - built with NextJS.β159Updated 6 months ago
- QA-Pilot is an interactive chat project that leverages online/local LLM for rapid understanding and navigation of GitHub code repository.β312Updated 2 months ago
- OpenβWebUI Tools is a modular toolkit designed to extend and enrich your Open WebUI instance, turning it into a powerful AI workstation. β¦β430Updated 2 weeks ago