aidatatools / ollama-benchmarkLinks
LLM Benchmark for Throughput via Ollama (Local LLMs)
☆313Updated 3 months ago
Alternatives and similar repositories for ollama-benchmark
Users that are interested in ollama-benchmark are comparing it to the libraries listed below
Sorting:
- Code execution utilities for Open WebUI & Ollama☆310Updated last year
- Dagger functions to import Hugging Face GGUF models into a local ollama instance and optionally push them to ollama.com.☆119Updated last year
- A proxy server for multiple ollama instances with Key security☆543Updated 3 weeks ago
- Download models from the Ollama library, without Ollama☆115Updated last year
- Benchmark llm performance☆108Updated last year
- beep boop 🤖 (experimental)☆117Updated 11 months ago
- Handy tool to measure the performance and efficiency of LLMs workloads.☆73Updated 7 months ago
- A simple to use Ollama autocompletion engine with options exposed and streaming functionality☆140Updated 8 months ago
- Link you Ollama models to LM-Studio☆146Updated last year
- A open webui function for better R1 experience☆78Updated 9 months ago
- A simple, intuitive toolkit for quickly implementing LLM powered applications.☆270Updated 11 months ago
- VSCode AI coding assistant powered by self-hosted llama.cpp endpoint.☆183Updated 10 months ago
- OpenAPI Tool Servers☆770Updated 2 months ago
- Web UI and API for managing MCP Orchestrator (mcpo) instances and configurations☆127Updated 6 months ago
- Open Deep Researcher with openai compatible endpoint, now completely local with ollama, local playwright via searxng with citations and p…☆148Updated 8 months ago
- ☆108Updated 3 months ago
- 🚀 Retrieval Augmented Generation (RAG) with txtai. Combine search and LLMs to find insights with your own data.☆421Updated last week
- Wraps any OpenAI API interface as Responses with MCPs support so it supports Codex. Adding any missing stateful features. Ollama and Vllm…☆137Updated last month
- InferX: Inference as a Service Platform☆142Updated this week
- The RunPod worker template for serving our large language model endpoints. Powered by vLLM.☆386Updated this week
- An OpenAI API compatible API for chat with image input and questions about the images. aka Multimodal.☆266Updated 9 months ago
- Tool to download models from Huggingface Hub and convert them to GGML/GGUF for llama.cpp☆163Updated 7 months ago
- Comparison of the output quality of quantization methods, using Llama 3, transformers, GGUF, EXL2.☆165Updated last year
- Parse files (e.g. code repos) and websites to clipboard or a file for ingestions by AI / LLMs☆315Updated last week
- An OpenAI API compatible speech to text server for audio transcription and translations, aka. Whisper.☆89Updated 10 months ago
- API up your Ollama Server.☆186Updated last month
- LLM plugin providing access to models running on an Ollama server☆343Updated last month
- Instantly calculate the maximum size of quantized language models that can fit in your available RAM, helping you optimize your models fo…☆242Updated 7 months ago
- LLMX; Easiest 3rd party Local LLM UI for the web!☆282Updated last month
- Nginx proxy server in a Docker container to Authenticate & Proxy requests to Ollama from Public Internet via Cloudflare Tunnel☆152Updated 3 months ago