aidatatools / ollama-benchmarkLinks
LLM Benchmark for Throughput via Ollama (Local LLMs)
☆244Updated 3 weeks ago
Alternatives and similar repositories for ollama-benchmark
Users that are interested in ollama-benchmark are comparing it to the libraries listed below
Sorting:
- Handy tool to measure the performance and efficiency of LLMs workloads.☆66Updated 2 months ago
- Link you Ollama models to LM-Studio☆140Updated 11 months ago
- A open webui function for better R1 experience☆79Updated 3 months ago
- 🚀 Retrieval Augmented Generation (RAG) with txtai. Combine search and LLMs to find insights with your own data.☆382Updated last month
- Download models from the Ollama library, without Ollama☆86Updated 7 months ago
- A proxy server for multiple ollama instances with Key security☆449Updated this week
- beep boop 🤖 (experimental)☆111Updated 5 months ago
- a Repository of Open-WebUI tools to use with your favourite LLMs☆232Updated last week
- Dagger functions to import Hugging Face GGUF models into a local ollama instance and optionally push them to ollama.com.☆115Updated last year
- Code execution utilities for Open WebUI & Ollama☆285Updated 7 months ago
- OpenAPI Tool Servers☆484Updated 2 weeks ago
- InferX is a Inference Function as a Service Platform☆111Updated last week
- ☆204Updated last month
- Web UI for ExLlamaV2☆502Updated 4 months ago
- Turns devices into a scalable LLM platform☆144Updated last week
- Local LLM Server with GPU and NPU Acceleration☆138Updated this week
- Local LLM Powered Recursive Search & Smart Knowledge Explorer☆243Updated 4 months ago
- Nginx proxy server in a Docker container to Authenticate & Proxy requests to Ollama from Public Internet via Cloudflare Tunnel☆119Updated 4 months ago
- What If Language Models Expertly Routed All Inference? WilmerAI allows prompts to be routed to specialized workflows based on the domain …☆705Updated last week
- Smart proxy for LLM APIs that enables model-specific parameter control, automatic mode switching (like Qwen3's /think and /no_think), and…☆48Updated last month
- LLM plugin providing access to models running on an Ollama server☆318Updated last week
- API up your Ollama Server.☆161Updated 2 weeks ago
- Comparison of the output quality of quantization methods, using Llama 3, transformers, GGUF, EXL2.☆154Updated last year
- This project was generated 100% by AI, with one prompt. NOTE: This neuroca project was generated in 3 hours on 3/3/2025. There are depend…☆49Updated 2 months ago
- Tool to download models from Huggingface Hub and convert them to GGML/GGUF for llama.cpp☆152Updated last month
- Model swapping for llama.cpp (or any local OpenAPI compatible server)☆969Updated this week
- ☆95Updated 6 months ago
- This project demonstrates a basic chain-of-thought interaction with any LLM (Large Language Model)☆319Updated 9 months ago
- Ollama client written in Python☆2Updated 6 months ago
- automatically quant GGUF models☆184Updated last week