aidatatools / ollama-benchmarkLinks
LLM Benchmark for Throughput via Ollama (Local LLMs)
☆255Updated 2 weeks ago
Alternatives and similar repositories for ollama-benchmark
Users that are interested in ollama-benchmark are comparing it to the libraries listed below
Sorting:
- Dagger functions to import Hugging Face GGUF models into a local ollama instance and optionally push them to ollama.com.☆116Updated last year
- A proxy server for multiple ollama instances with Key security☆461Updated last week
- Benchmark llm performance☆100Updated 11 months ago
- Local LLM Server with GPU and NPU Acceleration☆206Updated this week
- Code execution utilities for Open WebUI & Ollama☆290Updated 8 months ago
- Fully-featured, beautiful web interface for vLLM - built with NextJS.☆146Updated 2 months ago
- a Repository of Open-WebUI tools to use with your favourite LLMs☆247Updated this week
- InferX is a Inference Function as a Service Platform☆116Updated 2 weeks ago
- beep boop 🤖 (experimental)☆110Updated 6 months ago
- Nginx proxy server in a Docker container to Authenticate & Proxy requests to Ollama from Public Internet via Cloudflare Tunnel☆123Updated 2 weeks ago
- A open webui function for better R1 experience☆79Updated 4 months ago
- A simple to use Ollama autocompletion engine with options exposed and streaming functionality☆130Updated 3 months ago
- Lightweight Inference server for OpenVINO☆188Updated this week
- Download models from the Ollama library, without Ollama☆89Updated 8 months ago
- VSCode AI coding assistant powered by self-hosted llama.cpp endpoint.☆183Updated 5 months ago
- Handy tool to measure the performance and efficiency of LLMs workloads.☆67Updated 2 months ago
- Link you Ollama models to LM-Studio☆140Updated last year
- ☆95Updated 6 months ago
- Serving LLMs in the HF-Transformers format via a PyFlask API☆71Updated 10 months ago
- Wraps any OpenAI API interface as Responses with MCPs support so it supports Codex. Adding any missing stateful features. Ollama and Vllm…☆72Updated 2 weeks ago
- 🚀 Retrieval Augmented Generation (RAG) with txtai. Combine search and LLMs to find insights with your own data.☆385Updated 2 months ago
- Optimized Ollama LLM server configuration for Mac Studio and other Apple Silicon Macs. Headless setup with automatic startup, resource op…☆205Updated 4 months ago
- Create Linux commands from natural language, in the shell.☆111Updated last week
- Minimal Linux OS with a Model Context Protocol (MCP) gateway to expose local capabilities to LLMs.☆257Updated 3 weeks ago
- Generate train.jsonl and valid.jsonl files to use for fine-tuning Mistral and other LLMs.☆94Updated last year
- AI powered Chatbot with real time updates.☆60Updated 8 months ago
- OpenAPI Tool Servers☆531Updated 3 weeks ago
- Your gateway to both Ollama & Apple MlX models☆140Updated 4 months ago
- An OpenAI API compatible API for chat with image input and questions about the images. aka Multimodal.☆257Updated 4 months ago
- This repository contains custom pipelines developed for the OpenWebUI framework, including advanced workflows such as long-term memory fi…☆70Updated 2 months ago