aidatatools / ollama-benchmarkLinks
LLM Benchmark for Throughput via Ollama (Local LLMs)
☆291Updated last month
Alternatives and similar repositories for ollama-benchmark
Users that are interested in ollama-benchmark are comparing it to the libraries listed below
Sorting:
- Handy tool to measure the performance and efficiency of LLMs workloads.☆71Updated 4 months ago
- Code execution utilities for Open WebUI & Ollama☆296Updated 10 months ago
- Link you Ollama models to LM-Studio☆142Updated last year
- Benchmark llm performance☆104Updated last year
- A proxy server for multiple ollama instances with Key security☆489Updated last week
- beep boop 🤖 (experimental)☆114Updated 8 months ago
- VSCode AI coding assistant powered by self-hosted llama.cpp endpoint.☆183Updated 7 months ago
- Dagger functions to import Hugging Face GGUF models into a local ollama instance and optionally push them to ollama.com.☆117Updated last year
- 🚀 Retrieval Augmented Generation (RAG) with txtai. Combine search and LLMs to find insights with your own data.☆406Updated 4 months ago
- Download models from the Ollama library, without Ollama☆97Updated 10 months ago
- InferX is a Inference Function as a Service Platform☆133Updated this week
- Parse files (e.g. code repos) and websites to clipboard or a file for ingestions by AI / LLMs☆298Updated last month
- Instantly calculate the maximum size of quantized language models that can fit in your available RAM, helping you optimize your models fo…☆238Updated 4 months ago
- Lightweight Inference server for OpenVINO☆211Updated this week
- OpenAPI Tool Servers☆665Updated last week
- A open webui function for better R1 experience☆79Updated 6 months ago
- QA-Pilot is an interactive chat project that leverages online/local LLM for rapid understanding and navigation of GitHub code repository.☆306Updated 3 weeks ago
- Open‑WebUI Tools is a modular toolkit designed to extend and enrich your Open WebUI instance, turning it into a powerful AI workstation. …☆347Updated 2 weeks ago
- This repository contains custom pipelines developed for the OpenWebUI framework, including advanced workflows such as long-term memory fi…☆74Updated 4 months ago
- Lightweight & fast AI inference proxy for self-hosted LLMs backends like Ollama, LM Studio and others. Designed for speed, simplicity and…☆87Updated last week
- Nginx proxy server in a Docker container to Authenticate & Proxy requests to Ollama from Public Internet via Cloudflare Tunnel☆139Updated 2 weeks ago
- An OpenAI API compatible API for chat with image input and questions about the images. aka Multimodal.☆260Updated 6 months ago
- ☆223Updated 4 months ago
- LLM plugin providing access to models running on an Ollama server☆337Updated this week
- A simple to use Ollama autocompletion engine with options exposed and streaming functionality☆135Updated 5 months ago
- Fully-featured, beautiful web interface for vLLM - built with NextJS.☆152Updated 4 months ago
- ☆99Updated 3 weeks ago
- LLMX; Easiest 3rd party Local LLM UI for the web!☆272Updated 2 weeks ago
- FastMLX is a high performance production ready API to host MLX models.☆331Updated 6 months ago
- llmbasedos — Local-First OS Where Your AI Agents Wake Up and Work☆272Updated 3 weeks ago