aidatatools / ollama-benchmarkLinks
LLM Benchmark for Throughput via Ollama (Local LLMs)
β269Updated last month
Alternatives and similar repositories for ollama-benchmark
Users that are interested in ollama-benchmark are comparing it to the libraries listed below
Sorting:
- A proxy server for multiple ollama instances with Key securityβ470Updated last week
- beep boop π€ (experimental)β112Updated 7 months ago
- Code execution utilities for Open WebUI & Ollamaβ290Updated 8 months ago
- Lemonade helps users run local LLMs with the highest performance by configuring state-of-the-art inference engines for their NPUs and GPUβ¦β381Updated this week
- Download models from the Ollama library, without Ollamaβ90Updated 8 months ago
- Handy tool to measure the performance and efficiency of LLMs workloads.β69Updated 3 months ago
- OpenAPI Tool Serversβ573Updated last month
- Dagger functions to import Hugging Face GGUF models into a local ollama instance and optionally push them to ollama.com.β116Updated last year
- Link you Ollama models to LM-Studioβ141Updated last year
- π Retrieval Augmented Generation (RAG) with txtai. Combine search and LLMs to find insights with your own data.β390Updated 3 months ago
- VSCode AI coding assistant powered by self-hosted llama.cpp endpoint.β183Updated 6 months ago
- Benchmark llm performanceβ101Updated last year
- An OpenAI API compatible API for chat with image input and questions about the images. aka Multimodal.β260Updated 5 months ago
- Fully-featured, beautiful web interface for vLLM - built with NextJS.β149Updated 3 months ago
- Comparison of the output quality of quantization methods, using Llama 3, transformers, GGUF, EXL2.β158Updated last year
- Nginx proxy server in a Docker container to Authenticate & Proxy requests to Ollama from Public Internet via Cloudflare Tunnelβ128Updated last month
- A fast batching API to serve LLM modelsβ185Updated last year
- OpenβWebUI Tools is a modular toolkit designed to extend and enrich your Open WebUI instance, turning it into a powerful AI workstation. β¦β294Updated last week
- Lightweight Inference server for OpenVINOβ191Updated 2 weeks ago
- β207Updated 2 weeks ago
- This repository contains custom pipelines developed for the OpenWebUI framework, including advanced workflows such as long-term memory fiβ¦β71Updated 2 months ago
- A open webui function for better R1 experienceβ79Updated 5 months ago
- Dolphin System Messagesβ323Updated 5 months ago
- InferX is a Inference Function as a Service Platformβ119Updated 2 weeks ago
- This project demonstrates a basic chain-of-thought interaction with any LLM (Large Language Model)β321Updated 10 months ago
- You donβt need to read the code to understand how to build!β202Updated 6 months ago
- A platform to self-host AI on easy modeβ156Updated this week
- Official python implementation of the UTCPβ364Updated last week
- β95Updated 7 months ago
- Optimized Ollama LLM server configuration for Mac Studio and other Apple Silicon Macs. Headless setup with automatic startup, resource opβ¦β225Updated 5 months ago