aidatatools / ollama-benchmark
LLM Benchmark for Throughput via Ollama (Local LLMs)
☆93Updated last month
Related projects: ⓘ
- Dagger functions to import Hugging Face GGUF models into a local ollama instance and optionally push them to ollama.com.☆109Updated 3 months ago
- Easily view and modify JSON datasets for large language models☆55Updated this week
- PyPlexitas is an open-source Python CLI alternative to Perplexity AI, designed to perform web searches, scrape content, generate embeddin…☆29Updated 3 months ago
- Effortlessly run LLM backends, APIs, frontends, and services with one command.☆199Updated this week
- Complex RAG backend☆28Updated 5 months ago
- 👁️ Multimodal LLM vision multitool☆18Updated 2 weeks ago
- A local AI companion that uses a collection of free, open source AI models in order to create two virtual companions that will follow you…☆68Updated 3 weeks ago
- An OpenAI API compatible API for chat with image input and questions about the images. aka Multimodal.☆157Updated this week
- For inferring and serving local LLMs using the MLX framework☆77Updated 5 months ago
- GPU Power and Performance Manager☆39Updated 2 weeks ago
- ☆95Updated this week
- Something similar to Apple Intelligence?☆54Updated 2 months ago
- automatically quant GGUF models☆119Updated this week
- 🚀 Retrieval Augmented Generation (RAG) with txtai. Combine search and LLMs to find insights with your own data.☆226Updated 2 weeks ago
- Large Model Proxy is designed to make it easy to run multiple resource-heavy Large Models (LM) on the same machine with limited amount of…☆44Updated last month
- An endpoint server for efficiently serving quantized open-source LLMs for code.☆52Updated 11 months ago
- run ollama & gguf easily with a single command☆46Updated 4 months ago
- A collection of prompts to challenge the reasoning abilities of large language models in presence of misguiding information☆51Updated this week
- Serving LLMs in the HF-Transformers format via a PyFlask API☆68Updated last week
- multi1: create o1-like reasoning chains with multiple AI providers (and locally)☆64Updated this week
- Comparison of the output quality of quantization methods, using Llama 3, transformers, GGUF, EXL2.☆113Updated 4 months ago
- This is the Mixture-of-Agents (MoA) concept, adapted from the original work by TogetherAI. My version is tailored for local model usage a…☆101Updated 2 months ago
- ☆37Updated 2 months ago
- A simple light terminal style chat app that lets you use connect to your local llama.cpp server☆27Updated 2 months ago
- Self-host LLMs with vLLM and BentoML☆62Updated this week
- This small API downloads and exposes access to NeuML's txtai-wikipedia and full wikipedia datasets, taking in a query and returning full …☆41Updated last month
- Mixture-of-Ollamas☆25Updated last month
- Practical and advanced guide to LLMOps. It provides a solid understanding of large language models’ general concepts, deployment techniqu…☆46Updated last month
- ☆64Updated 3 months ago
- Local LLM inference & management server with built-in OpenAI API☆30Updated 5 months ago