yoziru / nextjs-vllm-uiLinks
Fully-featured, beautiful web interface for vLLM - built with NextJS.
☆149Updated 2 months ago
Alternatives and similar repositories for nextjs-vllm-ui
Users that are interested in nextjs-vllm-ui are comparing it to the libraries listed below
Sorting:
- An OpenAI API compatible API for chat with image input and questions about the images. aka Multimodal.☆259Updated 4 months ago
- ☆94Updated 7 months ago
- automatically quant GGUF models☆188Updated last week
- This is the Mixture-of-Agents (MoA) concept, adapted from the original work by TogetherAI. My version is tailored for local model usage a…☆117Updated last year
- ☆132Updated 3 months ago
- The RunPod worker template for serving our large language model endpoints. Powered by vLLM.☆337Updated this week
- A fast batching API to serve LLM models☆185Updated last year
- Dataset Crafting w/ RAG/Wikipedia ground truth and Efficient Fine-Tuning Using MLX and Unsloth. Includes configurable dataset annotation …☆184Updated last year
- Dagger functions to import Hugging Face GGUF models into a local ollama instance and optionally push them to ollama.com.☆116Updated last year
- An extension for oobabooga/text-generation-webui that enables the LLM to search the web using DuckDuckGo☆251Updated 2 weeks ago
- Docker compose to run vLLM on Windows☆98Updated last year
- Inference service for Qwen2.5-VL-7b model☆191Updated 4 months ago
- Unsloth Studio☆98Updated 3 months ago
- A open webui function for better R1 experience☆79Updated 4 months ago
- ☆152Updated last week
- Simple UI for Llama-3.2-11B-Vision & Molmo-7B-D☆136Updated 10 months ago
- multi1: create o1-like reasoning chains with multiple AI providers (and locally). Supports LiteLLM as backend too for 100+ providers at o…☆348Updated 6 months ago
- Auto Data is a library designed for quick and effortless creation of datasets tailored for fine-tuning Large Language Models (LLMs).☆102Updated 9 months ago
- A proxy server for multiple ollama instances with Key security☆470Updated this week
- 🗣️ Real‑time, low‑latency voice, vision, and conversational‑memory AI assistant built on LiveKit and local LLMs ✨☆77Updated last month
- Distributed Inference for mlx LLm☆94Updated last year
- Comparison of the output quality of quantization methods, using Llama 3, transformers, GGUF, EXL2.☆158Updated last year
- This extension enhances the capabilities of textgen-webui by integrating advanced vision models, allowing users to have contextualized co…☆57Updated 9 months ago
- Building open version of OpenAI o1 via reasoning traces (Groq, ollama, Anthropic, Gemini, OpenAI, Azure supported) Demo: https://hugging…☆183Updated 9 months ago
- ☆207Updated 2 weeks ago
- ☆117Updated 8 months ago
- Mycomind Daemon: A mycelium-inspired, advanced Mixture-of-Memory-RAG-Agents (MoMRA) cognitive assistant that combines multiple AI models …☆35Updated last year
- A python package for developing AI applications with local LLMs.☆151Updated 7 months ago
- A pipeline parallel training script for LLMs.☆153Updated 3 months ago
- One click templates for inferencing Language Models☆201Updated 2 weeks ago