intel-staging / text-generation-webuiLinks
A Gradio Web UI for running local LLM on Intel GPU (e.g., local PC with iGPU, discrete GPU such as Arc, Flex and Max) using IPEX-LLM.
☆17Updated last week
Alternatives and similar repositories for text-generation-webui
Users that are interested in text-generation-webui are comparing it to the libraries listed below
Sorting:
- EdgeInfer enables efficient edge intelligence by running small AI models, including embeddings and OnnxModels, on resource-constrained de…☆49Updated last year
- run chatglm3-6b in BM1684X☆40Updated last year
- The inference code of RVC-Boss/GPT-SoVITS that can be developer-friendly.☆15Updated last year
- Enable AI to quickly retrieve content, simply!☆53Updated last year
- This project support a WEB UI with Vicuna13B (using llama-cpp-python, chatbot-ui)☆46Updated 2 years ago
- Convert different model APIs into the OpenAI API format out of the box.☆160Updated last year
- Developer kits reference setup scripts for various kinds of Intel platforms and GPUs☆38Updated last week
- LLM Client with Function Calling | LLM客户端,支持联网,IO操作和执行代码☆68Updated 5 months ago
- ☆33Updated last year
- Instruct-tune LLaMA on consumer hardware with shareGPT data☆126Updated 2 years ago
- This is a web service built using FastAPI and OpenAI's GPT-3.5-Turbo model that enables translation of various types of documents. The se…☆135Updated 2 years ago
- A voice chatbot based on GPT4All and talkGPT, running on your local pc!☆152Updated last year
- OVALChat is a customizable Web app aimed at conducting user studies with chatbots☆28Updated last year
- 封装KimiAI的API,快速调用AI回复☆11Updated last year
- xllamacpp - a Python wrapper of llama.cpp☆62Updated last week
- 适用于sophon bm1684x,基于 Langchain 与 ChatGLM 等语言模型的本地知识库问答☆14Updated last year
- < 10MB, One-click self-hosted ChatGPT, allowing access to various data sources and non-OpenAI models.☆141Updated 2 years ago
- LocalAGI:Locally run AGI powered by LLaMA, ChatGLM and more. | 基于 ChatGLM, LLaMA 大模型的本地运行的 AGI☆81Updated 2 years ago
- SiliconCloud Cookbook☆22Updated 8 months ago
- LLM inference in C/C++☆103Updated last week
- instinct.cpp provides ready to use alternatives to OpenAI Assistant API and built-in utilities for developing AI Agent applications (RAG,…☆53Updated last year
- 简单、靠谱的 SDXL Docker 使用方案。☆41Updated last year
- One-click Docker APPs of Open-source Code Interpreter Projects.☆61Updated 2 years ago
- CodeShell model in C/C++☆105Updated last year
- llama.cpp fork used by GPT4All☆57Updated 8 months ago
- ChatGPT Style client-compatible Backend Server, open source implementation.☆95Updated last year
- My develoopment fork of llama.cpp. For now working on RK3588 NPU and Tenstorrent backend☆107Updated last week
- Auto Thinking Mode switch for Qwen3 in Open webui☆68Updated 6 months ago
- LM inference server implementation based on *.cpp.☆289Updated 2 months ago
- llama.cpp with unicode (windows) support☆54Updated 2 years ago