intel-staging / text-generation-webuiLinks
A Gradio Web UI for running local LLM on Intel GPU (e.g., local PC with iGPU, discrete GPU such as Arc, Flex and Max) using IPEX-LLM.
☆18Updated this week
Alternatives and similar repositories for text-generation-webui
Users that are interested in text-generation-webui are comparing it to the libraries listed below
Sorting:
- EdgeInfer enables efficient edge intelligence by running small AI models, including embeddings and OnnxModels, on resource-constrained de…☆45Updated last year
- Creating Interactive and Embedded Physics Simulations from Static Textbook Diagrams☆21Updated 4 months ago
- LLM inference in C/C++☆97Updated last week
- xllamacpp - a Python wrapper of llama.cpp☆46Updated this week
- a useful PDF Translate tool base on LLM/ 一个基于大语言模型的PDF翻译程序☆69Updated 11 months ago
- ☆58Updated this week
- 简单、靠谱的 SDXL Docker 使用方案。☆42Updated last year
- instinct.cpp provides ready to use alternatives to OpenAI Assistant API and built-in utilities for developing AI Agent applications (RAG,…☆52Updated last year
- CodeShell model in C/C++☆106Updated last year
- 封装KimiAI的API,快速调用AI回复☆11Updated last year
- Tech notes for mkdocs and gitbook☆18Updated 8 months ago
- Convert different model APIs into the OpenAI API format out of the box.☆157Updated last year
- Cross Platform Open Sourced Chinese NoteBookLM app based on Electron, Use DeepSeek + Reecho.ai☆74Updated 8 months ago
- AMD related optimizations for transformer models☆80Updated 3 weeks ago
- Developer kits reference setup scripts for various kinds of Intel platforms and GPUs☆31Updated this week
- A converter and basic tester for rwkv onnx☆42Updated last year
- ☆22Updated 2 months ago
- Nougat is a Meta AI's revolutionary OCR model designed to transcribe scientific PDFs into an easy-to-use Markdown format.☆24Updated last year
- Self-host llmapi server, make it really easy for accessing LLMs !☆37Updated 2 years ago
- 类似于锤子大爆炸的Linux OCR实现☆20Updated 3 years ago
- Instruct-tune LLaMA on consumer hardware with shareGPT data☆126Updated 2 years ago
- Cortex.Tensorrt-LLM is a C++ inference library that can be loaded by any server at runtime. It submodules NVIDIA’s TensorRT-LLM for GPU a…☆43Updated 9 months ago
- A single-file tkinter-based Ollama GUI project with no external dependencies.☆207Updated 4 months ago
- OVALChat is a customizable Web app aimed at conducting user studies with chatbots☆28Updated last year
- A voice chatbot based on GPT4All and talkGPT, running on your local pc!☆153Updated last year
- LM inference server implementation based on *.cpp.☆239Updated this week
- One-click Docker APPs of Open-source Code Interpreter Projects.☆60Updated 2 years ago
- ☆149Updated last year
- Get up and running with Llama 3, Mistral, Gemma, and other large language models.☆27Updated last week
- 模拟浏览器脚本操作,使用nodejs来批量读取和操作网盘文件信息。 这个代码库是`百度网盘批量清理重复文件计划`的一部分。☆11Updated 2 years ago