intel-staging / text-generation-webuiLinks
A Gradio Web UI for running local LLM on Intel GPU (e.g., local PC with iGPU, discrete GPU such as Arc, Flex and Max) using IPEX-LLM.
☆18Updated this week
Alternatives and similar repositories for text-generation-webui
Users that are interested in text-generation-webui are comparing it to the libraries listed below
Sorting:
- Knowledge Base QA using RAG pipeline on Intel CPU and GPU (e.g., local PC with iGPU, discrete GPU such as Arc, Flex and Max) with IPEX-LL…☆16Updated last month
- Developer kits reference setup scripts for various kinds of Intel platforms and GPUs☆28Updated this week
- EdgeInfer enables efficient edge intelligence by running small AI models, including embeddings and OnnxModels, on resource-constrained de…☆45Updated last year
- ☆110Updated last month
- This repository contains Dockerfiles, scripts, yaml files, Helm charts, etc. used to scale out AI containers with versions of TensorFlow …☆45Updated this week
- Intel® Extension for DeepSpeed* is an extension to DeepSpeed that brings feature support with SYCL kernels on Intel GPU(XPU) device. Note…☆61Updated 3 months ago
- Accelerate LLM with low-bit (FP4 / INT4 / FP8 / INT8) optimizations using ipex-llm☆164Updated last month
- OpenVINO LLM Benchmark☆11Updated last year
- ☆57Updated 3 weeks ago
- OVALChat is a customizable Web app aimed at conducting user studies with chatbots☆28Updated last year
- 适用于sophon bm1684x,基于 Langchain 与 ChatGLM 等语言模型的本地知识库问答☆12Updated last year
- ☆52Updated last year
- My develoopment fork of llama.cpp. For now working on RK3588 NPU and Tenstorrent backend☆94Updated last week
- ☆21Updated last year
- GLM-4 series: Open Multilingual Multimodal Chat LMs | 开源多语言多模态对话模型☆26Updated last month
- UnitEval is a benchmarking and evaluation tools for AutoDev Coder.☆12Updated last year
- xllamacpp - a Python wrapper of llama.cpp☆40Updated this week
- Self-host llmapi server, make it really easy for accessing LLMs !☆37Updated 2 years ago
- Easy-GPT4O opensource version☆77Updated last year
- llm-inference is a platform for publishing and managing llm inference, providing a wide range of out-of-the-box features for model deploy…☆81Updated last year
- rwkv finetuning☆36Updated last year
- TPO 是一个优化 LLM 输出文本的框架,通过迭代反馈和优化提示的方式来“微调模型”,而非直接调整模型的参数,使模型在推理过程中与人类偏好对齐以生成更好的结果。本项目提供了一个友好的 WebUI 来加载模型,实时优化基础模型并展示最佳结果。☆10Updated 3 months ago
- Get up and running with Llama 3, Mistral, Gemma, and other large language models.☆26Updated 3 weeks ago
- SiliconCloud Cookbook☆22Updated 2 months ago
- Various LLM Benchmarks☆20Updated last week
- CodeShell model in C/C++☆106Updated 10 months ago
- A Web-UI for Llama_index. Allow ChatGPT to access your own database.☆22Updated 2 years ago
- A converter and basic tester for rwkv onnx☆41Updated last year
- run chatglm3-6b in BM1684X☆39Updated last year
- LocalAGI:Locally run AGI powered by LLaMA, ChatGLM and more. | 基于 ChatGLM, LLaMA 大模型的本地运行的 AGI☆76Updated last year