liltom-eth / llama2-webui
Run any Llama 2 locally with gradio UI on GPU or CPU from anywhere (Linux/Windows/Mac). Use `llama2-wrapper` as your local llama2 backend for Generative Agents/Apps.
☆1,962Updated 10 months ago
Alternatives and similar repositories for llama2-webui:
Users that are interested in llama2-webui are comparing it to the libraries listed below
- LLaMA v2 Chatbot☆1,400Updated last year
- Python bindings for the Transformer models implemented in C/C++ using GGML library.☆1,830Updated last year
- Running Llama 2 and other Open-Source LLMs on CPU Inference Locally for Document Q&A☆958Updated last year
- A more memory-efficient rewrite of the HF transformers implementation of Llama for use with quantized weights.☆2,806Updated last year
- LLM as a Chatbot Service☆3,297Updated last year
- A fast inference library for running LLMs locally on modern consumer-class GPUs☆3,893Updated this week
- 4 bits quantization of LLaMA using GPTQ☆3,032Updated 6 months ago
- An easy-to-use LLMs quantization package with user-friendly apis, based on GPTQ algorithm.☆4,650Updated last week
- Large language models (LLMs) made easy, EasyLM is a one stop solution for pre-training, finetuning, evaluating and serving LLMs in JAX/Fl…