WapaMario63 / GPTQ-for-LLaMa-ROCmLinks
4 bits quantization of LLaMA using GPTQ, ported to HIP for use in AMD GPUs.
☆32Updated 2 years ago
Alternatives and similar repositories for GPTQ-for-LLaMa-ROCm
Users that are interested in GPTQ-for-LLaMa-ROCm are comparing it to the libraries listed below
Sorting:
- DEPRECATED!☆50Updated last year
- Web UI for ExLlamaV2☆514Updated 9 months ago
- A manual for helping using tesla p40 gpu☆137Updated last year
- A prompt/context management system☆170Updated 2 years ago
- 8-bit CUDA functions for PyTorch Rocm compatible☆41Updated last year
- A fork of textgen that kept some things like Exllama and old GPTQ.☆22Updated last year
- Falcon LLM ggml framework with CPU and GPU support☆247Updated last year
- A simple Gradio WebUI for loading/unloading models and loras in tabbyAPI.☆20Updated last year
- Automated prompting and scoring framework to evaluate LLMs using updated human knowledge prompts☆109Updated 2 years ago
- AMD (Radeon GPU) ROCm based setup for popular AI tools on Ubuntu 24.04.1☆216Updated 2 weeks ago
- An extension for oobabooga's text-generation-webui that adds syntax highlighting to code snippets☆66Updated last year
- ☆156Updated 2 years ago
- Landmark Attention: Random-Access Infinite Context Length for Transformers QLoRA☆123Updated 2 years ago
- ☆534Updated last year
- Comparison of the output quality of quantization methods, using Llama 3, transformers, GGUF, EXL2.☆165Updated last year
- A more memory-efficient rewrite of the HF transformers implementation of Llama for use with quantized weights.☆63Updated 2 years ago
- A gradio web UI for running Large Language Models like GPT-J 6B, OPT, GALACTICA, LLaMA, and Pygmalion.☆308Updated 2 years ago
- Prompt Jinja2 templates for LLMs☆34Updated 4 months ago
- ☆37Updated 2 years ago
- An extension for oobabooga/text-generation-webui that enables the LLM to search the web☆268Updated this week
- Lord of LLMS☆294Updated last month
- A community list of common phrases generated by GPT and Claude models☆79Updated 2 years ago
- A simple webui for stable-diffusion.cpp☆51Updated this week
- TheBloke's Dockerfiles☆307Updated last year
- Download models from the Ollama library, without Ollama☆109Updated last year
- LLaMA Server combines the power of LLaMA C++ with the beauty of Chatbot UI.☆129Updated 2 years ago
- A multimodal, function calling powered LLM webui.☆216Updated last year
- Experimental LLM Inference UX to aid in creative writing☆125Updated 11 months ago
- ☆50Updated 2 years ago
- BabyAGI to run with locally hosted models using the API from https://github.com/oobabooga/text-generation-webui☆87Updated 2 years ago