Highlyhotgames / fast_txtgen_7B
Installation Script for LLaMa 7B 4bit 128g on WSL
☆28Updated 9 months ago
Alternatives and similar repositories for fast_txtgen_7B:
Users that are interested in fast_txtgen_7B are comparing it to the libraries listed below
- ☆16Updated last year
- Experimental sampler to make LLMs more creative☆30Updated last year
- ☆12Updated 9 months ago
- Creates an Langchain Agent which uses the WebUI's API and Wikipedia to work☆72Updated last year
- An auto generated wiki.☆21Updated last year
- An Extension for oobabooga/text-generation-webui☆36Updated last year
- A Qt GUI for large language models☆40Updated last year
- Just a simple HowTo for https://github.com/johnsmith0031/alpaca_lora_4bit☆31Updated last year
- A repository to store helpful information and emerging insights in regard to LLMs☆20Updated last year
- ☆39Updated 11 months ago
- Accepts a Hugging Face model URL, automatically downloads and quantizes it using Bits and Bytes.☆38Updated 10 months ago
- ☆27Updated last year
- Reweight GPT - a simple neural network using transformer architecture for next character prediction☆50Updated last year
- An extension to Oobabooga to add a simple memory function for chat☆23Updated last year
- The hearth of The Pulsar App, fast, secure and shared inference with modern UI☆43Updated last month
- oobaboga -text-generation-webui implementation of wafflecomposite - langchain-ask-pdf-local☆68Updated last year
- Genertaes control vectors for use with llama.cpp in GGUF format.☆17Updated 4 months ago
- These agents work based on any local model. You ask your question and simply indicate the number of agents and experts who will answer it…☆18Updated 10 months ago
- ☆40Updated 9 months ago
- Merge LLM that are split in to parts☆25Updated last year
- run ollama & gguf easily with a single command☆49Updated 8 months ago
- Small repository for my video on LoRA☆16Updated last year
- ☆29Updated last year
- Prompt-Promptor is a python library for automatically generating prompts using LLMs☆72Updated last year
- 4bit bitsandbytes quants of the best 7B vlms☆25Updated 3 months ago
- Host the GPTQ model using AutoGPTQ as an API that is compatible with text generation UI API.☆91Updated last year
- ☆48Updated last year
- Simple and fast server for GPTQ-quantized LLaMA inference☆24Updated last year