thekevinscott / vicuna-7b
Vicuna 7B is a large language model that runs in the browser. Exposes programmatic access with minimal configuration.
☆21Updated 2 years ago
Alternatives and similar repositories for vicuna-7b:
Users that are interested in vicuna-7b are comparing it to the libraries listed below
- Self-host LLMs with LMDeploy and BentoML☆18Updated last month
- HuggingChat like UI in Gradio☆72Updated last year
- A simple package for leveraging Falcon 180B and the HF ecosystem's tools, including training/inference scripts, safetensors, integrations…☆13Updated last year
- Advanced Ultra-Low Bitrate Compression Techniques for the LLaMA Family of LLMs☆111Updated last year
- Generate High Quality textual or multi-modal datasets with Agents☆18Updated last year
- BUD-E (Buddy) is an open-source voice assistant framework that facilitates seamless interaction with AI models and APIs, enabling the cre…☆19Updated 6 months ago
- Data preparation code for Amber 7B LLM☆88Updated 11 months ago
- Parameter-Efficient Sparsity Crafting From Dense to Mixture-of-Experts for Instruction Tuning on General Tasks☆31Updated 11 months ago
- Zephyr 7B beta RAG Demo inside a Gradio app powered by BGE Embeddings, ChromaDB, and Zephyr 7B Beta LLM.☆34Updated last year
- Proteus is an experimental platform that combines the power of Large Language Models with the Genesis physics engine☆21Updated 4 months ago
- XmodelLM☆39Updated 5 months ago
- AI Assistant running within your browser.☆62Updated 4 months ago
- ☆21Updated last year
- ☆37Updated 6 months ago
- Repo hosting codes and materials related to speeding LLMs' inference using token merging.☆36Updated 11 months ago
- Ultra Fast Multi-Modality Vector Database☆18Updated last year
- [ICLR 2024] Skeleton-of-Thought: Prompting LLMs for Efficient Parallel Generation☆166Updated last year
- vLLM: A high-throughput and memory-efficient inference and serving engine for LLMs☆86Updated this week
- ☆38Updated last year
- A toolkit for fine-tuning, inferencing, and evaluating GreenBitAI's LLMs.☆82Updated last month
- ☆20Updated last year
- Simple Autogpt with tree of thoughts☆15Updated last year
- Repository for CPU Kernel Generation for LLM Inference☆26Updated last year
- QuIP quantization☆51Updated last year
- ☆73Updated last year
- Modified Beam Search with periodical restart☆12Updated 7 months ago
- Some simple scripts that I use day-to-day when working with LLMs and Huggingface Hub☆160Updated last year
- Deploy your GGML models to HuggingFace Spaces with Docker and gradio☆36Updated last year
- Yet another LLM☆10Updated 2 years ago
- ☆46Updated 9 months ago