ItsPi3141 / alpaca.cppLinks
Locally run an Instruction-Tuned Chat-Style LLM
☆30Updated 2 years ago
Alternatives and similar repositories for alpaca.cpp
Users that are interested in alpaca.cpp are comparing it to the libraries listed below
Sorting:
- OpenAI API webserver☆189Updated 4 years ago
- Falcon LLM ggml framework with CPU and GPU support☆249Updated 2 years ago
- LLaMa retrieval plugin script using OpenAI's retrieval plugin☆323Updated 2 years ago
- Supercharge Open-Source AI Models☆349Updated 2 years ago
- Landmark Attention: Random-Access Infinite Context Length for Transformers QLoRA☆124Updated 2 years ago
- LLM powered development for IntelliJ☆84Updated last year
- 💬 Chatbot web app + HTTP and Websocket endpoints for LLM inference with the Petals client☆316Updated last year
- Lossless normalization of uppercase characters☆11Updated 2 years ago
- Harnessing the Memory Power of the Camelids☆147Updated 2 years ago
- Python bindings for llama.cpp☆198Updated 2 years ago
- Inference code for LLaMA models☆189Updated 2 years ago
- An easy way to host your own AI API and expose alternative models, while being compatible with "open" AI clients.☆332Updated last year
- An OpenAI-like LLaMA inference API☆113Updated 2 years ago
- Instruct-tuning LLaMA on consumer hardware☆66Updated 2 years ago
- Visual Studio Code extension for WizardCoder☆148Updated 2 years ago
- Locally run an Instruction-Tuned Chat-Style LLM (Android/Linux/Windows/Mac)☆262Updated 2 years ago
- Tune any FALCON in 4-bit☆463Updated 2 years ago
- LLM-based code completion engine☆190Updated last year
- Client-side toolkit for using large language models, including where self-hosted☆115Updated this week
- Unofficial python bindings for the rust llm library. 🐍❤️🦀☆76Updated 2 years ago
- Locally run an Assistant-Tuned Chat-Style LLM☆496Updated 2 years ago
- TheBloke's Dockerfiles☆308Updated last year
- Framework agnostic python runtime for RWKV models☆147Updated 2 years ago
- The code we currently use to fine-tune models.☆117Updated last year
- GPT-2 small trained on phi-like data☆68Updated last year
- An Autonomous LLM Agent that runs on Wizcoder-15B☆333Updated last year
- fastLLaMa: An experimental high-performance framework for running Decoder-only LLMs with 4-bit quantization in Python using a C/C++ backe…☆412Updated 2 years ago
- A torchless, c++ rwkv implementation using 8bit quantization, written in cuda/hip/vulkan for maximum compatibility and minimum dependenci…☆313Updated 2 years ago
- A OpenAI API compatible REST server for llama.☆208Updated 11 months ago
- Patch for MPT-7B which allows using and training a LoRA☆58Updated 2 years ago