abacaj / mpt-30B-inference
Run inference on MPT-30B using CPU
☆575Updated last year
Alternatives and similar repositories for mpt-30B-inference:
Users that are interested in mpt-30B-inference are comparing it to the libraries listed below
- C++ implementation for 💫StarCoder☆454Updated last year
- An Autonomous LLM Agent that runs on Wizcoder-15B☆335Updated 6 months ago
- Salesforce open-source LLMs with 8k sequence length.☆717Updated 3 months ago
- Customizable implementation of the self-instruct paper.☆1,044Updated last year
- ☆535Updated last year
- ☆592Updated last year
- A school for camelids☆1,209Updated 2 years ago
- LLM that combines the principles of wizardLM and vicunaLM☆717Updated last year
- ☆277Updated last year
- LaMini-LM: A Diverse Herd of Distilled Models from Large-Scale Instructions☆820Updated 2 years ago
- LongLLaMA is a large language model capable of handling long contexts. It is based on OpenLLaMA and fine-tuned with the Focused Transform…☆1,452Updated last year
- C++ implementation for BLOOM☆811Updated last year
- Landmark Attention: Random-Access Infinite Context Length for Transformers QLoRA☆123Updated last year
- Evaluation tool for LLM QA chains☆1,075Updated last year
- A collection of modular datasets generated by GPT-4, General-Instruct - Roleplay-Instruct - Code-Instruct - and Toolformer☆1,630Updated last year
- UI tool for fine-tuning and testing your own LoRA models base on LLaMA, GPT-J and more. One-click run on Google Colab. + A Gradio ChatGPT…☆469Updated last year
- fastLLaMa: An experimental high-performance framework for running Decoder-only LLMs with 4-bit quantization in Python using a C/C++ backe…☆410Updated last year
- Chain together LLMs for reasoning & orchestrate multiple large models for accomplishing complex tasks☆604Updated 2 years ago
- This repository contains code and tooling for the Abacus.AI LLM Context Expansion project. Also included are evaluation scripts and bench…☆586Updated last year
- The code we currently use to fine-tune models.☆114Updated last year
- Tune any FALCON in 4-bit☆466Updated last year
- Directly Connecting Python to LLMs via Strongly-Typed Functions, Dataclasses, Interfaces & Generic Types☆397Updated 2 months ago
- Falcon LLM ggml framework with CPU and GPU support☆246Updated last year
- LLaMa retrieval plugin script using OpenAI's retrieval plugin☆324Updated 2 years ago
- ☆1,029Updated last year
- Fine-tune mistral-7B on 3090s, a100s, h100s☆711Updated last year
- SoTA Transformers with C-backend for fast inference on your CPU.☆310Updated last year
- Large Language Models for All, 🦙 Cult and More, Stay in touch !☆446Updated last year
- OpenAI-compatible Python client that can call any LLM☆371Updated last year
- ggml implementation of BERT☆488Updated last year