abacaj / mpt-30B-inferenceLinks
Run inference on MPT-30B using CPU
β576Updated 2 years ago
Alternatives and similar repositories for mpt-30B-inference
Users that are interested in mpt-30B-inference are comparing it to the libraries listed below
Sorting:
- C++ implementation for BLOOMβ808Updated 2 years ago
- C++ implementation for π«StarCoderβ457Updated 2 years ago
- A school for camelidsβ1,209Updated 2 years ago
- β599Updated 2 years ago
- LaMini-LM: A Diverse Herd of Distilled Models from Large-Scale Instructionsβ823Updated 2 years ago
- Evaluation tool for LLM QA chainsβ1,087Updated 2 years ago
- Salesforce open-source LLMs with 8k sequence length.β722Updated 10 months ago
- kani (γ«γ) is a highly hackable microframework for tool-calling language models. (NLP-OSS @ EMNLP 2023)β594Updated 3 weeks ago
- Scale LLM Engine public repositoryβ816Updated this week
- LLaMa retrieval plugin script using OpenAI's retrieval pluginβ324Updated 2 years ago
- β276Updated 2 years ago
- Directly Connecting Python to LLMs via Strongly-Typed Functions, Dataclasses, Interfaces & Generic Typesβ401Updated 9 months ago
- OpenAI-compatible Python client that can call any LLMβ373Updated 2 years ago
- Chain together LLMs for reasoning & orchestrate multiple large models for accomplishing complex tasksβ607Updated 2 years ago
- Locally hosted tool that connects documents to LLMs for summarization and querying, with a simple GUI.β799Updated 2 years ago
- A voice chat appβ1,178Updated 6 months ago
- A tiny implementation of an autonomous agent powered by LLMs (OpenAI GPT-4)β439Updated 2 years ago
- An Autonomous LLM Agent that runs on Wizcoder-15Bβ334Updated last year
- Large Language Models for All, π¦ Cult and More, Stay in touch !β450Updated 2 years ago
- A collection of modular datasets generated by GPT-4, General-Instruct - Roleplay-Instruct - Code-Instruct - and Toolformerβ1,631Updated 2 years ago
- This repository contains code and tooling for the Abacus.AI LLM Context Expansion project. Also included are evaluation scripts and benchβ¦β598Updated 2 years ago
- fastLLaMa: An experimental high-performance framework for running Decoder-only LLMs with 4-bit quantization in Python using a C/C++ backeβ¦β412Updated 2 years ago
- Tune any FALCON in 4-bitβ465Updated 2 years ago
- [NeurIPS 22] [AAAI 24] Recurrent Transformer-based long-context architecture.β775Updated last year
- β535Updated 2 years ago
- howdoi.aiβ257Updated 2 years ago
- Use GPT4 and GPT3.5 on inputs of unlimited size. Uses multithreading to process multiple chunks in parallel. Useful for tasks like Named β¦β269Updated 2 years ago
- Finetuning Large Language Models on One Consumer GPU in 2 Bitsβ733Updated last year
- LLM-based tool for parsing information and chatting with itβ214Updated 2 years ago
- Build robust LLM applications with true composability πβ422Updated last year