lastmile-ai / llama-retrieval-pluginLinks
LLaMa retrieval plugin script using OpenAI's retrieval plugin
β323Updated 2 years ago
Alternatives and similar repositories for llama-retrieval-plugin
Users that are interested in llama-retrieval-plugin are comparing it to the libraries listed below
Sorting:
- β457Updated last year
- C++ implementation for π«StarCoderβ452Updated last year
- Extend the original llama.cpp repo to support redpajama model.β117Updated 9 months ago
- Tune any FALCON in 4-bitβ466Updated last year
- Landmark Attention: Random-Access Infinite Context Length for Transformers QLoRAβ122Updated last year
- β405Updated 2 years ago
- This project is an attempt to create a common metric to test LLM's for progress in eliminating hallucinations which is the most serious cβ¦β222Updated 2 years ago
- OpenAlpaca: A Fully Open-Source Instruction-Following Model Based On OpenLLaMAβ302Updated last year
- A school for camelidsβ1,208Updated 2 years ago
- β534Updated last year
- Run Alpaca LLM in LangChainβ217Updated last year
- Load local LLMs effortlessly in a Jupyter notebook for testing purposes alongside Langchain or other agents. Contains Oobagooga and Kobolβ¦β212Updated last year
- β275Updated 2 years ago
- Reimplementation of the task generation part from the Alpaca paperβ119Updated 2 years ago
- Complex LLM Workflows from Simple JSON.β301Updated last year
- Instruct-tuning LLaMA on consumer hardwareβ65Updated 2 years ago
- β412Updated last year
- Falcon LLM ggml framework with CPU and GPU supportβ245Updated last year
- Command-line script for inferencing from models such as MPT-7B-Chatβ101Updated last year
- This repository contains code for extending the Stanford Alpaca synthetic instruction tuning to existing instruction-tuned models such asβ¦β352Updated last year
- Simple, hackable and fast implementation for training/finetuning medium-sized LLaMA-based modelsβ171Updated last month
- SoTA Transformers with C-backend for fast inference on your CPU.β308Updated last year
- React app implementing OpenAI and Google APIs to re-create behavior of the toolformer paper.β233Updated 2 years ago
- Instruct-tune LLaMA on consumer hardwareβ362Updated 2 years ago
- 4 bits quantization of SantaCoder using GPTQβ50Updated last year
- β269Updated 2 years ago
- C++ implementation for BLOOMβ809Updated 2 years ago
- A joint community effort to create one central leaderboard for LLMs.β299Updated 9 months ago
- fastLLaMa: An experimental high-performance framework for running Decoder-only LLMs with 4-bit quantization in Python using a C/C++ backeβ¦β408Updated 2 years ago
- Harnessing the Memory Power of the Camelidsβ146Updated last year