shawwn / openai-serverLinks
OpenAI API webserver
☆187Updated 3 years ago
Alternatives and similar repositories for openai-server
Users that are interested in openai-server are comparing it to the libraries listed below
Sorting:
- SoTA Transformers with C-backend for fast inference on your CPU.☆309Updated last year
- Landmark Attention: Random-Access Infinite Context Length for Transformers QLoRA☆122Updated last year
- A Simple Discord Bot for the Alpaca LLM☆101Updated last year
- 4 bits quantization of SantaCoder using GPTQ☆50Updated 2 years ago
- An easy way to host your own AI API and expose alternative models, while being compatible with "open" AI clients.☆330Updated 10 months ago
- LLaMa retrieval plugin script using OpenAI's retrieval plugin☆323Updated 2 years ago
- Reimplementation of the task generation part from the Alpaca paper☆119Updated 2 years ago
- ☆406Updated 2 years ago
- C++ implementation for BLOOM☆809Updated 2 years ago
- howdoi.ai☆255Updated 2 years ago
- Inference code for facebook LLaMA models with Wrapyfi support☆129Updated 2 years ago
- Command-line script for inferencing from models such as MPT-7B-Chat☆101Updated last year
- Prompt programming with FMs.☆442Updated 10 months ago
- fastLLaMa: An experimental high-performance framework for running Decoder-only LLMs with 4-bit quantization in Python using a C/C++ backe…☆408Updated 2 years ago
- 💬 Chatbot web app + HTTP and Websocket endpoints for LLM inference with the Petals client☆312Updated last year
- The GeoV model is a large langauge model designed by Georges Harik and uses Rotary Positional Embeddings with Relative distances (RoPER).…☆121Updated 2 years ago
- ☆175Updated 2 years ago
- Used for adaptive human in the loop evaluation of language and embedding models.☆308Updated 2 years ago
- A discord bot that roleplays!☆149Updated last year
- ☆275Updated 2 years ago
- Inference code for LLaMA models☆187Updated 2 years ago
- An OpenAI-like LLaMA inference API☆112Updated last year
- ☆130Updated 2 years ago
- Embeddings focused small version of Llama NLP model☆104Updated 2 years ago
- Directly Connecting Python to LLMs via Strongly-Typed Functions, Dataclasses, Interfaces & Generic Types☆399Updated 3 months ago
- A repository to run gpt-j-6b on low vram machines (4.2 gb minimum vram for 2000 token context, 3.5 gb for 1000 token context). Model load…☆115Updated 3 years ago
- A command-line interface to generate textual and conversational datasets with LLMs.☆295Updated last year
- An HTTP serving framework by Banana☆99Updated last year
- Smol but mighty language model☆61Updated 2 years ago
- ☆83Updated last year