bigscience-workshop / petalsLinks
πΈ Run LLMs at home, BitTorrent-style. Fine-tuning and inference up to 10x faster than offloading
β9,839Updated last year
Alternatives and similar repositories for petals
Users that are interested in petals are comparing it to the libraries listed below
Sorting:
- Running large language models on a single GPU for throughput-oriented scenarios.β9,379Updated last year
- LLMs build upon Evol Insturct: WizardLM, WizardCoder, WizardMathβ9,469Updated 5 months ago
- OpenLLaMA, a permissively licensed open source reproduction of Meta AIβs LLaMA 7B trained on the RedPajama datasetβ7,528Updated 2 years ago
- Large Language Model Text Generation Inferenceβ10,684Updated 2 weeks ago
- StableLM: Stability AI Language Modelsβ15,787Updated last year
- Implementation of the LLaMA language model based on nanoGPT. Supports flash attention, Int8 and GPTQ 4bit quantization, LoRA and LLaMA-Adβ¦β6,087Updated 5 months ago
- Tensor library for machine learningβ13,648Updated last week
- Locally run an Instruction-Tuned Chat-Style LLMβ10,197Updated 2 years ago
- Instruct-tune LLaMA on consumer hardwareβ18,983Updated last year
- Build, personalize and control your own LLMs. From data pre-processing to fine-tuning, xTuring provides an easy way to personalize open-sβ¦β2,664Updated last week
- Gorilla: Training and Evaluating LLMs for Function Calls (Tool Calls)β12,578Updated 2 weeks ago
- Python bindings for llama.cppβ9,786Updated 3 months ago
- QLoRA: Efficient Finetuning of Quantized LLMsβ10,778Updated last year
- OpenAssistant is a chat-based assistant that understands tasks, can interact with third-party systems, and retrieve information dynamicalβ¦β37,501Updated last year
- Universal LLM Deployment Engine with ML Compilationβ21,691Updated last week
- An implementation of model parallel autoregressive transformers on GPUs, based on the Megatron and DeepSpeed librariesβ7,343Updated 2 months ago
- Run, manage, and scale AI workloads on any AI infrastructure. Use one system to access & manage all AI compute (Kubernetes, 20+ clouds, oβ¦β9,016Updated last week
- π A list of open LLMs available for commercial use.β12,527Updated 9 months ago
- The RedPajama-Data repository contains code for preparing large datasets for training large language models.β4,861Updated 11 months ago
- An LLM playground you can run on your laptopβ6,365Updated 2 months ago
- CodeGen is a family of open-source model for program synthesis. Trained on TPU-v4. Competitive with OpenAI Codex.β5,154Updated last month
- Run any open-source LLMs, such as DeepSeek and Llama, as OpenAI compatible API endpoint in the cloud.β11,965Updated this week
- A collection of libraries to optimise AI model performancesβ8,363Updated last year
- A language for constraint-guided and efficient LLM programming.β4,090Updated 6 months ago
- Home of StarCoder: fine-tuning & inference!β7,474Updated last year
- [ICLR 2024] Efficient Streaming Language Models with Attention Sinksβ7,135Updated last year
- Code and documentation to train Stanford's Alpaca models, and generate the data.β30,236Updated last year
- An open platform for training, serving, and evaluating large language models. Release repo for Vicuna and Chatbot Arena.β39,287Updated 6 months ago
- Simple UI for LLM Model Finetuningβ2,065Updated last year
- Semantic cache for LLMs. Fully integrated with LangChain and llama_index.β7,849Updated 4 months ago