bigscience-workshop / petalsLinks
πΈ Run LLMs at home, BitTorrent-style. Fine-tuning and inference up to 10x faster than offloading
β9,915Updated last year
Alternatives and similar repositories for petals
Users that are interested in petals are comparing it to the libraries listed below
Sorting:
- Instruct-tune LLaMA on consumer hardwareβ18,979Updated last year
- The RedPajama-Data repository contains code for preparing large datasets for training large language models.β4,923Updated last year
- LLMs build upon Evol Insturct: WizardLM, WizardCoder, WizardMathβ9,475Updated 8 months ago
- Implementation of the LLaMA language model based on nanoGPT. Supports flash attention, Int8 and GPTQ 4bit quantization, LoRA and LLaMA-Adβ¦β6,092Updated 7 months ago
- OpenLLaMA, a permissively licensed open source reproduction of Meta AIβs LLaMA 7B trained on the RedPajama datasetβ7,530Updated 2 years ago
- Running large language models on a single GPU for throughput-oriented scenarios.β9,383Updated last year
- Build, personalize and control your own LLMs. From data pre-processing to fine-tuning, xTuring provides an easy way to personalize open-sβ¦β2,664Updated this week
- QLoRA: Efficient Finetuning of Quantized LLMsβ10,830Updated last year
- Tensor library for machine learningβ13,907Updated last week
- Large Language Model Text Generation Inferenceβ10,749Updated 3 weeks ago
- The simplest way to run LLaMA on your local machineβ12,997Updated last year
- An LLM playground you can run on your laptopβ6,370Updated last week
- Universal LLM Deployment Engine with ML Compilationβ21,981Updated last week
- An implementation of model parallel autoregressive transformers on GPUs, based on the Megatron and DeepSpeed librariesβ7,374Updated last month
- Home of StarCoder: fine-tuning & inference!β7,529Updated last year
- Run, manage, and scale AI workloads on any AI infrastructure. Use one system to access & manage all AI compute (Kubernetes, 20+ clouds, oβ¦β9,418Updated this week
- Locally run an Instruction-Tuned Chat-Style LLMβ10,187Updated 2 years ago
- StableLM: Stability AI Language Modelsβ15,768Updated last year
- OpenAssistant is a chat-based assistant that understands tasks, can interact with third-party systems, and retrieve information dynamicalβ¦β37,466Updated last year
- A collection of libraries to optimise AI model performancesβ8,354Updated last year
- Simple UI for LLM Model Finetuningβ2,063Updated 2 years ago
- An open platform for training, serving, and evaluating large language models. Release repo for Vicuna and Chatbot Arena.β39,392Updated 8 months ago
- Implementation of RLHF (Reinforcement Learning with Human Feedback) on top of the PaLM architecture. Basically ChatGPT but with PaLMβ7,882Updated 3 months ago
- LLM as a Chatbot Serviceβ3,331Updated 2 years ago
- RWKV (pronounced RwaKuv) is an RNN with great LLM performance, which can also be directly trained like a GPT transformer (parallelizable)β¦β14,346Updated this week
- An unnecessarily tiny implementation of GPT-2 in NumPy.β3,439Updated 2 years ago
- Turbopilot is an open source large-language-model based code completion engine that runs locally on CPUβ3,806Updated 2 years ago
- Gorilla: Training and Evaluating LLMs for Function Calls (Tool Calls)β12,710Updated 2 weeks ago
- A more memory-efficient rewrite of the HF transformers implementation of Llama for use with quantized weights.β2,908Updated 2 years ago
- β22,113Updated last week