Stability-AI / gpt-neoxLinks
An implementation of model parallel autoregressive transformers on GPUs, based on the DeepSpeed library.
☆13Updated 2 years ago
Alternatives and similar repositories for gpt-neox
Users that are interested in gpt-neox are comparing it to the libraries listed below
Sorting:
- Large Scale Distributed Model Training strategy with Colossal AI and Lightning AI☆56Updated 2 years ago
- Adversarial Training and SFT for Bot Safety Models☆40Updated 2 years ago
- Evaluation of bm42 sparse indexing algorithm☆72Updated last year
- ☆94Updated 3 years ago
- This repository contains code for cleaning your training data of benchmark data to help combat data snooping.☆27Updated 2 years ago
- Evaluating LLMs with CommonGen-Lite☆94Updated last year
- ☆128Updated 2 years ago
- QAmeleon introduces synthetic multilingual QA data using PaLM, a 540B large language model. This dataset was generated by prompt tuning P…☆35Updated 2 years ago
- The data processing pipeline for the Koala chatbot language model☆118Updated 2 years ago
- The collection of bulding blocks building fine-tunable metric learning models☆35Updated last month
- Source codes for the paper "Bounding the Capabilities of Large Language Models in Open Text Generation with Prompt Constraints"☆27Updated 3 years ago
- Hugging Face Inference Toolkit used to serve transformers, sentence-transformers, and diffusers models.☆90Updated last month
- ☆56Updated 7 months ago
- ☆16Updated last year
- Seahorse is a dataset for multilingual, multi-faceted summarization evaluation. It consists of 96K summaries with human ratings along 6 q…☆89Updated last year
- NeurIPS 2023 - Cappy: Outperforming and Boosting Large Multi-Task LMs with a Small Scorer☆46Updated last year
- RWKV is a RNN with transformer-level LLM performance. It can be directly trained like a GPT (parallelizable). So it's combining the best …☆10Updated 2 years ago
- Using open source LLMs to build synthetic datasets for direct preference optimization☆72Updated last year
- [EMNLP 2023 Industry Track] A simple prompting approach that enables the LLMs to run inference in batches.☆77Updated last year
- Small and Efficient Mathematical Reasoning LLMs☆73Updated 2 years ago
- C++ inference wrappers for running blazing fast embedding services on your favourite serverless like AWS Lambda. By Prithivi Da, PRs welc…☆23Updated last year
- Repo for "Smart Word Suggestions" (SWS) task and benchmark☆20Updated 2 years ago
- Safety Score for Pre-Trained Language Models☆96Updated 2 years ago
- A Multilingual Dataset for Parsing Realistic Task-Oriented Dialogs☆115Updated 2 years ago
- ☆53Updated 3 months ago
- Pre-train Static Word Embeddings☆94Updated 5 months ago
- minimal scripts for 24GB VRAM GPUs. training, inference, whatever☆50Updated last month
- A Python wrapper around HuggingFace's TGI (text-generation-inference) and TEI (text-embedding-inference) servers.☆32Updated 4 months ago
- An Implementation of "Orca: Progressive Learning from Complex Explanation Traces of GPT-4"☆43Updated last year
- Lightweight demos for finetuning LLMs. Powered by 🤗 transformers and open-source datasets.☆77Updated last year