xrsrke / instructGOOSE
Implementation of Reinforcement Learning from Human Feedback (RLHF)
☆169Updated last year
Related projects ⓘ
Alternatives and complementary repositories for instructGOOSE
- ☆94Updated last year
- ☆158Updated last year
- An experimental implementation of the retrieval-enhanced language model☆75Updated last year
- Chain-of-Hindsight, A Scalable RLHF Method☆220Updated last year
- Code accompanying the paper Pretraining Language Models with Human Preferences☆177Updated 9 months ago
- ☆221Updated last year
- Tk-Instruct is a Transformer model that is tuned to solve many NLP tasks by following instructions.☆177Updated 2 years ago
- A (somewhat) minimal library for finetuning language models with PPO on human feedback.☆86Updated last year
- This is the repo for the paper Shepherd -- A Critic for Language Model Generation☆213Updated last year
- ☆259Updated 11 months ago
- RLHF implementation details of OAI's 2019 codebase☆152Updated 10 months ago
- ☆175Updated last year
- Implementation of Toolformer: Language Models Can Teach Themselves to Use Tools☆136Updated last year
- A simulation framework for RLHF and alternatives. Develop your RLHF method without collecting human data.☆782Updated 4 months ago
- Reverse Instructions to generate instruction tuning data with corpus examples☆206Updated 8 months ago
- Open Instruction Generalist is an assistant trained on massive synthetic instructions to perform many millions of tasks☆206Updated 10 months ago
- All available datasets for Instruction Tuning of Large Language Models☆237Updated 11 months ago
- Learning to Compress Prompts with Gist Tokens - https://arxiv.org/abs/2304.08467☆266Updated last year
- ☆263Updated last year
- Official repository of NEFTune: Noisy Embeddings Improves Instruction Finetuning☆384Updated 6 months ago
- Code and model release for the paper "Task-aware Retrieval with Instructions" by Asai et al.☆160Updated last year
- A minimum example of aligning language models with RLHF similar to ChatGPT☆214Updated last year
- Self-Alignment with Principle-Following Reward Models☆148Updated 8 months ago
- Scaling Data-Constrained Language Models☆321Updated last month
- Implementation of ChatGPT RLHF (Reinforcement Learning with Human Feedback) on any generation model in huggingface's transformer (blommz-…☆543Updated 6 months ago
- [AAAI 2024] Investigating the Effectiveness of Task-Agnostic Prefix Prompt for Instruction Following☆79Updated 2 months ago
- A repository for transformer critique learning and generation☆86Updated 11 months ago
- LLaMA-TRL: Fine-tuning LLaMA with PPO and LoRA☆185Updated last year
- Official code from the paper "Offline RL for Natural Language Generation with Implicit Language Q Learning"☆199Updated last year
- Exploring finetuning public checkpoints on filter 8K sequences on Pile☆115Updated last year