openai / following-instructions-human-feedback
☆1,178Updated last year
Related projects ⓘ
Alternatives and complementary repositories for following-instructions-human-feedback
- Human preference data for "Training a Helpful and Harmless Assistant with Reinforcement Learning from Human Feedback"☆1,625Updated last year
- Code for the paper Fine-Tuning Language Models from Human Preferences☆1,232Updated last year
- Code for "Learning to summarize from human feedback"☆992Updated last year
- ☆1,475Updated 3 weeks ago
- Beyond the Imitation Game collaborative benchmark for measuring and extrapolating the capabilities of language models☆2,875Updated 4 months ago
- ☆1,080Updated 10 months ago
- Implementation of Toolformer, Language Models That Can Use Tools, by MetaAI☆1,971Updated 4 months ago
- Dromedary: towards helpful, ethical and reliable LLMs.☆1,128Updated last year
- Toolkit for creating, sharing and using natural language prompts.☆2,700Updated last year
- A repo for distributed training of language models with Reinforcement Learning via Human Feedback (RLHF)☆4,509Updated 10 months ago
- Expanding natural instructions☆960Updated 11 months ago
- Implementation of the specific Transformer architecture from PaLM - Scaling Language Modeling with Pathways☆821Updated 2 years ago
- The hub for EleutherAI's work on interpretability and learning dynamics☆2,285Updated 3 weeks ago
- A simulation framework for RLHF and alternatives. Develop your RLHF method without collecting human data.☆783Updated 4 months ago
- AutoPrompt: Automatic Prompt Construction for Masked Language Models.☆595Updated 3 months ago
- [NeurIPS 22] [AAAI 24] Recurrent Transformer-based long-context architecture.☆755Updated 3 weeks ago
- ☆710Updated 5 months ago
- A modular RL library to fine-tune language models to human preferences☆2,218Updated 8 months ago
- Reproduce results and replicate training fo T0 (Multitask Prompted Training Enables Zero-Shot Task Generalization)☆457Updated 2 years ago
- Benchmarking large language models' complex reasoning ability with chain-of-thought prompting☆2,579Updated 3 months ago
- ☆1,507Updated last year
- A prize for finding tasks that cause large language models to show inverse scaling☆597Updated last year
- Ongoing research training transformer language models at scale, including: BERT & GPT-2☆1,338Updated 8 months ago
- TruthfulQA: Measuring How Models Imitate Human Falsehoods☆621Updated last year
- PaL: Program-Aided Language Models (ICML 2023)☆474Updated last year
- A central, open resource for data and tools related to chain-of-thought reasoning in large language models. Developed @ Samwald research …☆895Updated 5 months ago
- Original Implementation of Prompt Tuning from Lester, et al, 2021☆657Updated 6 months ago
- 800,000 step-level correctness labels on LLM solutions to MATH problems☆1,669Updated last year
- Data and code for NeurIPS 2022 Paper "Learn to Explain: Multimodal Reasoning via Thought Chains for Science Question Answering".☆606Updated 2 months ago
- Implementation of RETRO, Deepmind's Retrieval based Attention net, in Pytorch☆852Updated last year