ArmelRandy / Self-instructLinks
A repository to perform self-instruct with a model on HF Hub
☆32Updated 2 years ago
Alternatives and similar repositories for Self-instruct
Users that are interested in Self-instruct are comparing it to the libraries listed below
Sorting:
- The Official Repository for "Bring Your Own Data! Self-Supervised Evaluation for Large Language Models"☆107Updated 2 years ago
- A repository for transformer critique learning and generation☆89Updated 2 years ago
- Code and data accompanying our paper on arXiv "Faithful Chain-of-Thought Reasoning".☆165Updated last year
- Evaluating LLMs with fewer examples☆170Updated last year
- [ICLR 2024 Spotlight] FLASK: Fine-grained Language Model Evaluation based on Alignment Skill Sets☆218Updated 2 years ago
- [ICLR 2023] Guess the Instruction! Flipped Learning Makes Language Models Stronger Zero-Shot Learners☆116Updated 6 months ago
- evol augment any dataset online☆61Updated 2 years ago
- ☆38Updated last year
- Code repository for the c-BTM paper☆108Updated 2 years ago
- Scripts for generating synthetic finetuning data for reducing sycophancy.☆117Updated 2 years ago
- Code for paper "LEVER: Learning to Verifiy Language-to-Code Generation with Execution" (ICML'23)☆90Updated 2 years ago
- ☆180Updated 2 years ago
- ☆76Updated last year
- ☆34Updated 2 years ago
- A framework for few-shot evaluation of autoregressive language models.☆105Updated 2 years ago
- Scaling Data-Constrained Language Models☆343Updated 6 months ago
- ☆95Updated 2 years ago
- Simple next-token-prediction for RLHF☆227Updated 2 years ago
- Spherical Merge Pytorch/HF format Language Models with minimal feature loss.☆141Updated 2 years ago
- ☆44Updated last year
- Skill-It! A Data-Driven Skills Framework for Understanding and Training Language Models☆47Updated 2 years ago
- [Data + code] ExpertQA : Expert-Curated Questions and Attributed Answers☆136Updated last year
- Code accompanying the paper Pretraining Language Models with Human Preferences☆180Updated last year
- Code accompanying "How I learned to start worrying about prompt formatting".☆113Updated 6 months ago
- Multipack distributed sampler for fast padding-free training of LLMs☆202Updated last year
- This project studies the performance and robustness of language models and task-adaptation methods.☆155Updated last year
- TART: A plug-and-play Transformer module for task-agnostic reasoning☆202Updated 2 years ago
- [EMNLP 2023] The CoT Collection: Improving Zero-shot and Few-shot Learning of Language Models via Chain-of-Thought Fine-Tuning☆252Updated 2 years ago
- PASTA: Post-hoc Attention Steering for LLMs☆132Updated last year
- Multi-Domain Expert Learning☆67Updated last year