soochan-lee / RoTLinks
Official code for ACL 2023 (short, findings) paper "Recursion of Thought: A Divide and Conquer Approach to Multi-Context Reasoning with Language Models"
☆43Updated 2 years ago
Alternatives and similar repositories for RoT
Users that are interested in RoT are comparing it to the libraries listed below
Sorting:
- Small and Efficient Mathematical Reasoning LLMs☆71Updated last year
- [ICLR 2023] Guess the Instruction! Flipped Learning Makes Language Models Stronger Zero-Shot Learners☆116Updated last month
- ☆74Updated last year
- Official repo for NAACL 2024 Findings paper "LeTI: Learning to Generate from Textual Interactions."☆65Updated 2 years ago
- ☆125Updated 10 months ago
- Code repository for the c-BTM paper☆107Updated last year
- Repository for "I am a Strange Dataset: Metalinguistic Tests for Language Models"☆44Updated last year
- Patch for MPT-7B which allows using and training a LoRA☆58Updated 2 years ago
- QLoRA with Enhanced Multi GPU Support☆37Updated last year
- ☆57Updated 10 months ago
- Evaluating LLMs with CommonGen-Lite☆90Updated last year
- Mixing Language Models with Self-Verification and Meta-Verification☆105Updated 7 months ago
- An experiment to see if chatgpt can improve the output of the stanford alpaca dataset☆12Updated 2 years ago
- Spherical Merge Pytorch/HF format Language Models with minimal feature loss.☆135Updated last year
- QLoRA: Efficient Finetuning of Quantized LLMs☆78Updated last year
- Repository containing the SPIN experiments on the DIBT 10k ranked prompts☆24Updated last year
- This is the repo for the paper Shepherd -- A Critic for Language Model Generation☆219Updated last year
- ☆31Updated last year
- [NeurIPS 2023] PyTorch code for Can Language Models Teach? Teacher Explanations Improve Student Performance via Theory of Mind☆66Updated last year
- Multi-Domain Expert Learning☆67Updated last year
- A set of utilities for running few-shot prompting experiments on large-language models☆122Updated last year
- Code repo for "Agent Instructs Large Language Models to be General Zero-Shot Reasoners"☆114Updated 10 months ago
- ☆29Updated this week
- Low-Rank adapter extraction for fine-tuned transformers models☆175Updated last year
- Load multiple LoRA modules simultaneously and automatically switch the appropriate combination of LoRA modules to generate the best answe…☆156Updated last year
- Sakura-SOLAR-DPO: Merge, SFT, and DPO☆116Updated last year
- TART: A plug-and-play Transformer module for task-agnostic reasoning☆200Updated 2 years ago
- Flacuna was developed by fine-tuning Vicuna on Flan-mini, a comprehensive instruction collection encompassing various tasks. Vicuna is al…☆111Updated last year
- Advanced Reasoning Benchmark Dataset for LLMs☆47Updated last year
- ☆118Updated 11 months ago