microsoft / deep-language-networksLinks
We view Large Language Models as stochastic language layers in a network, where the learnable parameters are the natural language prompts at each layer. We stack two such layers, feeding the output of one layer to the next. We call the stacked architecture a Deep Language Network - DLN
☆95Updated last year
Alternatives and similar repositories for deep-language-networks
Users that are interested in deep-language-networks are comparing it to the libraries listed below
Sorting:
- The Official Repository for "Bring Your Own Data! Self-Supervised Evaluation for Large Language Models"☆107Updated 2 years ago
- Repository for NPHardEval, a quantified-dynamic benchmark of LLMs☆62Updated last year
- [EMNLP'23] Execution-Based Evaluation for Open Domain Code Generation☆49Updated 2 years ago
- SILO Language Models code repository☆83Updated last year
- A repository for transformer critique learning and generation☆89Updated 2 years ago
- RL algorithm: Advantage induced policy alignment☆66Updated 2 years ago
- Scalable Meta-Evaluation of LLMs as Evaluators☆43Updated last year
- Code for the arXiv preprint "The Unreasonable Effectiveness of Easy Training Data"☆48Updated last year
- Code of ICLR paper: https://openreview.net/forum?id=-cqvvvb-NkI☆95Updated 2 years ago
- ☆44Updated last year
- ☆129Updated last year
- Reference implementation for Reward-Augmented Decoding: Efficient Controlled Text Generation With a Unidirectional Reward Model☆45Updated 3 months ago
- Code for the ICLR 2024 paper "How to catch an AI liar: Lie detection in black-box LLMs by asking unrelated questions"☆71Updated last year
- PASTA: Post-hoc Attention Steering for LLMs☆132Updated last year
- [ICLR 2023] Guess the Instruction! Flipped Learning Makes Language Models Stronger Zero-Shot Learners☆116Updated 6 months ago
- Self-Alignment with Principle-Following Reward Models☆169Updated 3 months ago
- A dataset of LLM-generated chain-of-thought steps annotated with mistake location.☆85Updated last year
- Advanced Reasoning Benchmark Dataset for LLMs☆47Updated 2 years ago
- ☆76Updated last year
- ☆150Updated 2 years ago
- Large language models (LLMs) made easy, EasyLM is a one stop solution for pre-training, finetuning, evaluating and serving LLMs in JAX/Fl…☆78Updated last year
- ☆32Updated last year
- CodeUltraFeedback: aligning large language models to coding preferences (TOSEM 2025)☆73Updated last year
- Language models scale reliably with over-training and on downstream tasks☆100Updated last year
- ☆80Updated 9 months ago
- A unified benchmark for math reasoning☆89Updated 2 years ago
- Code for the arXiv paper: "LLMs as Factual Reasoners: Insights from Existing Benchmarks and Beyond"☆61Updated 11 months ago
- Code and data accompanying our paper on arXiv "Faithful Chain-of-Thought Reasoning".☆165Updated last year
- This repository includes code for the paper "Does Localization Inform Editing? Surprising Differences in Where Knowledge Is Stored vs. Ca…☆60Updated 2 years ago
- Aioli: A unified optimization framework for language model data mixing☆32Updated 11 months ago