microsoft / deep-language-networksLinks
We view Large Language Models as stochastic language layers in a network, where the learnable parameters are the natural language prompts at each layer. We stack two such layers, feeding the output of one layer to the next. We call the stacked architecture a Deep Language Network - DLN
☆95Updated last year
Alternatives and similar repositories for deep-language-networks
Users that are interested in deep-language-networks are comparing it to the libraries listed below
Sorting:
- The Official Repository for "Bring Your Own Data! Self-Supervised Evaluation for Large Language Models"☆107Updated 2 years ago
- [EMNLP'23] Execution-Based Evaluation for Open Domain Code Generation☆49Updated 2 years ago
- SILO Language Models code repository☆83Updated last year
- ☆44Updated last year
- RL algorithm: Advantage induced policy alignment☆66Updated 2 years ago
- ☆150Updated 2 years ago
- A repository for transformer critique learning and generation☆89Updated 2 years ago
- ☆77Updated last year
- [NeurIPS 2023 Main Track] This is the repository for the paper titled "Don’t Stop Pretraining? Make Prompt-based Fine-tuning Powerful Lea…☆76Updated 2 years ago
- Code of ICLR paper: https://openreview.net/forum?id=-cqvvvb-NkI☆95Updated 2 years ago
- [ICLR 2023] Guess the Instruction! Flipped Learning Makes Language Models Stronger Zero-Shot Learners☆116Updated 7 months ago
- Repository for NPHardEval, a quantified-dynamic benchmark of LLMs☆63Updated last year
- The GitHub repo for Goal Driven Discovery of Distributional Differences via Language Descriptions☆72Updated 2 years ago
- [AAAI 2024] Investigating the Effectiveness of Task-Agnostic Prefix Prompt for Instruction Following☆78Updated last year
- ☆57Updated 2 years ago
- Self-Alignment with Principle-Following Reward Models☆169Updated 4 months ago
- CodeUltraFeedback: aligning large language models to coding preferences (TOSEM 2025)☆73Updated last year
- ☆80Updated 10 months ago
- Code accompanying the paper Pretraining Language Models with Human Preferences☆180Updated last year
- A unified benchmark for math reasoning☆89Updated 3 years ago
- A dataset of LLM-generated chain-of-thought steps annotated with mistake location.☆85Updated last year
- This project studies the performance and robustness of language models and task-adaptation methods.☆155Updated last year
- Code for the ICLR 2024 paper "How to catch an AI liar: Lie detection in black-box LLMs by asking unrelated questions"☆71Updated last year
- Language models scale reliably with over-training and on downstream tasks☆99Updated last year
- Code for the arXiv paper: "LLMs as Factual Reasoners: Insights from Existing Benchmarks and Beyond"☆61Updated last year
- Code and data accompanying our paper on arXiv "Faithful Chain-of-Thought Reasoning".☆165Updated last year
- Code for paper "LEVER: Learning to Verifiy Language-to-Code Generation with Execution" (ICML'23)☆90Updated 2 years ago
- Scalable Meta-Evaluation of LLMs as Evaluators☆43Updated last year
- ☆158Updated 2 years ago
- Building modular LMs with parameter-efficient fine-tuning.☆114Updated 3 weeks ago