microsoft / deep-language-networks
We view Large Language Models as stochastic language layers in a network, where the learnable parameters are the natural language prompts at each layer. We stack two such layers, feeding the output of one layer to the next. We call the stacked architecture a Deep Language Network - DLN
☆93Updated 7 months ago
Alternatives and similar repositories for deep-language-networks:
Users that are interested in deep-language-networks are comparing it to the libraries listed below
- SILO Language Models code repository☆81Updated last year
- RL algorithm: Advantage induced policy alignment☆64Updated last year
- Building modular LMs with parameter-efficient fine-tuning.☆97Updated last week
- This repository includes code for the paper "Does Localization Inform Editing? Surprising Differences in Where Knowledge Is Stored vs. Ca…☆59Updated last year
- ☆34Updated last year
- ☆45Updated last year
- ☆160Updated last year
- Skill-It! A Data-Driven Skills Framework for Understanding and Training Language Models☆43Updated last year
- The Official Repository for "Bring Your Own Data! Self-Supervised Evaluation for Large Language Models"☆108Updated last year
- Code for the ICLR 2024 paper "How to catch an AI liar: Lie detection in black-box LLMs by asking unrelated questions"☆65Updated 8 months ago
- The official code of EMNLP 2022, "SCROLLS: Standardized CompaRison Over Long Language Sequences".☆69Updated last year
- Code for the arXiv preprint "The Unreasonable Effectiveness of Easy Training Data"☆46Updated last year
- [AAAI 2024] Investigating the Effectiveness of Task-Agnostic Prefix Prompt for Instruction Following☆79Updated 5 months ago
- A framework for few-shot evaluation of autoregressive language models.☆102Updated last year
- A repository for transformer critique learning and generation☆88Updated last year
- This project studies the performance and robustness of language models and task-adaptation methods.☆147Updated 9 months ago
- ☆34Updated 4 months ago
- ☆38Updated 10 months ago
- Code for the arXiv paper: "LLMs as Factual Reasoners: Insights from Existing Benchmarks and Beyond"☆59Updated last month
- [EMNLP'23] Execution-Based Evaluation for Open Domain Code Generation☆46Updated last year
- The GitHub repo for Goal Driven Discovery of Distributional Differences via Language Descriptions☆69Updated last year
- ☆26Updated 7 months ago
- Personalized Soups: Personalized Large Language Model Alignment via Post-hoc Parameter Merging☆98Updated last year
- ☆73Updated 10 months ago
- PyTorch code for the RetoMaton paper: "Neuro-Symbolic Language Modeling with Automaton-augmented Retrieval" (ICML 2022)☆71Updated 2 years ago
- Code and data accompanying our paper on arXiv "Faithful Chain-of-Thought Reasoning".☆157Updated 9 months ago
- Large language models (LLMs) made easy, EasyLM is a one stop solution for pre-training, finetuning, evaluating and serving LLMs in JAX/Fl…☆66Updated 6 months ago
- ☆67Updated last year
- Code of ICLR paper: https://openreview.net/forum?id=-cqvvvb-NkI☆94Updated 2 years ago
- ☆111Updated 7 months ago