iPieter / universal-distillation
π§ͺCreate domain-adapted language models by distilling from many pre-trained LMs
β10Updated 2 years ago
Alternatives and similar repositories for universal-distillation:
Users that are interested in universal-distillation are comparing it to the libraries listed below
- β12Updated last month
- Reference implementation for Reward-Augmented Decoding: Efficient Controlled Text Generation With a Unidirectional Reward Modelβ42Updated last year
- Repository for Skill Set Optimizationβ12Updated 7 months ago
- β21Updated last month
- Minimum Description Length probing for neural network representationsβ19Updated last month
- Aioli: A unified optimization framework for language model data mixingβ22Updated 2 months ago
- This repo contains code for the paper "Psychologically-informed chain-of-thought prompts for metaphor understanding in large language modβ¦β14Updated last year
- Exploration using DSPy to optimize modules to maximize performance on the OpenToM datasetβ15Updated last year
- β23Updated 6 months ago
- β19Updated 4 months ago
- This is a new metric that can be used to evaluate faithfulness of text generated by LLMs. The work behind this repository can be found heβ¦β31Updated last year
- Starbucks: Improved Training for 2D Matryoshka Embeddingsβ18Updated last month
- The code implementation of MAGDi: Structured Distillation of Multi-Agent Interaction Graphs Improves Reasoning in Smaller Language Modelsβ¦β33Updated last year
- Few-shot Learning with Auxiliary Dataβ27Updated last year
- This is official project in our paper: Is Bigger and Deeper Always Better? Probing LLaMA Across Scales and Layersβ30Updated last year
- β14Updated 5 months ago
- Implementation of the model: "Reka Core, Flash, and Edge: A Series of Powerful Multimodal Language Models" in PyTorchβ30Updated last week
- This repository includes a benchmark and code for the paper "Evaluating LLMs at Detecting Errors in LLM Responses".β27Updated 7 months ago
- IntructIR, a novel benchmark specifically designed to evaluate the instruction following ability in information retrieval models. Our focβ¦β31Updated 9 months ago
- Script for processing OpenAI's PRM800K process supervision dataset into an Alpaca-style instruction-response formatβ27Updated last year
- Tasks for describing differences between text distributions.β16Updated 7 months ago
- Repository containing the SPIN experiments on the DIBT 10k ranked promptsβ24Updated last year
- Code for Stage-wise Fine-tuning for Graph-to-Text Generationβ26Updated 2 years ago
- Plug-and-play Search Interfaces with Pyserini and Hugging Faceβ31Updated last year
- β17Updated last month
- β13Updated 3 months ago
- Python package for generating datasets to evaluate reasoning and retrieval of large language modelsβ15Updated this week
- In-Context Alignment: Chat with Vanilla Language Models Before Fine-Tuningβ34Updated last year
- Measuring and Controlling Persona Drift in Language Model Dialogsβ17Updated last year
- Mixture of Expert (MoE) techniques for enhancing LLM performance through expert-driven prompt mapping and adapter combinations.β12Updated last year