MadryLab / DsDm
☆43Updated 9 months ago
Related projects ⓘ
Alternatives and complementary repositories for DsDm
- Revisiting Efficient Training Algorithms For Transformer-based Language Models (NeurIPS 2023)☆79Updated last year
- AI Logging for Interpretability and Explainability🔬☆87Updated 5 months ago
- ☆49Updated last year
- Long Is More for Alignment: A Simple but Tough-to-Beat Baseline for Instruction Fine-Tuning [ICML 2024]☆14Updated 6 months ago
- Repo accompanying our paper "Do Llamas Work in English? On the Latent Language of Multilingual Transformers".☆56Updated 7 months ago
- ☆50Updated 5 months ago
- [SafeGenAi @ NeurIPS 2024] Cheating Automatic LLM Benchmarks: Null Models Achieve High Win Rates☆58Updated 2 weeks ago
- ☆26Updated last week
- Language models scale reliably with over-training and on downstream tasks☆94Updated 7 months ago
- ☆34Updated 3 months ago
- A fusion of a linear layer and a cross entropy loss, written for pytorch in triton.☆54Updated 3 months ago
- Skill-It! A Data-Driven Skills Framework for Understanding and Training Language Models☆41Updated last year
- Code accompanying the paper "Massive Activations in Large Language Models"☆121Updated 8 months ago
- Implementation of PaCE: Parsimonious Concept Engineering for Large Language Models (NeurIPS 2024)☆26Updated this week
- Is In-Context Learning Sufficient for Instruction Following in LLMs?☆23Updated 5 months ago
- Function Vectors in Large Language Models (ICLR 2024)☆116Updated 3 weeks ago
- ☆15Updated last week
- Code release for "Debating with More Persuasive LLMs Leads to More Truthful Answers"☆83Updated 7 months ago
- ☆32Updated last year
- ☆78Updated last year
- Codebase for Instruction Following without Instruction Tuning☆29Updated last month
- [ICLR'24] RAIN: Your Language Models Can Align Themselves without Finetuning☆83Updated 5 months ago
- ☆50Updated last week
- Replicating O1 inference-time scaling laws☆48Updated last month
- Stick-breaking attention☆32Updated last week
- Test-time-training on nearest neighbors for large language models☆25Updated 6 months ago
- Röttger et al. (2023): "XSTest: A Test Suite for Identifying Exaggerated Safety Behaviours in Large Language Models"☆61Updated 10 months ago
- Repository for NPHardEval, a quantified-dynamic benchmark of LLMs☆48Updated 7 months ago
- ☆65Updated 7 months ago
- [NeurIPS-2024] 📈 Scaling Laws with Vocabulary: Larger Models Deserve Larger Vocabularies https://arxiv.org/abs/2407.13623☆67Updated last month