lancopku / FedMNMT
[Findings of ACL 2023] Communication Efficient Federated Learning for Multilingual Machine Translation with Adapter
☆12Updated last year
Alternatives and similar repositories for FedMNMT:
Users that are interested in FedMNMT are comparing it to the libraries listed below
- Code for the paper "Pretrained Models for Multilingual Federated Learning" at NAACL 2022☆11Updated 2 years ago
- The official implement of paper "Does Federated Learning Really Need Backpropagation?"☆23Updated 2 years ago
- Code for LLM_Catastrophic_Forgetting via SAM.☆10Updated 9 months ago
- Crawl & visualize ICLR papers and reviews.☆18Updated 2 years ago
- Official implementation of Privacy Implications of Retrieval-Based Language Models (EMNLP 2023). https://arxiv.org/abs/2305.14888☆35Updated 9 months ago
- ☆21Updated last year
- [ICLR 2024] "Data Distillation Can Be Like Vodka: Distilling More Times For Better Quality" by Xuxi Chen*, Yu Yang*, Zhangyang Wang, Baha…☆12Updated 10 months ago
- Source code for the TMLR paper "Black-Box Prompt Learning for Pre-trained Language Models"☆55Updated last year
- On the Effectiveness of Parameter-Efficient Fine-Tuning☆38Updated last year
- Codebase for decoding compressed trust.☆23Updated 10 months ago
- Representation Surgery for Multi-Task Model Merging. ICML, 2024.☆42Updated 5 months ago
- Code for EMNLP 2021 main conference paper "Dynamic Knowledge Distillation for Pre-trained Language Models"☆40Updated 2 years ago
- [Findings of EMNLP22] From Mimicking to Integrating: Knowledge Integration for Pre-Trained Language Models☆19Updated 2 years ago
- ☆14Updated 5 months ago
- Mixture of Attention Heads☆43Updated 2 years ago
- Official repo for the paper: Recovering Private Text in Federated Learning of Language Models (in NeurIPS 2022)☆57Updated 2 years ago
- EMNLP 2024: Model Editing Harms General Abilities of Large Language Models: Regularization to the Rescue☆35Updated 4 months ago
- ☆10Updated 2 months ago
- Long Is More for Alignment: A Simple but Tough-to-Beat Baseline for Instruction Fine-Tuning [ICML 2024]☆17Updated 10 months ago
- ☆20Updated 3 months ago
- [ICML 2024] Junk DNA Hypothesis: A Task-Centric Angle of LLM Pre-trained Weights through Sparsity; Lu Yin*, Ajay Jaiswal*, Shiwei Liu, So…☆16Updated 9 months ago
- Less is More: Task-aware Layer-wise Distillation for Language Model Compression (ICML2023)☆34Updated last year
- ☆15Updated last year
- Restore safety in fine-tuned language models through task arithmetic☆28Updated last year
- [ICLR 2025] Code&Data for the paper "Super(ficial)-alignment: Strong Models May Deceive Weak Models in Weak-to-Strong Generalization"☆13Updated 9 months ago
- Code for "Universal Adversarial Triggers Are Not Universal."☆16Updated 10 months ago
- Data-free knowledge distillation using Gaussian noise (NeurIPS paper)☆15Updated 2 years ago
- [NeurIPS 2024] Official code of $\beta$-DPO: Direct Preference Optimization with Dynamic $\beta$☆41Updated 5 months ago
- ☆42Updated last year
- Code of Robust Lottery Tickets for Pre-trained Language Models (ACL2022)☆20Updated 2 years ago