Spico197 / watchmen
😎 A simple and easy-to-use toolkit for GPU scheduling.
☆42Updated 3 years ago
Alternatives and similar repositories for watchmen:
Users that are interested in watchmen are comparing it to the libraries listed below
- Source code for our EMNLP'21 paper 《Raise a Child in Large Language Model: Towards Effective and Generalizable Fine-tuning》☆57Updated 3 years ago
- ☆32Updated 3 years ago
- 服务器 GPU 监控程序,当 GPU 属性满足预设条件时通过微信发送提示消息☆29Updated 3 years ago
- ☆73Updated 2 years ago
- Code for EMNLP 2021 main conference paper "Dynamic Knowledge Distillation for Pre-trained Language Models"☆40Updated 2 years ago
- [NeurIPS 2022] "A Win-win Deal: Towards Sparse and Robust Pre-trained Language Models", Yuanxin Liu, Fandong Meng, Zheng Lin, Jiangnan Li…☆21Updated last year
- [ICLR 2024]EMO: Earth Mover Distance Optimization for Auto-Regressive Language Modeling(https://arxiv.org/abs/2310.04691)☆119Updated 11 months ago
- Ladder Side-Tuning在CLUE上的简单尝试☆19Updated 2 years ago
- 🎮 A toolkit for Relation Extraction and more...☆24Updated 4 months ago
- A light-weight script for maintaining a LOT of machine learning experiments.☆91Updated 2 years ago
- ☆46Updated last month
- ICLR2023 - Tailoring Language Generation Models under Total Variation Distance☆21Updated 2 years ago
- Mixture of Attention Heads☆41Updated 2 years ago
- 擂台赛3-大规模预训练调优比赛的示例代码与baseline实现☆38Updated 2 years ago
- [EVA ICLR'23; LARA ICML'22] Efficient attention mechanisms via control variates, random features, and importance sampling☆80Updated last year
- The code and data for the paper JiuZhang3.0☆40Updated 8 months ago
- Must-read papers on improving efficiency for pre-trained language models.☆102Updated 2 years ago
- Code for ACL 2023 paper titled "Lifting the Curse of Capacity Gap in Distilling Language Models"☆28Updated last year
- Implementation of ICLR 2022 paper "Enhancing Cross-lingual Transfer by Manifold Mixup".☆21Updated 2 years ago
- ☆65Updated 9 months ago
- Code for EMNLP 2022 paper “Distilled Dual-Encoder Model for Vision-Language Understanding”☆29Updated last year
- ☆45Updated 5 months ago
- Code for the ACL-2022 paper "StableMoE: Stable Routing Strategy for Mixture of Experts"☆45Updated 2 years ago
- my commonly-used tools☆50Updated last month
- Code for the AAAI 2022 publication "Well-classified Examples are Underestimated in Classification with Deep Neural Networks"☆49Updated 2 years ago
- 基于Gated Attention Unit的Transformer模型(尝鲜版)☆97Updated last year
- ☆39Updated last year
- 本文旨在整理文本生成领域国内外工业界和企业家的研究者和研究机构。排名不分 先后。更新中,欢迎大家补充☆48Updated 4 years ago
- ☆16Updated last year
- Feeling confused about super alignment? Here is a reading list☆42Updated last year