Spico197 / watchmenLinks
😎 A simple and easy-to-use toolkit for GPU scheduling.
☆45Updated 5 months ago
Alternatives and similar repositories for watchmen
Users that are interested in watchmen are comparing it to the libraries listed below
Sorting:
- 服务器 GPU 监控程序,当 GPU 属性满足预设条件时通过微信发送提示消息☆32Updated 4 years ago
 - Must-read papers on improving efficiency for pre-trained language models.☆105Updated 3 years ago
 - Source code for our AAAI'22 paper 《From Dense to Sparse: Contrastive Pruning for Better Pre-trained Language Model Compression》☆25Updated 3 years ago
 - [EVA ICLR'23; LARA ICML'22] Efficient attention mechanisms via control variates, random features, and importance sampling☆87Updated 2 years ago
 - ☆73Updated 3 years ago
 - Source code for our EMNLP'21 paper 《Raise a Child in Large Language Model: Towards Effective and Generalizable Fine-tuning》☆61Updated 3 years ago
 - ☆48Updated last year
 - 🎮 A toolkit for Relation Extraction and more...☆24Updated 5 months ago
 - A light-weight script for maintaining a LOT of machine learning experiments.☆92Updated 2 years ago
 - [ICLR 2024]EMO: Earth Mover Distance Optimization for Auto-Regressive Language Modeling(https://arxiv.org/abs/2310.04691)☆126Updated last year
 - [NeurIPS 2022] "A Win-win Deal: Towards Sparse and Robust Pre-trained Language Models", Yuanxin Liu, Fandong Meng, Zheng Lin, Jiangnan Li…☆21Updated last year
 - 基于Gated Attention Unit的Transformer模型(尝鲜版)☆98Updated 2 years ago
 - ☆45Updated 2 weeks ago
 - The official repo for the paper "Teacher Forcing Recovers Reward Functions for Text Generation"☆31Updated 2 years ago
 - self-adaptive in-context learning☆45Updated 2 years ago
 - ☆33Updated 4 years ago
 - Released code for our ICLR23 paper.☆66Updated 2 years ago
 - Code for ACL 2022 paper "BERT Learns to Teach: Knowledge Distillation with Meta Learning".☆86Updated 3 years ago
 - Code for EMNLP 2021 main conference paper "Dynamic Knowledge Distillation for Pre-trained Language Models"☆41Updated 3 years ago
 - Code for paper "UniPELT: A Unified Framework for Parameter-Efficient Language Model Tuning", ACL 2022☆63Updated 3 years ago
 - Feeling confused about super alignment? Here is a reading list☆43Updated last year
 - ☆43Updated 2 years ago
 - Princeton NLP's pre-training library based on fairseq with DeepSpeed kernel integration 🚃☆114Updated 3 years ago
 - code for promptCSE, emnlp 2022☆11Updated 2 years ago
 - CMATH: Can your language model pass Chinese elementary school math test?☆47Updated 2 years ago
 - Code for EMNLP 2022 paper “Distilled Dual-Encoder Model for Vision-Language Understanding”☆31Updated 2 years ago
 - [NeurIPS'24] Weak-to-Strong Search: Align Large Language Models via Searching over Small Language Models☆62Updated 10 months ago
 - This is the repo for our paper "Mr-Ben: A Comprehensive Meta-Reasoning Benchmark for Large Language Models"☆50Updated last year
 - 本文旨在整理文本生成领域国内外工业界和企业家的研究者和研究机构。排名不分先后。更新中,欢迎大家补充☆51Updated 4 years ago
 - ☆23Updated 3 years ago