cnstark / gputaskerLinks
An awesome gpu tasks scheduler. 轻量好用的GPU机群任务调度工具。觉得有用可以点个star
☆188Updated 2 years ago
Alternatives and similar repositories for gputasker
Users that are interested in gputasker are comparing it to the libraries listed below
Sorting:
- 📊 A simple command-line utility for querying and monitoring GPU status☆92Updated 2 years ago
- 一款便捷的抢占显卡脚本☆343Updated 6 months ago
- Cool Papers - Immersive Paper Discovery☆590Updated 2 months ago
- pytorch单精度、半精度、混合精度、单卡、多卡(DP / DDP)、FSDP、DeepSpeed模型训练代码,并对比不同方法的训练速度以及GPU内存的使用☆114Updated last year
- The record of what I‘ve been through.☆100Updated 6 months ago
- DeepSpeed教程 & 示例注释 & 学习笔记 (大模型高效训练)☆173Updated last year
- DeepSpeed Tutorial☆100Updated 11 months ago
- The pure and clear PyTorch Distributed Training Framework.☆275Updated last year
- A light-weight script for maintaining a LOT of machine learning experiments.☆92Updated 2 years ago
- Download papers and supplemental materials from open-access paper website, such as AAAI, AAMAS, AISTATS, COLT, CORL, CVPR, ECCV, ICCV, IC…☆268Updated this week
- ☆61Updated last year
- 实验室服务器管理☆29Updated last year
- ☆172Updated this week
- How to use wandb?☆671Updated last year
- More light-weight pytorch experiment management library!☆67Updated 2 years ago
- Yet another PyTorch Trainer and some core components for deep learning.☆222Updated last year
- 多模态 MM +Chat 合集☆273Updated 2 months ago
- ZO2 (Zeroth-Order Offloading): Full Parameter Fine-Tuning 175B LLMs with 18GB GPU Memory☆166Updated 3 weeks ago
- Simple tutorials on Pytorch DDP training☆281Updated 2 years ago
- ☆260Updated 5 months ago
- A pupil in the computer world.(Felix Fu)☆241Updated last year
- mindspore implementation of transformers☆68Updated 2 years ago
- 主要记录大语言大模型(LLMs) 算法(应用)工程师多模态相关知识☆222Updated last year
- 抢占显卡☆75Updated 9 months ago
- deep learning template code☆66Updated last year
- ☆79Updated last year
- Survey Paper List - Efficient LLM and Foundation Models☆253Updated 10 months ago
- 看图学大模型☆316Updated last year
- LiBai(李白): A Toolbox for Large-Scale Distributed Parallel Training☆409Updated last week
- ☆166Updated this week