megvii-research / megfile
Megvii FILE Library - Working with Files in Python same as the standard library
☆128Updated this week
Related projects ⓘ
Alternatives and complementary repositories for megfile
- Simple Dynamic Batching Inference☆145Updated 2 years ago
- MegEngine到其他框架的转换器☆67Updated last year
- FireFlyer Record file format, writer and reader for DL training samples.☆116Updated last year
- A hyperparameter manager for deep learning experiments.☆95Updated 2 years ago
- optimized BERT transformer inference on NVIDIA GPU. https://arxiv.org/abs/2210.03052☆457Updated 8 months ago
- Models and examples built with OneFlow☆96Updated last month
- ☆74Updated 11 months ago
- 📒A small curated list of Awesome Diffusion Inference Papers with codes.☆96Updated this week
- A collection of memory efficient attention operators implemented in the Triton language.☆219Updated 5 months ago
- Datasets, Transforms and Models specific to Computer Vision☆83Updated last year
- Ascend PyTorch adapter (torch_npu). Mirror of https://gitee.com/ascend/pytorch☆256Updated this week
- Easy Parallel Library (EPL) is a general and efficient deep learning framework for distributed model training.☆263Updated last year
- Zero Bubble Pipeline Parallelism☆282Updated last week
- Decode JPEG image on GPU using PyTorch☆84Updated last year
- Demystify RAM Usage in Multi-Process Data Loaders☆183Updated last year
- useful dotfiles included vim, zsh, tmux and vscode☆17Updated 3 weeks ago
- NART = NART is not A RunTime, a deep learning inference framework.☆38Updated last year
- mllm-npu: training multimodal large language models on Ascend NPUs☆83Updated 2 months ago
- ☆55Updated 4 years ago
- A high-performance, extensible Python AOT compiler.☆414Updated last year
- ActNN: Reducing Training Memory Footprint via 2-Bit Activation Compressed Training☆201Updated last year
- A model compression and acceleration toolbox based on pytorch.☆327Updated 10 months ago
- A codebase & model zoo for pretrained backbone based on MegEngine.☆33Updated last year
- USP: Unified (a.k.a. Hybrid, 2D) Sequence Parallel Attention for Long Context Transformers Model Training and Inference☆362Updated this week
- Transformer related optimization, including BERT, GPT☆60Updated last year
- TVMScript kernel for deformable attention☆24Updated 2 years ago
- LiBai(李白): A Toolbox for Large-Scale Distributed Parallel Training☆390Updated last week
- 一种任务级GPU算力分时调度的高性能深度学习训练平台☆311Updated last year
- GLake: optimizing GPU memory management and IO transmission.☆381Updated 3 months ago
- Microsoft Automatic Mixed Precision Library☆525Updated last month