megvii-research / megfileLinks
Megvii FILE Library - Working with Files in Python same as the standard library
☆163Updated this week
Alternatives and similar repositories for megfile
Users that are interested in megfile are comparing it to the libraries listed below
Sorting:
- To pioneer training long-context multi-modal transformer models☆61Updated 3 months ago
- mllm-npu: training multimodal large language models on Ascend NPUs☆94Updated last year
- FireFlyer Record file format, writer and reader for DL training samples.☆236Updated 2 years ago
- A Distributed Attention Towards Linear Scalability for Ultra-Long Context, Heterogeneous Data Training☆562Updated last week
- ☆439Updated 3 months ago
- Demystify RAM Usage in Multi-Process Data Loaders☆204Updated 2 years ago
- Large scale image dataset visiualization tool.☆121Updated last week
- VeOmni: Scaling Any Modality Model Training with Model-Centric Distributed Recipe Zoo☆1,338Updated this week
- A parallelism VAE avoids OOM for high resolution image generation☆83Updated 3 months ago
- An industrial extension library of pytorch to accelerate large scale model training☆52Updated 3 months ago
- Patch convolution to avoid large GPU memory usage of Conv2D☆93Updated 10 months ago
- 🤗A PyTorch-native Inference Engine with Hybrid Cache Acceleration and Parallelism for DiTs.☆588Updated this week
- useful dotfiles included vim, zsh, tmux and vscode☆18Updated 2 months ago
- USP: Unified (a.k.a. Hybrid, 2D) Sequence Parallel Attention for Long Context Transformers Model Training and Inference☆605Updated last month
- A hyperparameter manager for deep learning experiments.☆96Updated 3 years ago
- flex-block-attn: an efficient block sparse attention computation library☆88Updated this week
- High performance inference engine for diffusion models☆95Updated 2 months ago
- Datasets, Transforms and Models specific to Computer Vision☆90Updated 2 years ago
- Megatron's multi-modal data loader☆278Updated last week
- CVFusion is an open-source deep learning compiler to fuse the OpenCV operators.☆32Updated 3 years ago
- A light-weight and high-efficient training framework for accelerating diffusion tasks.☆50Updated last year
- Code for Draft Attention☆93Updated 6 months ago
- Ascend PyTorch adapter (torch_npu). Mirror of https://gitee.com/ascend/pytorch☆458Updated last week
- TVMScript kernel for deformable attention☆25Updated 3 years ago
- ☆187Updated 10 months ago
- A sparse attention kernel supporting mix sparse patterns☆385Updated 9 months ago
- ByteCheckpoint: An Unified Checkpointing Library for LFMs☆252Updated 4 months ago
- [ICLR 2025] COAT: Compressing Optimizer States and Activation for Memory-Efficient FP8 Training☆250Updated 3 months ago
- MegEngine到其他框架的转换器☆70Updated 2 years ago
- pytorch-profiler☆51Updated 2 years ago