nvidia-cosmos / cosmos-xennaLinks
Python library for building and running distributed data pipelines using Ray
☆46Updated 3 weeks ago
Alternatives and similar repositories for cosmos-xenna
Users that are interested in cosmos-xenna are comparing it to the libraries listed below
Sorting:
- Cosmos-RL is a flexible and scalable Reinforcement Learning framework specialized for Physical AI applications.☆274Updated this week
- CATransformers is a framework for joint neural network and hardware architecture search.☆19Updated 8 months ago
- Cosmos-Curate is a powerful video curation system that processes, analyzes, and organizes video content using advanced AI models and dist…☆131Updated this week
- Experimental scripts for researching data adaptive learning rate scheduling.☆22Updated 2 years ago
- MobileLLM-R1☆72Updated 3 months ago
- Easily run PyTorch on multiple GPUs & machines☆56Updated last month
- ☆47Updated last year
- imagetokenizer is a python package, helps you encoder visuals and generate visuals token ids from codebook, supports both image and video…☆37Updated last year
- Multi-Turn RL Training System with AgentTrainer for Language Model Game Reinforcement Learning☆57Updated 3 weeks ago
- Official Code Repository for EnvGen: Generating and Adapting Environments via LLMs for Training Embodied Agents (COLM 2024)☆39Updated last year
- Train, tune, and infer Bamba model☆137Updated 7 months ago
- Implementation of SmoothCache, a project aimed at speeding-up Diffusion Transformer (DiT) based GenAI models with error-guided caching.☆46Updated 5 months ago
- Code for “Pretrained Language Models as Visual Planners for Human Assistance”☆61Updated 2 years ago
- Pipeline parallelism for the minimalist☆37Updated 5 months ago
- Pytorch Distributed native training library for LLMs/VLMs with OOTB Hugging Face support☆232Updated this week
- Patch convolution to avoid large GPU memory usage of Conv2D☆93Updated 11 months ago
- Tree Attention: Topology-aware Decoding for Long-Context Attention on GPU clusters☆131Updated last year
- Defeating the Training-Inference Mismatch via FP16☆172Updated last month
- [ACL 2024] RelayAttention for Efficient Large Language Model Serving with Long System Prompts☆40Updated last year
- Pixel Parsing. A reproduction of OCR-free end-to-end document understanding models with open data☆23Updated last year
- Odysseus: Playground of LLM Sequence Parallelism☆79Updated last year
- ☆102Updated last year
- RWKV is an RNN with transformer-level LLM performance. It can be directly trained like a GPT (parallelizable). So it's combining the best…☆59Updated 9 months ago
- ☆91Updated last year
- A Video Tokenizer Evaluation Dataset☆147Updated 11 months ago
- JORA: JAX Tensor-Parallel LoRA Library (ACL 2024)☆36Updated last year
- Compression for Foundation Models☆35Updated 5 months ago
- Training-free Post-training Efficient Sub-quadratic Complexity Attention. Implemented with OpenAI Triton.☆147Updated 2 months ago
- Mixture-of-Transformers: A Sparse and Scalable Architecture for Multi-Modal Foundation Models. TMLR 2025.☆134Updated 3 months ago
- My attempt to improve the speed of the newton schulz algorithm, starting from the dion implementation.☆25Updated last month