NVIDIA / multi-storage-clientLinks
Unified high-performance Python client for object and file stores.
☆29Updated 2 weeks ago
Alternatives and similar repositories for multi-storage-client
Users that are interested in multi-storage-client are comparing it to the libraries listed below
Sorting:
- Container plugin for Slurm Workload Manager☆356Updated last week
- AIStore: scalable storage for AI applications☆1,564Updated this week
- NVIDIA Resiliency Extension is a python package for framework developers and users to implement fault-tolerant features. It improves the …☆193Updated last week
- Fault tolerance for PyTorch (HSDP, LocalSGD, DiLoCo, Streaming DiLoCo)☆365Updated last week
- cuVS - a library for vector search and clustering on the GPU☆462Updated this week
- RAFT contains fundamental widely-used algorithms and primitives for machine learning and information retrieval. The algorithms are CUDA-a…☆915Updated this week
- NVIDIA Inference Xfer Library (NIXL)☆484Updated this week
- Pipeline Parallelism for PyTorch☆772Updated 11 months ago
- DGXC Benchmarking provides recipes in ready-to-use templates for evaluating performance of specific AI use cases across hardware and soft…☆18Updated 3 weeks ago
- Tools to deploy GPU clusters in the Cloud☆31Updated 2 years ago
- A high-throughput and memory-efficient inference and serving engine for LLMs☆78Updated this week
- TorchX is a universal job launcher for PyTorch applications. TorchX is designed to have fast iteration time for training/research and sup…☆376Updated this week
- A library to analyze PyTorch traces.☆397Updated last week
- A tool to configure, launch and manage your machine learning experiments.☆173Updated this week
- ☆52Updated 4 months ago
- KvikIO - High Performance File IO☆218Updated this week
- A Fusion Code Generator for NVIDIA GPUs (commonly known as "nvFuser")☆341Updated this week
- A CPU+GPU Profiling library that provides access to timeline traces and hardware performance counters.☆840Updated this week
- A simple yet powerful tool to turn traditional container/OS images into unprivileged sandboxes.☆798Updated 7 months ago
- JetStream is a throughput and memory optimized engine for LLM inference on XLA devices, starting with TPUs (and GPUs in future -- PRs wel…☆359Updated last month
- 🚀 Collection of components for development, training, tuning, and inference of foundation models leveraging PyTorch native components.☆207Updated last week
- ☆60Updated 3 weeks ago
- The Amazon S3 Connector for PyTorch delivers high throughput for PyTorch training jobs that access and store data in Amazon S3.☆171Updated last week
- Examples demonstrating available options to program multiple GPUs in a single node or a cluster☆757Updated 5 months ago
- A tool for bandwidth measurements on NVIDIA GPUs.☆487Updated 3 months ago
- 🚀 Efficiently (pre)training foundation models with native PyTorch features, including FSDP for training and SDPA implementation of Flash…☆256Updated 2 weeks ago
- Scalable and Performant Data Loading☆289Updated this week
- Easy and lightning fast training of 🤗 Transformers on Habana Gaudi processor (HPU)☆190Updated this week
- VUA stands for 'VAST Undivided Attention'. It's a global KVCache storage solution optimizing LLM time to first token (TTFT) and GPU utili…☆18Updated last month
- A throughput-oriented high-performance serving framework for LLMs☆846Updated 2 weeks ago