project-codeflare / zero-copy-model-loading
In-depth code associated with my Medium blog post, "How to Load PyTorch Models 340 Times Faster with Ray"
☆26Updated 2 years ago
Alternatives and similar repositories for zero-copy-model-loading
Users that are interested in zero-copy-model-loading are comparing it to the libraries listed below
Sorting:
- Simple dependency injection framework for Python☆21Updated last year
- Pygloo provides Python bindings for Gloo.☆22Updated 2 months ago
- Productionize machine learning predictions, with ONNX or without☆65Updated last year
- Module, Model, and Tensor Serialization/Deserialization☆229Updated this week
- ☆16Updated 2 years ago
- A collection of reproducible inference engine benchmarks☆30Updated 3 weeks ago
- Utilities for Training Very Large Models☆58Updated 7 months ago
- Notes and artifacts from the ONNX steering committee☆26Updated this week
- Serialize JAX, Flax, Haiku, or Objax model params with 🤗`safetensors`☆44Updated 11 months ago
- The Triton backend for the PyTorch TorchScript models.☆150Updated last week
- 🐍 Python bidding for the Hora Approximate Nearest Neighbor Search Algorithm library☆72Updated 3 years ago
- A lightweight wrapper for PyTorch that provides a simple declarative API for context switching between devices, distributed modes, mixed-…☆67Updated last year
- 🚀 Collection of libraries used with fms-hf-tuning to accelerate fine-tuning and training of large models.☆9Updated 2 weeks ago
- A performant, memory-efficient checkpointing library for PyTorch applications, designed with large, complex distributed workloads in mind…☆157Updated 5 months ago
- Demo of the unit_scaling library, showing how a model can be easily adapted to train in FP8.☆45Updated 10 months ago
- The largest multilingual image-text classification dataset. It contains fashion products.☆72Updated last year
- No-GIL Python environment featuring NVIDIA Deep Learning libraries.☆60Updated last month
- A stand-alone implementation of several NumPy dtype extensions used in machine learning.☆262Updated this week
- torch::deploy (multipy for non-torch uses) is a system that lets you get around the GIL problem by running multiple Python interpreters i…☆180Updated 5 months ago
- TorchFix - a linter for PyTorch-using code with autofix support☆141Updated 3 months ago
- Deploy your HPC Cluster on AWS in 20min. with just 1-Click.☆55Updated last month
- High-performance safetensors model loader☆30Updated last month
- ☆13Updated last year
- ☆32Updated this week
- Provide Python access to the NVML library for GPU diagnostics☆235Updated 5 months ago
- ☆207Updated last week
- Accelerated inference of 🤗 models using FuriosaAI NPU chips.☆26Updated 11 months ago
- A file utility for accessing both local and remote files through a unified interface.☆42Updated this week
- Model compression for ONNX☆92Updated 6 months ago
- Locality Sensitive Hashing☆73Updated last year