huggingface / safetensorsLinks
Simple, safe way to store and distribute tensors
β3,462Updated last week
Alternatives and similar repositories for safetensors
Users that are interested in safetensors are comparing it to the libraries listed below
Sorting:
- π Accelerate inference and training of π€ Transformers, Diffusers, TIMM and Sentence Transformers with easy to use hardware optimizationβ¦β3,104Updated this week
- Accessible large language models via k-bit quantization for PyTorch.β7,619Updated this week
- Hackable and optimized Transformers building blocks, supporting a composable construction.β9,967Updated this week
- AITemplate is a Python framework which renders neural network into high performance CUDA/HIP C++ code. Specialized for FP16 TensorCore (Nβ¦β4,681Updated 2 weeks ago
- π A simple way to launch, train, and use PyTorch models on almost any device and distributed configuration, automatic mixed precision (iβ¦β9,165Updated last week
- MII makes low-latency and high-throughput inference possible, powered by DeepSpeed.β2,060Updated 3 months ago
- Large Language Model Text Generation Inferenceβ10,539Updated 2 weeks ago
- PyTorch native quantization and sparsity for training and inferenceβ2,384Updated this week
- A machine learning compiler for GPUs, CPUs, and ML acceleratorsβ3,564Updated this week
- Fast inference engine for Transformer modelsβ4,044Updated 5 months ago
- PyTorch extensions for high performance and large scale training.β3,376Updated 5 months ago
- Tensor library for machine learningβ13,222Updated this week
- Kernl lets you run PyTorch transformer models several times faster on GPU with a single line of code, and is designed to be easily hackabβ¦β1,584Updated last year
- β‘ Build your chatbot within minutes on your favorite device; offer SOTA compression techniques for LLMs; run LLMs efficiently on Intel Plβ¦β2,168Updated 11 months ago
- A library for accelerating Transformer models on NVIDIA GPUs, including using 8-bit floating point (FP8) precision on Hopper, Ada and Blaβ¦β2,755Updated this week
- Inference Llama 2 in one file of pure π₯β2,117Updated last year
- Transformer related optimization, including BERT, GPTβ6,312Updated last year
- Python bindings for the Transformer models implemented in C/C++ using GGML library.β1,878Updated last year
- PyTorch compiler that accelerates training and inference. Get built-in optimizations for performance, memory, parallelism, and easily wriβ¦β1,416Updated this week
- π€ Evaluate: A library for easily evaluating machine learning models and datasets.β2,326Updated this week
- Serve, optimize and scale PyTorch models in productionβ4,350Updated last month
- A Python package for extending the official PyTorch that can easily obtain performance on Intel platformβ1,969Updated this week
- Python bindings for llama.cppβ9,609Updated last month
- A high-performance ML model serving framework, offers dynamic batching and CPU/GPU pipelines to fully exploit your compute machineβ860Updated last month
- PyTorch native post-training libraryβ5,517Updated this week
- β2,889Updated this week
- PyTriton is a Flask/FastAPI-like interface that simplifies Triton's deployment in Python environments.β821Updated last month
- A high-performance Python-based I/O system for large (and small) deep learning problems, with strong support for PyTorch.β2,825Updated 3 months ago
- A blazing fast inference solution for text embeddings modelsβ4,045Updated this week
- A Python-level JIT compiler designed to make unmodified PyTorch programs faster.β1,063Updated last year