nrasadi / split-keras-tensorflow-model
Splits Keras with Tensorflow backends into two or more submodels.
☆18Updated 2 years ago
Alternatives and similar repositories for split-keras-tensorflow-model:
Users that are interested in split-keras-tensorflow-model are comparing it to the libraries listed below
- Reference implementations of popular Binarized Neural Networks☆107Updated last week
- Model-less Inference Serving☆88Updated last year
- This is a list of awesome edgeAI inference related papers.☆95Updated last year
- FilterForward: Scaling Video Analytics on Constrained Edge Nodes☆28Updated 5 years ago
- ☆19Updated 3 years ago
- LegoDNN: a block-grained scaling tool for mobile vision systems☆50Updated last year
- Source code and datasets for Ekya, a system for continuous learning on the edge.☆105Updated 3 years ago
- [MLSys 2021] IOS: Inter-Operator Scheduler for CNN Acceleration☆197Updated 2 years ago
- ☆127Updated last year
- An analytical performance modeling tool for deep neural networks.☆88Updated 4 years ago
- Post-training sparsity-aware quantization☆34Updated 2 years ago
- Inference of quantization aware trained networks using TensorRT☆80Updated 2 years ago
- The code for paper: Neuralpower: Predict and deploy energy-efficient convolutional neural networks☆21Updated 5 years ago
- A Portable C Library for Distributed CNN Inference on IoT Edge Clusters☆83Updated 5 years ago
- Code for High-Capacity Expert Binary Networks (ICLR 2021).☆27Updated 3 years ago
- [CVPR'20] ZeroQ: A Novel Zero Shot Quantization Framework☆277Updated last year
- Quantization of Convolutional Neural networks.☆244Updated 8 months ago
- ☆14Updated 3 years ago
- Train neural networks with joint quantization and pruning on both weights and activations using any pytorch modules☆41Updated 2 years ago
- Implementation of "NITI: Training Integer Neural Networks Using Integer-only Arithmetic" on arxiv☆80Updated 2 years ago
- Multi-Instance-GPU profiling tool☆57Updated 2 years ago
- Official PyTorch Implementation of HELP: Hardware-adaptive Efficient Latency Prediction for NAS via Meta-Learning (NeurIPS 2021 Spotlight…☆63Updated 8 months ago
- ML model training for edge devices☆162Updated last year
- A Tool for Automatic Parallelization of Deep Learning Training in Distributed Multi-GPU Environments.☆130Updated 3 years ago
- PipeTransformer: Automated Elastic Pipelining for Distributed Training of Large-scale Models. ICML 2021☆56Updated 3 years ago
- ☆19Updated 2 years ago
- [ICLR 2021] HW-NAS-Bench: Hardware-Aware Neural Architecture Search Benchmark☆110Updated 2 years ago
- Simple Training and Deployment of Fast End-to-End Binary Networks☆157Updated 3 years ago
- sensAI: ConvNets Decomposition via Class Parallelism for Fast Inference on Live Data☆64Updated 9 months ago
- Fine-grained GPU sharing primitives☆141Updated 5 years ago