DMALab / Reading_GroupLinks
DMALab's reading group slides and papers.
☆16Updated 4 years ago
Alternatives and similar repositories for Reading_Group
Users that are interested in Reading_Group are comparing it to the libraries listed below
Sorting:
- HogWild++: A New Mechanism for Decentralized Asynchronous Stochastic Gradient Descent☆33Updated 8 years ago
- implement distributed machine learning with Pytorch + OpenMPI☆51Updated 6 years ago
- Tutorial code on how to build your own Deep Learning System in 2k Lines☆125Updated 8 years ago
- code for https//research.fb.com/publications/towards-automated-neural-interaction-discovery-for-click-through-rate-prediction/☆18Updated 4 years ago
- Simple Distributed Deep Learning on TensorFlow☆133Updated last week
- GPU-specialized parameter server for GPU machine learning.☆101Updated 7 years ago
- Pytorch implementation of λOpt: Learn to Regularize Recommender Models in Finer Levels, KDD 2019☆53Updated 5 years ago
- Some tensorflow examples☆19Updated 7 years ago
- CS294; AI For Systems and Systems For AI☆224Updated 5 years ago
- Light-weight GPU kernel interface for graph operations☆15Updated 5 years ago
- A distributed logistic regression system based on ps-lite.☆45Updated 8 years ago
- Differentiable Product Quantization for End-to-End Embedding Compression.☆62Updated 2 years ago
- Implementation of Parameter Server using PyTorch communication lib☆42Updated 6 years ago
- A PyTorch implementation of Ranking Distillation☆90Updated 4 years ago
- A Fast Muti-processing BERT-Inference System☆101Updated 2 years ago
- 位置敏感哈希LSH应用用计算向量最大内积和☆11Updated 7 years ago
- this is the release repository of superneurons☆52Updated 4 years ago
- FRED simulator and associated paper☆26Updated 9 years ago
- Asynchronous Stochastic Gradient Descent with Delay Compensation☆21Updated 8 years ago
- Ternary Gradients to Reduce Communication in Distributed Deep Learning (TensorFlow)☆182Updated 6 years ago
- ddl-benchmarks: Benchmarks for Distributed Deep Learning☆37Updated 5 years ago
- Practical low-rank gradient compression for distributed optimization: https://arxiv.org/abs/1905.13727☆147Updated 7 months ago
- A compressed adaptive optimizer for training large-scale deep learning models using PyTorch☆27Updated 5 years ago
- ☆22Updated 6 years ago
- FTPipe and related pipeline model parallelism research.☆41Updated 2 years ago
- A fully adaptive, zero-tuning parameter manager that enables efficient distributed machine learning training☆20Updated 2 years ago
- ☆63Updated 4 years ago
- CUDA Matrix Factorization Library with Stochastic Gradient Descent (SGD)☆71Updated 7 years ago
- ☆43Updated last year
- distributed version of product-nets☆82Updated 5 years ago