junstar92 / parallel_programming_studyLinks
Study parallel programming - CUDA, OpenMP, MPI, Pthread
☆57Updated 2 years ago
Alternatives and similar repositories for parallel_programming_study
Users that are interested in parallel_programming_study are comparing it to the libraries listed below
Sorting:
- ☆53Updated 6 months ago
- ☆43Updated last year
- Study Group of Deep Learning Compiler☆160Updated 2 years ago
- ☆100Updated last year
- CUDA based GPU Programming☆34Updated last year
- A performance library for machine learning applications.☆183Updated last year
- ☆56Updated 2 years ago
- PyTorch CoreSIG☆55Updated 5 months ago
- ☆88Updated last year
- OwLite is a low-code AI model compression toolkit for AI models.☆45Updated 3 weeks ago
- FriendliAI Model Hub☆91Updated 2 years ago
- ☆18Updated 7 months ago
- NEST Compiler☆116Updated 4 months ago
- ☆10Updated last year
- OwLite Examples repository offers illustrative example codes to help users seamlessly compress PyTorch deep learning models and transform…☆10Updated 8 months ago
- Neural Network Acceleration using CPU/GPU, ASIC, FPGA☆60Updated 4 years ago
- ☆25Updated 2 years ago
- Getting GPU Util 99%☆34Updated 4 years ago
- ☆68Updated last week
- Introduction to CUDA programming☆118Updated 8 years ago
- CUDA Hands-on training material by Jack☆53Updated 5 years ago
- Official Github repository for the SIGCOMM '24 paper "Accelerating Model Training in Multi-cluster Environments with Consumer-grade GPUs"☆70Updated 10 months ago
- CUDA 8-bit Tensor Core Matrix Multiplication based on m16n16k16 WMMA API☆30Updated last year
- [AAAI 2025] SMMF: Square-Matricized Momentum Factorization for Memory-Efficient Optimization☆20Updated last week
- FuriosaAI SDK☆45Updated 9 months ago
- Parallel Programming with CUDA @ Hallym University, 2019☆16Updated 5 years ago
- Optimized Parallel Tiled Approach to perform 2D Convolution by taking advantage of the lower latency, higher bandwidth shared memory as w…☆14Updated 7 years ago
- ☆15Updated 3 years ago
- 42dot LLM consists of a pre-trained language model, 42dot LLM-PLM, and a fine-tuned model, 42dot LLM-SFT, which is trained to respond to …☆131Updated last year
- 2018 International Student Car Competition: Autonomous Car SKKU Team. HEVEN☆15Updated 4 years ago