HPC-SJTU / xfoldLinks
Democratizing AlphaFold3: an PyTorch reimplementation to accelerate protein structure prediction
☆21Updated 8 months ago
Alternatives and similar repositories for xfold
Users that are interested in xfold are comparing it to the libraries listed below
Sorting:
- Democratizing AlphaFold3: an PyTorch reimplementation to accelerate protein structure prediction☆54Updated last year
- The Zaychik Power Controller server☆13Updated last year
- OpenCAEPoro for ASC 2024☆38Updated 2 years ago
- 🧪 Ultrafast bisulfite☆38Updated last year
- Repository for HPCGame 1st Problems.☆70Updated last year
- A Throughput-Optimized Pipeline Parallel Inference System for Large Language Models☆46Updated last month
- Documentation for HPC course☆160Updated 7 months ago
- The dataset and baseline code for ASC23 LLM inference optimization challenge.☆32Updated 2 years ago
- HPC-Lab for High Performance Computing course, 2023 Spring , Tsinghua Universit. 高性能计算导论 @ THU.☆24Updated 2 years ago
- Learning TileLang with 10 puzzles!☆56Updated this week
- Intel® Tensor Processing Primitives extension for Pytorch*☆18Updated 2 weeks ago
- ☆284Updated last week
- A disitributed implementation of alphafold3 base on xfold and tpp-pytorch-extension☆12Updated 8 months ago
- Solution of Programming Massively Parallel Processors☆49Updated 2 years ago
- performance engineering☆30Updated last year
- Summary of the Specs of Commonly Used GPUs for Training and Inference of LLM☆73Updated 5 months ago
- [DAC2024] A Holistic Functionalization Approach to Optimizing Imperative Tensor Programs in Deep Learning☆15Updated 2 years ago
- ☆14Updated this week
- gLLM: Global Balanced Pipeline Parallelism System for Distributed LLM Serving with Token Throttling☆53Updated 3 weeks ago
- The repository has collected a batch of noteworthy MLSys bloggers (Algorithms/Systems)☆317Updated last year
- High performance Transformer implementation in C++.☆148Updated last year
- Flash Attention from Scratch on CUDA Ampere☆122Updated 5 months ago
- Summary of some awesome work for optimizing LLM inference☆172Updated 2 months ago
- The official implementation for the intra-stage fusion technique introduced in https://arxiv.org/abs/2409.13221☆29Updated 9 months ago
- LeetGPU Solutions☆98Updated 3 months ago
- Stepwise optimizations of DGEMM on CPU, reaching performance faster than Intel MKL eventually, even under multithreading.☆163Updated 3 years ago
- Wiki fo HPC☆126Updated 6 months ago
- ☆28Updated last week
- NEO is a LLM inference engine built to save the GPU memory crisis by CPU offloading☆81Updated 7 months ago
- Codes & examples for "CUDA - From Correctness to Performance"☆121Updated last year