Ying1123 / awesome-neural-symbolic
A list of awesome neural symbolic papers.
☆47Updated 2 years ago
Alternatives and similar repositories for awesome-neural-symbolic:
Users that are interested in awesome-neural-symbolic are comparing it to the libraries listed below
- Framework to reduce autotune overhead to zero for well known deployments.☆65Updated last week
- NAACL '24 (Best Demo Paper RunnerUp) / MlSys @ NeurIPS '23 - RedCoast: A Lightweight Tool to Automate Distributed Training and Inference☆64Updated 4 months ago
- Tritonbench is a collection of PyTorch custom operators with example inputs to measure their performance.☆111Updated last week
- TritonBench: Benchmarking Large Language Model Capabilities for Generating Triton Operators☆35Updated last month
- Simple and efficient pytorch-native transformer training and inference (batched)☆73Updated last year
- SMT-LIB benchmarks for shape computations from deep learning models in PyTorch☆17Updated 2 years ago
- The source code of our work "Prepacking: A Simple Method for Fast Prefilling and Increased Throughput in Large Language Models" [AISTATS …☆59Updated 6 months ago
- SparseTIR: Sparse Tensor Compiler for Deep Learning☆135Updated 2 years ago
- ☆36Updated 7 months ago
- ☆103Updated 8 months ago
- Triton-based implementation of Sparse Mixture of Experts.☆210Updated 4 months ago
- Python package for rematerialization-aware gradient checkpointing☆24Updated last year
- Personal solutions to the Triton Puzzles☆18Updated 9 months ago
- Experiment of using Tangent to autodiff triton☆78Updated last year
- Benchmark PyTorch Custom Operators☆14Updated last year
- TileFusion is an experimental C++ macro kernel template library that elevates the abstraction level in CUDA C for tile processing.☆82Updated this week
- Repository of the paper "Accelerating Transformer Inference for Translation via Parallel Decoding"☆116Updated last year
- Training language models to make programs faster☆87Updated last year
- A schedule language for large model training☆146Updated 10 months ago
- CUDA and Triton implementations of Flash Attention with SoftmaxN.☆69Updated 10 months ago
- Memory Optimizations for Deep Learning (ICML 2023)☆64Updated last year
- ☆18Updated last year
- ☆13Updated 4 months ago
- Sparsity support for PyTorch☆34Updated last month
- Make triton easier☆47Updated 10 months ago
- PyTorch compilation tutorial covering TorchScript, torch.fx, and Slapo☆18Updated 2 years ago
- Boosting 4-bit inference kernels with 2:4 Sparsity☆72Updated 7 months ago
- ☆38Updated last year
- GPTQ inference TVM kernel☆38Updated last year
- extensible collectives library in triton☆85Updated 3 weeks ago