Ying1123 / awesome-neural-symbolicLinks
A list of awesome neural symbolic papers.
☆47Updated 2 years ago
Alternatives and similar repositories for awesome-neural-symbolic
Users that are interested in awesome-neural-symbolic are comparing it to the libraries listed below
Sorting:
- NAACL '24 (Best Demo Paper RunnerUp) / MlSys @ NeurIPS '23 - RedCoast: A Lightweight Tool to Automate Distributed Training and Inference☆66Updated 6 months ago
- Python package for rematerialization-aware gradient checkpointing☆25Updated last year
- ☆38Updated last year
- ☆13Updated 6 months ago
- ☆105Updated 9 months ago
- ☆71Updated 3 weeks ago
- GPTQ inference TVM kernel☆40Updated last year
- TritonBench: Benchmarking Large Language Model Capabilities for Generating Triton Operators☆55Updated 3 months ago
- Accelerate LLM preference tuning via prefix sharing with a single line of code☆41Updated last month
- An Attention Superoptimizer☆21Updated 4 months ago
- Repository of the paper "Accelerating Transformer Inference for Translation via Parallel Decoding"☆117Updated last year
- Odysseus: Playground of LLM Sequence Parallelism☆70Updated 11 months ago
- ☆37Updated 9 months ago
- ☆93Updated last week
- Triton-based implementation of Sparse Mixture of Experts.☆217Updated 6 months ago
- (NeurIPS 2022) Automatically finding good model-parallel strategies, especially for complex models and clusters.☆39Updated 2 years ago
- Hydragen: High-Throughput LLM Inference with Shared Prefixes☆36Updated last year
- SparseTIR: Sparse Tensor Compiler for Deep Learning☆138Updated 2 years ago
- PyTorch compilation tutorial covering TorchScript, torch.fx, and Slapo☆18Updated 2 years ago
- Benchmark PyTorch Custom Operators☆14Updated last year
- A curated list for Efficient Large Language Models☆11Updated last year
- A schedule language for large model training☆148Updated 11 months ago
- Simple and efficient pytorch-native transformer training and inference (batched)☆75Updated last year
- TileFusion is an experimental C++ macro kernel template library that elevates the abstraction level in CUDA C for tile processing.☆88Updated this week
- ☆19Updated 2 years ago
- Framework to reduce autotune overhead to zero for well known deployments.☆74Updated 3 weeks ago
- ☆43Updated last year
- GPU operators for sparse tensor operations☆32Updated last year
- Experiment of using Tangent to autodiff triton☆79Updated last year
- Memory Optimizations for Deep Learning (ICML 2023)☆64Updated last year