hyhieu / easy_pybindLinks
☆32Updated last year
Alternatives and similar repositories for easy_pybind
Users that are interested in easy_pybind are comparing it to the libraries listed below
Sorting:
- ☆89Updated last year
- Learning about CUDA by writing PTX code.☆145Updated last year
- ☆91Updated last year
- Tree Attention: Topology-aware Decoding for Long-Context Attention on GPU clusters☆130Updated 11 months ago
- Fast and memory efficient PyTorch implementation of the Perceiver with FlashAttention.☆30Updated 11 months ago
- ☆252Updated 4 months ago
- FlashRNN - Fast RNN Kernels with I/O Awareness☆103Updated last week
- Experimental GPU language with meta-programming☆23Updated last year
- Experiment of using Tangent to autodiff triton☆80Updated last year
- ☆174Updated last year
- train with kittens!☆63Updated last year
- Flash Attention Triton kernel with support for second-order derivatives☆106Updated last week
- σ-GPT: A New Approach to Autoregressive Models☆68Updated last year
- ☆53Updated last year
- Focused on fast experimentation and simplicity☆75Updated 10 months ago
- Landing repository for the paper "Softpick: No Attention Sink, No Massive Activations with Rectified Softmax"☆85Updated last month
- Implementation of the proposed Spline-Based Transformer from Disney Research☆104Updated 11 months ago
- Quick implementation of nGPT, learning entirely on the hypersphere, from NvidiaAI☆291Updated 4 months ago
- CUDA-L1: Improving CUDA Optimization via Contrastive Reinforcement Learning☆195Updated last week
- Normalized Transformer (nGPT)☆192Updated 11 months ago
- ring-attention experiments☆155Updated last year
- Minimal (400 LOC) implementation Maximum (multi-node, FSDP) GPT training☆132Updated last year
- Personal solutions to the Triton Puzzles☆20Updated last year
- High-Performance SGEMM on CUDA devices☆107Updated 9 months ago
- Samples of good AI generated CUDA kernels☆91Updated 5 months ago
- FlexAttention w/ FlashAttention3 Support☆27Updated last year
- ☆28Updated last month
- WIP☆93Updated last year
- Quantized LLM training in pure CUDA/C++.☆209Updated this week
- JAX bindings for Flash Attention v2☆97Updated last week