dhcode-cpp / online-softmaxLinks
simplest online-softmax notebook for explain Flash Attention
☆10Updated 5 months ago
Alternatives and similar repositories for online-softmax
Users that are interested in online-softmax are comparing it to the libraries listed below
Sorting:
- ☆52Updated last year
- A MoE impl for PyTorch, [ATC'23] SmartMoE☆63Updated last year
- ☆11Updated last year
- 模型压缩的小白入门教程☆22Updated 11 months ago
- qwen-nsa☆66Updated last month
- [ACL 2024] A novel QAT with Self-Distillation framework to enhance ultra low-bit LLMs.☆114Updated last year
- ☆76Updated last month
- 天池 NVIDIA TensorRT Hackathon 2023 —— 生成式AI模型优化赛 初赛第三名方案☆49Updated last year
- ☢️ TensorRT 2023复赛——基于TensorRT-LLM的Llama模型推断加速优化☆48Updated last year
- [ICML 2023] SmoothQuant: Accurate and Efficient Post-Training Quantization for Large Language Models☆23Updated last year
- TensorRT-in-Action 是一个 GitHub 代码库,提供了使用 TensorRT 的代码示例,并有对应 Jupyter Notebook。☆16Updated 2 years ago
- from MHA, MQA, GQA to MLA by 苏剑林, with code☆19Updated 3 months ago
- ☆131Updated last month
- The Official Implementation of Ada-KV: Optimizing KV Cache Eviction by Adaptive Budget Allocation for Efficient LLM Inference☆76Updated 4 months ago
- ☆24Updated last year
- ☆79Updated last year
- Implementation of FlashAttention in PyTorch☆150Updated 4 months ago
- Transformer related optimization, including BERT, GPT☆17Updated last year
- 使用 CUDA C++ 实现的 llama 模型推理框架☆57Updated 6 months ago
- More Tokens, Lower Precision: Towards the Optimal Token-Precision Trade-off in KV Cache Compression☆11Updated 4 months ago
- Manages vllm-nccl dependency☆17Updated last year
- Train llm (bloom, llama, baichuan2-7b, chatglm3-6b) with deepspeed pipeline mode. Faster than zero/zero++/fsdp.☆95Updated last year
- DeepSeek Native Sparse Attention pytorch implementation☆70Updated 3 months ago
- ☆90Updated last year
- Efficient Mixture of Experts for LLM Paper List☆68Updated 5 months ago
- lightweighted deep learning inference service framework☆39Updated 3 years ago
- This is a personal reimplementation of Google's Infini-transformer, utilizing a small 2b model. The project includes both model and train…☆56Updated last year
- An easy-to-use package for implementing SmoothQuant for LLMs☆99Updated 2 months ago
- Decoding Attention is specially optimized for MHA, MQA, GQA and MLA using CUDA core for the decoding stage of LLM inference.☆36Updated 2 months ago
- simplify >2GB large onnx model☆57Updated 6 months ago