Lightning-AI / lightning-Habana
Lightning support for Intel Habana accelerators.
☆27Updated 3 weeks ago
Alternatives and similar repositories for lightning-Habana:
Users that are interested in lightning-Habana are comparing it to the libraries listed below
- The source code of our work "Prepacking: A Simple Method for Fast Prefilling and Increased Throughput in Large Language Models" [AISTATS …☆59Updated 6 months ago
- Implementation of Infini-Transformer in Pytorch☆110Updated 3 months ago
- Code for NeurIPS LLM Efficiency Challenge☆57Updated last year
- Implementation of TableFormer, Robust Transformer Modeling for Table-Text Encoding, in Pytorch☆37Updated 3 years ago
- Simple and efficient pytorch-native transformer training and inference (batched)☆73Updated last year
- JORA: JAX Tensor-Parallel LoRA Library (ACL 2024)☆33Updated last year
- ☆101Updated 11 months ago
- 삼각형의 실전! Triton☆16Updated last year
- Load compute kernels from the Hub☆115Updated this week
- Triton Implementation of HyperAttention Algorithm☆47Updated last year
- ☆47Updated 7 months ago
- ☆81Updated last year
- Experiment of using Tangent to autodiff triton☆78Updated last year
- Randomized Positional Encodings Boost Length Generalization of Transformers☆80Updated last year
- Awesome Triton Resources☆26Updated 3 weeks ago
- ☆37Updated last year
- Large Scale Distributed Model Training strategy with Colossal AI and Lightning AI☆57Updated last year
- some common Huggingface transformers in maximal update parametrization (µP)☆80Updated 3 years ago
- Transformers components but in Triton☆32Updated last month
- Repository for Sparse Finetuning of LLMs via modified version of the MosaicML llmfoundry☆40Updated last year
- Elixir: Train a Large Language Model on a Small GPU Cluster☆14Updated last year
- Large scale 4D parallelism pre-training for 🤗 transformers in Mixture of Experts *(still work in progress)*☆82Updated last year
- ☆47Updated 5 months ago
- ☆29Updated 2 years ago
- Tritonbench is a collection of PyTorch custom operators with example inputs to measure their performance.☆116Updated this week
- DPO, but faster 🚀☆41Updated 4 months ago
- ☆48Updated last year
- ☆43Updated last year
- ☆78Updated 9 months ago
- ☆125Updated last year