Lightning-AI / lightning-Habana
Lightning support for Intel Habana accelerators.
☆26Updated last month
Alternatives and similar repositories for lightning-Habana:
Users that are interested in lightning-Habana are comparing it to the libraries listed below
- Implementation of Infini-Transformer in Pytorch☆109Updated 2 months ago
- JORA: JAX Tensor-Parallel LoRA Library (ACL 2024)☆32Updated 10 months ago
- The source code of our work "Prepacking: A Simple Method for Fast Prefilling and Increased Throughput in Large Language Models"☆59Updated 5 months ago
- Train, tune, and infer Bamba model☆86Updated last month
- Triton Implementation of HyperAttention Algorithm☆47Updated last year
- Prototype routines for GPU quantization written using PyTorch.☆19Updated 3 weeks ago
- ☆94Updated 9 months ago
- ☆47Updated 6 months ago
- Code for studying the super weight in LLM☆91Updated 3 months ago
- ☆37Updated 11 months ago
- Implementation of TableFormer, Robust Transformer Modeling for Table-Text Encoding, in Pytorch☆37Updated 2 years ago
- Code for NeurIPS LLM Efficiency Challenge☆57Updated 11 months ago
- One Initialization to Rule them All: Fine-tuning via Explained Variance Adaptation☆37Updated 4 months ago
- ☆80Updated last year
- Large Scale Distributed Model Training strategy with Colossal AI and Lightning AI☆57Updated last year
- A fast implementation of T5/UL2 in PyTorch using Flash Attention☆83Updated last month
- Experiment of using Tangent to autodiff triton☆76Updated last year
- DPO, but faster 🚀☆40Updated 3 months ago
- 삼각형의 실전! Triton☆15Updated last year
- Large scale 4D parallelism pre-training for 🤗 transformers in Mixture of Experts *(still work in progress)*☆81Updated last year
- The simplest implementation of recent Sparse Attention patterns for efficient LLM inference.☆58Updated last month
- Randomized Positional Encodings Boost Length Generalization of Transformers☆79Updated 11 months ago
- Code and Configs for Asynchronous RLHF: Faster and More Efficient RL for Language Models☆32Updated 3 months ago
- Some personal experiments around routing tokens to different autoregressive attention, akin to mixture-of-experts☆116Updated 4 months ago
- Simple and efficient pytorch-native transformer training and inference (batched)☆68Updated 11 months ago
- Implementation of a Light Recurrent Unit in Pytorch☆47Updated 5 months ago
- ☆43Updated last year
- Quickest way to share everything about your research within a single app☆17Updated last year