Lightning-AI / lightning-HabanaLinks
Lightning support for Intel Habana accelerators.
☆27Updated last month
Alternatives and similar repositories for lightning-Habana
Users that are interested in lightning-Habana are comparing it to the libraries listed below
Sorting:
- ☆108Updated last year
- JORA: JAX Tensor-Parallel LoRA Library (ACL 2024)☆33Updated last year
- Triton Implementation of HyperAttention Algorithm☆48Updated last year
- some common Huggingface transformers in maximal update parametrization (µP)☆80Updated 3 years ago
- 삼각형의 실전! Triton☆16Updated last year
- Simple and efficient pytorch-native transformer training and inference (batched)☆75Updated last year
- ☆44Updated last year
- Awesome Triton Resources☆28Updated last month
- Code for the examples presented in the talk "Training a Llama in your backyard: fine-tuning very large models on consumer hardware" given…☆14Updated last year
- Randomized Positional Encodings Boost Length Generalization of Transformers☆82Updated last year
- ☆47Updated 9 months ago
- The source code of our work "Prepacking: A Simple Method for Fast Prefilling and Increased Throughput in Large Language Models" [AISTATS …☆59Updated 7 months ago
- Experiment of using Tangent to autodiff triton☆79Updated last year
- Implementation of Infini-Transformer in Pytorch☆111Updated 5 months ago
- ☆81Updated last year
- Load compute kernels from the Hub☆139Updated this week
- The simplest implementation of recent Sparse Attention patterns for efficient LLM inference.☆62Updated 4 months ago
- Implementation of TableFormer, Robust Transformer Modeling for Table-Text Encoding, in Pytorch☆39Updated 3 years ago
- ☆13Updated 3 weeks ago
- A block oriented training approach for inference time optimization.☆33Updated 9 months ago
- Official implementation of "The Sparse Frontier: Sparse Attention Trade-offs in Transformer LLMs"☆32Updated last month
- ☆80Updated last year
- ArcticTraining is a framework designed to simplify and accelerate the post-training process for large language models (LLMs)☆105Updated this week
- Repository for Sparse Finetuning of LLMs via modified version of the MosaicML llmfoundry☆42Updated last year
- See details in https://github.com/pytorch/xla/blob/r1.12/torch_xla/distributed/fsdp/README.md☆24Updated 2 years ago
- Prototype routines for GPU quantization written using PyTorch.☆21Updated 3 months ago
- Code and Configs for Asynchronous RLHF: Faster and More Efficient RL for Language Models☆54Updated last month
- CUDA implementation of autoregressive linear attention, with all the latest research findings☆44Updated 2 years ago
- ☆14Updated 10 months ago
- ☆37Updated last year