mli / transformers-benchmarksLinks
real Transformer TeraFLOPS on various GPUs
☆905Updated last year
Alternatives and similar repositories for transformers-benchmarks
Users that are interested in transformers-benchmarks are comparing it to the libraries listed below
Sorting:
- LiBai(李白): A Toolbox for Large-Scale Distributed Parallel Training☆406Updated last week
- SwissArmyTransformer is a flexible and powerful library to develop your own Transformer variants.☆1,080Updated 5 months ago
- Rotary Transformer☆970Updated 3 years ago
- Best practice for training LLaMA models in Megatron-LM☆656Updated last year
- Several simple examples for popular neural network toolkits calling custom CUDA operators.☆1,481Updated 4 years ago
- ☆609Updated last year
- Ongoing research training transformer language models at scale, including: BERT & GPT-2☆2,088Updated 2 months ago
- An easy/swift-to-adapt PyTorch-Lighting template. 套壳模板,简单易用,稍改原来Pytorch代码,即可适配Lightning。You can translate your previous Pytorch code much…☆1,473Updated last year
- Efficient Training (including pre-training and fine-tuning) for Big Models☆596Updated 3 weeks ago
- A fast MoE impl for PyTorch☆1,744Updated 4 months ago
- How to use wandb?☆655Updated last year
- huggingface mirror download☆581Updated 2 months ago
- 更纯粹、更高压缩率的Tokenizer☆480Updated 6 months ago
- Tencent Pre-training framework in PyTorch & Pre-trained Model Zoo☆1,073Updated 10 months ago
- A plug-and-play library for parameter-efficient-tuning (Delta Tuning)☆1,028Updated 9 months ago
- Ongoing research training transformer language models at scale, including: BERT & GPT-2☆1,395Updated last year
- pytorch distribute tutorials☆138Updated this week
- RoFormer V1 & V2 pytorch☆501Updated 3 years ago
- The official repo of Pai-Megatron-Patch for LLM & VLM large scale training developed by Alibaba Cloud.☆1,151Updated last week
- Tuning LLMs with no tears💦; Sample Design Engineering (SDE) for more efficient downstream-tuning.☆1,005Updated last year
- A collection of phenomenons observed during the scaling of big foundation models, which may be developed into consensus, principles, or l…☆282Updated last year
- PyTorch Project Specification.☆679Updated 3 years ago
- Cool Papers - Immersive Paper Discovery☆559Updated 2 weeks ago
- Open Academic Research on Improving LLaMA to SOTA LLM☆1,618Updated last year
- Train a 1B LLM with 1T tokens from scratch by personal☆679Updated last month
- an implementation of transformer, bert, gpt, and diffusion models for learning purposes☆154Updated 8 months ago
- A quickstart and benchmark for pytorch distributed training.☆1,668Updated 10 months ago
- Ascend PyTorch adapter (torch_npu). Mirror of https://gitee.com/ascend/pytorch☆375Updated this week
- Tutel MoE: Optimized Mixture-of-Experts Library, Support DeepSeek FP8/FP4☆842Updated this week
- Official github repo for C-Eval, a Chinese evaluation suite for foundation models [NeurIPS 2023]☆1,741Updated last year