ruimalheiro / training-custom-llamaView on GitHub
Llama-style transformer in PyTorch with multi-node / multi-GPU training. Includes pretraining, fine-tuning, DPO, LoRA, and knowledge distillation. Scripts for dataset mixing and training from scratch.
21Mar 6, 2026Updated this week

Alternatives and similar repositories for training-custom-llama

Users that are interested in training-custom-llama are comparing it to the libraries listed below

Sorting:

Are these results useful?