ruimalheiro / training-custom-llamaView on GitHub
Llama-style transformer in PyTorch with multi-node / multi-GPU training. Includes pretraining, fine-tuning, DPO, LoRA, and knowledge distillation. Scripts for dataset mixing and training from scratch.
22Mar 24, 2026Updated this week

Alternatives and similar repositories for training-custom-llama

Users that are interested in training-custom-llama are comparing it to the libraries listed below. We may earn a commission when you buy through links labeled 'Ad' on this page.

Sorting:

Are these results useful?