ruimalheiro / training-custom-llama
View external linksLinks

Llama-style transformer in PyTorch with multi-node / multi-GPU training. Includes pretraining, fine-tuning, DPO, LoRA, and knowledge distillation. Scripts for dataset mixing and training from scratch.
21Updated this week

Alternatives and similar repositories for training-custom-llama

Users that are interested in training-custom-llama are comparing it to the libraries listed below

Sorting:

Are these results useful?