uygarkurt / Llama-3-PyTorchLinks
☆38Updated 10 months ago
Alternatives and similar repositories for Llama-3-PyTorch
Users that are interested in Llama-3-PyTorch are comparing it to the libraries listed below
Sorting:
- ☆45Updated 5 months ago
 - Unofficial implementation of https://arxiv.org/pdf/2407.14679☆50Updated last year
 - Complete implementation of Llama2 with/without KV cache & inference 🚀☆48Updated last year
 - LORA: Low-Rank Adaptation of Large Language Models implemented using PyTorch☆117Updated 2 years ago
 - ☆45Updated 5 months ago
 - Distributed training (multi-node) of a Transformer model☆86Updated last year
 - A repository of Python scripts to scrape code contents of the public repositories of `huggingface`.☆52Updated last year
 - ☆210Updated 10 months ago
 - Accelerate Model Training with PyTorch 2.X, published by Packt☆47Updated last month
 - Various installation guides for Large Language Models☆75Updated 6 months ago
 - Experimenting with small language models☆74Updated last year
 - LLaMA 2 implemented from scratch in PyTorch☆358Updated 2 years ago
 - GPU Kernels☆203Updated 6 months ago
 - A repository to unravel the language of GPUs, making their kernel conversations easy to understand☆194Updated 5 months ago
 - LLaMA 3 is one of the most promising open-source model after Mistral, we will recreate it's architecture in a simpler manner.☆188Updated last year
 - Implementation of a GPT-4o like Multimodal from Scratch using Python☆73Updated 6 months ago
 - Learn the building blocks of how to build gpt-oss from scratch☆91Updated last month
 - minimal GRPO implementation from scratch☆98Updated 7 months ago
 - Collection of autoregressive model implementation☆86Updated 6 months ago
 - Simple repository for training small reasoning models☆44Updated 8 months ago
 - ☆88Updated last year
 - Recreating PyTorch from scratch (C/C++, CUDA, NCCL and Python, with multi-GPU support and automatic differentiation!)☆160Updated last year
 - making the official triton tutorials actually comprehensible☆57Updated 2 months ago
 - Micro Llama is a small Llama based model with 300M parameters trained from scratch with $500 budget☆161Updated 2 months ago
 - ☆34Updated last year
 - A single repo with all scripts and utils to train / fine-tune the Mamba model with or without FIM☆59Updated last year
 - 👷 Build compute kernels☆163Updated last week
 - An overview of GRPO & DeepSeek-R1 Training with Open Source GRPO Model Fine Tuning☆37Updated 5 months ago
 - Notes about "Attention is all you need" video (https://www.youtube.com/watch?v=bCz4OMemCcA)☆319Updated 2 years ago
 - The simplest, fastest repository for training/finetuning medium-sized xLSTMs.☆41Updated last year