uygarkurt / Llama-3-PyTorchLinks
☆38Updated 10 months ago
Alternatives and similar repositories for Llama-3-PyTorch
Users that are interested in Llama-3-PyTorch are comparing it to the libraries listed below
Sorting:
- LLaMA 3 is one of the most promising open-source model after Mistral, we will recreate it's architecture in a simpler manner.☆190Updated last year
- ☆36Updated last year
- LORA: Low-Rank Adaptation of Large Language Models implemented using PyTorch☆117Updated 2 years ago
- ☆45Updated 6 months ago
- Implementation of a GPT-4o like Multimodal from Scratch using Python☆73Updated 7 months ago
- ☆45Updated 6 months ago
- Learn the building blocks of how to build gpt-oss from scratch☆105Updated 2 months ago
- An overview of GRPO & DeepSeek-R1 Training with Open Source GRPO Model Fine Tuning☆36Updated 6 months ago
- Fine-tune an LLM to perform batch inference and online serving.☆113Updated 6 months ago
- Building a 2.3M-parameter LLM from scratch with LLaMA 1 architecture.☆191Updated last year
- Complete implementation of Llama2 with/without KV cache & inference 🚀☆48Updated last year
- Train LLM on Hugging Face infra☆67Updated 2 weeks ago
- Collection of autoregressive model implementation☆86Updated 7 months ago
- GPU Kernels☆209Updated 7 months ago
- Various installation guides for Large Language Models☆77Updated 7 months ago
- Fine tune Gemma 3 on an object detection task☆89Updated 4 months ago
- Notes about "Attention is all you need" video (https://www.youtube.com/watch?v=bCz4OMemCcA)☆324Updated 2 years ago
- An extension of the nanoGPT repository for training small MOE models.☆215Updated 8 months ago
- A set of scripts and notebooks on LLM finetunning and dataset creation☆111Updated last year
- A repository to unravel the language of GPUs, making their kernel conversations easy to understand☆196Updated 5 months ago
- From scratch implementation of a vision language model in pure PyTorch☆250Updated last year
- Google TPU optimizations for transformers models☆122Updated 10 months ago
- Distributed training (multi-node) of a Transformer model☆87Updated last year
- This is the code that went into our practical dive using mamba as information extraction☆57Updated last year
- ☆136Updated last year
- ☆46Updated 7 months ago
- A collection of lightweight interpretability scripts to understand how LLMs think☆66Updated last week
- 👷 Build compute kernels☆186Updated this week
- Micro Llama is a small Llama based model with 300M parameters trained from scratch with $500 budget☆162Updated 3 months ago
- minimal GRPO implementation from scratch☆99Updated 8 months ago