FareedKhan-dev / train-llama4Links
Building LLaMA 4 MoE from Scratch
☆64Updated 5 months ago
Alternatives and similar repositories for train-llama4
Users that are interested in train-llama4 are comparing it to the libraries listed below
Sorting:
- minimal GRPO implementation from scratch☆98Updated 6 months ago
- Implementation of a GPT-4o like Multimodal from Scratch using Python☆72Updated 6 months ago
- Maximizing the Performance of a Simple RAG using RL☆81Updated 6 months ago
- LLaMA 3 is one of the most promising open-source model after Mistral, we will recreate it's architecture in a simpler manner.☆186Updated last year
- From scratch implementation of a vision language model in pure PyTorch☆243Updated last year
- Building a 2.3M-parameter LLM from scratch with LLaMA 1 architecture.☆186Updated last year
- ☆45Updated 5 months ago
- Composition of Multimodal Language Models From Scratch☆15Updated last year
- A Straightforward, Step-by-Step Implementation of a Video Diffusion Model☆59Updated last month
- An overview of GRPO & DeepSeek-R1 Training with Open Source GRPO Model Fine Tuning☆37Updated 4 months ago
- First-principle implementations of groundbreaking AI algorithms using a wide range of deep learning frameworks, accompanied by supporting…☆177Updated 2 months ago
- ☆95Updated 6 months ago
- Parameter-efficient finetuning script for Phi-3-vision, the strong multimodal language model by Microsoft.☆58Updated last year
- Fine tune Gemma 3 on an object detection task☆85Updated 2 months ago
- An NVIDIA AI Workbench example project for fine-tuning a Nemotron-3 8B model☆54Updated last year
- Lightweight toolkit package to train and fine-tune 1.58bit Language models☆90Updated 4 months ago
- A minimal implementation of LLaVA-style VLM with interleaved image & text & video processing ability.☆96Updated 9 months ago
- (ICCV 2025) OCR Hinders RAG: Evaluating the Cascading Impact of OCR on Retrieval-Augmented Generation☆87Updated 3 months ago
- [ICLR'25] ApolloMoE: Efficiently Democratizing Medical LLMs for 50 Languages via a Mixture of Language Family Experts☆49Updated 10 months ago
- Distributed training (multi-node) of a Transformer model☆84Updated last year
- Fine-Tuning Llama3-8B LLM in a multi-GPU environment using DeepSpeed☆18Updated last year
- Utils for Unsloth https://github.com/unslothai/unsloth☆153Updated this week
- Inference, Fine Tuning and many more recipes with Gemma family of models☆269Updated 2 months ago
- nanoGRPO is a lightweight implementation of Group Relative Policy Optimization (GRPO)☆121Updated 5 months ago
- Easy to use, High Performant Knowledge Distillation for LLMs☆93Updated 5 months ago
- Notes and commented code for RLHF (PPO)☆110Updated last year
- EvolKit is an innovative framework designed to automatically enhance the complexity of instructions used for fine-tuning Large Language M…☆240Updated 11 months ago
- Model compression toolkit engineered for enhanced usability, comprehensiveness, and efficiency.☆136Updated last week
- RL significantly the reasoning capability of Qwen2.5-1.5B-Instruct☆30Updated 7 months ago
- Tina: Tiny Reasoning Models via LoRA☆290Updated 2 weeks ago