evintunador / minLlama3
a simplified version of Meta's Llama 3 model to be used for learning
☆41Updated 11 months ago
Alternatives and similar repositories for minLlama3:
Users that are interested in minLlama3 are comparing it to the libraries listed below
- Micro Llama is a small Llama based model with 300M parameters trained from scratch with $500 budget☆146Updated last year
- This repository contains an implementation of the LLaMA 2 (Large Language Model Meta AI) model, a Generative Pretrained Transformer (GPT)…☆64Updated last year
- minimal GRPO implementation from scratch☆85Updated last month
- Notes and commented code for RLHF (PPO)☆86Updated last year
- LLaMA 3 is one of the most promising open-source model after Mistral, we will recreate it's architecture in a simpler manner.☆157Updated 8 months ago
- Implementation of the LongRoPE: Extending LLM Context Window Beyond 2 Million Tokens Paper☆135Updated 9 months ago
- Reference implementation of Mistral AI 7B v0.1 model.☆28Updated last year
- Training and Fine-tuning an llm in Python and PyTorch.☆41Updated last year
- a simplified version of Google's Gemma model to be used for learning☆24Updated last year
- Distributed training (multi-node) of a Transformer model☆64Updated last year
- Building a 2.3M-parameter LLM from scratch with LLaMA 1 architecture.☆158Updated 11 months ago
- Lightweight demos for finetuning LLMs. Powered by 🤗 transformers and open-source datasets.☆74Updated 6 months ago
- Direct Preference Optimization from scratch in PyTorch☆90Updated 2 weeks ago
- Tutorial for how to build BERT from scratch☆92Updated 11 months ago
- RL significantly the reasoning capability of Qwen2.5-1.5B-Instruct☆28Updated 2 months ago
- Official repository for ORPO☆448Updated 10 months ago
- [ACL'24] Selective Reflection-Tuning: Student-Selected Data Recycling for LLM Instruction-Tuning☆354Updated 7 months ago
- Notes about LLaMA 2 model☆59Updated last year
- Starter pack for NeurIPS LLM Efficiency Challenge 2023.☆124Updated last year
- Q-GaLore: Quantized GaLore with INT4 Projection and Layer-Adaptive Low-Rank Gradients.☆198Updated 9 months ago
- Scripts for fine-tuning Llama2 via SFT and DPO.☆197Updated last year
- Training small GPT-2 style models using Kolmogorov-Arnold networks.☆116Updated 11 months ago
- [ICLR 2024] Family of LLMs for mathematical reasoning.☆261Updated 4 months ago
- An extension of the nanoGPT repository for training small MOE models.☆131Updated last month
- ☆255Updated last year
- The Truth Is In There: Improving Reasoning in Language Models with Layer-Selective Rank Reduction☆386Updated 9 months ago
- A compact LLM pretrained in 9 days by using high quality data☆310Updated 2 weeks ago
- ☆129Updated 8 months ago
- ☆219Updated 10 months ago
- An overview of GRPO & DeepSeek-R1 Training with Open Source GRPO Model Fine Tuning☆31Updated 2 months ago