angry-kratos / Simple_Llama3_from_scratch
☆31Updated 10 months ago
Alternatives and similar repositories for Simple_Llama3_from_scratch
Users that are interested in Simple_Llama3_from_scratch are comparing it to the libraries listed below
Sorting:
- ☆30Updated last week
- Notebook and Scripts that showcase running quantized diffusion models on consumer GPUs☆38Updated 6 months ago
- making the official triton tutorials actually comprehensible☆30Updated last month
- Collection of autoregressive model implementation☆85Updated 3 weeks ago
- ☆46Updated last month
- Deep learning library implemented from scratch in numpy. Mixtral, Mamba, LLaMA, GPT, ResNet, and other experiments.☆51Updated last year
- ☆47Updated 8 months ago
- This is the code that went into our practical dive using mamba as information extraction☆54Updated last year
- Complete implementation of Llama2 with/without KV cache & inference 🚀☆46Updated 11 months ago
- Documented and Unit Tested educational Deep Learning framework with Autograd from scratch.☆111Updated last year
- several types of attention modules written in PyTorch for learning purposes☆51Updated 7 months ago
- A single repo with all scripts and utils to train / fine-tune the Mamba model with or without FIM☆54Updated last year
- Flexible Python library providing building blocks (layers) for reproducible Transformers research (Tensorflow ✅, Pytorch 🔜, and Jax 🔜)☆53Updated last year
- Collection of tests performed during the study of the new Kolmogorov-Arnold Neural Networks (KAN)☆40Updated 2 months ago
- Prune transformer layers☆69Updated 11 months ago
- ☆27Updated 10 months ago
- My fork os allen AI's OLMo for educational purposes.☆30Updated 5 months ago
- NanoGPT-speedrunning for the poor T4 enjoyers☆65Updated 3 weeks ago
- Reference implementation of Mistral AI 7B v0.1 model.☆28Updated last year
- working implimention of deepseek MLA☆41Updated 4 months ago
- RAGs: Simple implementations of Retrieval Augmented Generation (RAG) Systems☆104Updated 3 months ago
- Explorations into the proposal from the paper "Grokfast, Accelerated Grokking by Amplifying Slow Gradients"☆99Updated 4 months ago
- Training small GPT-2 style models using Kolmogorov-Arnold networks.☆117Updated 11 months ago
- ☆129Updated 8 months ago
- World's Smallest Vision-Language Model☆27Updated last year
- This code repository contains the code used for my "Optimizing Memory Usage for Training LLMs and Vision Transformers in PyTorch" blog po…☆92Updated last year
- Video+code lecture on building nanoGPT from scratch☆67Updated 11 months ago
- Repo hosting codes and materials related to speeding LLMs' inference using token merging.☆36Updated last year
- RL significantly the reasoning capability of Qwen2.5-1.5B-Instruct☆29Updated 2 months ago
- a simplified version of Google's Gemma model to be used for learning☆24Updated last year