hkproj / mistral-src-commentedLinks
Reference implementation of Mistral AI 7B v0.1 model.
☆28Updated last year
Alternatives and similar repositories for mistral-src-commented
Users that are interested in mistral-src-commented are comparing it to the libraries listed below
Sorting:
- One click templates for inferencing Language Models☆195Updated 3 weeks ago
- LORA: Low-Rank Adaptation of Large Language Models implemented using PyTorch☆110Updated last year
- ☆40Updated last month
- Notes about "Attention is all you need" video (https://www.youtube.com/watch?v=bCz4OMemCcA)☆293Updated 2 years ago
- Various installation guides for Large Language Models☆70Updated 2 months ago
- From scratch implementation of a vision language model in pure PyTorch☆227Updated last year
- Fully fine-tune large models like Mistral, Llama-2-13B, or Qwen-14B completely for free☆232Updated 8 months ago
- Starter pack for NeurIPS LLM Efficiency Challenge 2023.☆125Updated last year
- A set of scripts and notebooks on LLM finetunning and dataset creation☆110Updated 9 months ago
- LLaMA 2 implemented from scratch in PyTorch☆337Updated last year
- Fine-tune an LLM to perform batch inference and online serving.☆112Updated last month
- customizable template GPT code designed for easy novel architecture experimentation☆26Updated 3 months ago
- Notes on quantization in neural networks☆89Updated last year
- Complete implementation of Llama2 with/without KV cache & inference 🚀☆47Updated last year
- Notes on the Mistral AI model☆19Updated last year
- Banishing LLM Hallucinations Requires Rethinking Generalization☆276Updated last year
- An extension of the nanoGPT repository for training small MOE models.☆160Updated 4 months ago
- ☆54Updated 5 months ago
- ML algorithms implementations that are good for learning the underlying principles☆24Updated 7 months ago
- Building a 2.3M-parameter LLM from scratch with LLaMA 1 architecture.☆180Updated last year
- An overview of GRPO & DeepSeek-R1 Training with Open Source GRPO Model Fine Tuning☆34Updated last month
- Q-GaLore: Quantized GaLore with INT4 Projection and Layer-Adaptive Low-Rank Gradients.☆198Updated 11 months ago
- Toolkit for attaching, training, saving and loading of new heads for transformer models☆282Updated 4 months ago
- Micro Llama is a small Llama based model with 300M parameters trained from scratch with $500 budget☆153Updated last year
- LLaMA 3 is one of the most promising open-source model after Mistral, we will recreate it's architecture in a simpler manner.☆171Updated 10 months ago
- A compact LLM pretrained in 9 days by using high quality data☆318Updated 3 months ago
- Video+code lecture on building nanoGPT from scratch☆69Updated last year
- a simplified version of Meta's Llama 3 model to be used for learning☆41Updated last year
- Set of scripts to finetune LLMs☆37Updated last year
- This project showcases an LLMOps pipeline that fine-tunes a small-size LLM model to prepare for the outage of the service LLM.☆307Updated 3 months ago