josehoras / Knowledge-DistillationLinks
☆10Updated 5 years ago
Alternatives and similar repositories for Knowledge-Distillation
Users that are interested in Knowledge-Distillation are comparing it to the libraries listed below
Sorting:
- Playground for Transformers☆52Updated last year
- ☆132Updated last year
- several types of attention modules written in PyTorch for learning purposes☆53Updated 8 months ago
- Distributed training (multi-node) of a Transformer model☆68Updated last year
- Notebook and Scripts that showcase running quantized diffusion models on consumer GPUs☆38Updated 7 months ago
- ☆38Updated last month
- Notes on quantization in neural networks☆83Updated last year
- LORA: Low-Rank Adaptation of Large Language Models implemented using PyTorch☆105Updated last year
- LoRA and DoRA from Scratch Implementations☆204Updated last year
- A minimal implementation of LLaVA-style VLM with interleaved image & text & video processing ability.☆93Updated 5 months ago
- Making of cuda kernel☆16Updated last week
- A single repo with all scripts and utils to train / fine-tune the Mamba model with or without FIM☆54Updated last year
- Official repository of "LiNeS: Post-training Layer Scaling Prevents Forgetting and Enhances Model Merging"☆26Updated 7 months ago
- Awesome list of papers that extend Mamba to various applications.☆133Updated 2 months ago
- A fork of the PEFT library, supporting Robust Adaptation (RoSA)☆14Updated 9 months ago
- This code repository contains the code used for my "Optimizing Memory Usage for Training LLMs and Vision Transformers in PyTorch" blog po…☆91Updated last year
- Repository containing awesome resources regarding Hugging Face tooling.☆47Updated last year
- Complete implementation of Llama2 with/without KV cache & inference 🚀☆46Updated last year
- Official repository for the paper "SwitchHead: Accelerating Transformers with Mixture-of-Experts Attention"☆97Updated 8 months ago
- Implementation of the paper: "Mixture-of-Depths: Dynamically allocating compute in transformer-based language models"☆94Updated this week
- PyTorch implementation of Retentive Network: A Successor to Transformer for Large Language Models☆14Updated last year
- Implementation of the proposed DeepCrossAttention by Heddes et al at Google research, in Pytorch☆86Updated 3 months ago
- Collection of autoregressive model implementation☆85Updated last month
- Implementation of MoE Mamba from the paper: "MoE-Mamba: Efficient Selective State Space Models with Mixture of Experts" in Pytorch and Ze…☆105Updated 2 months ago
- working implimention of deepseek MLA☆41Updated 4 months ago
- Notes on the Mamba and the S4 model (Mamba: Linear-Time Sequence Modeling with Selective State Spaces)☆168Updated last year
- Pytorch Implementation of the paper: "Learning to (Learn at Test Time): RNNs with Expressive Hidden States"☆24Updated this week
- ☆39Updated last month
- Conference schedule, top papers, and analysis of the data for NeurIPS 2023!☆119Updated last year
- ☆43Updated 4 months ago