ddidacus / llama-titansLinks
Adaptation of titans-pytorch to llama models on HF
☆22Updated 8 months ago
Alternatives and similar repositories for llama-titans
Users that are interested in llama-titans are comparing it to the libraries listed below
Sorting:
- Official implementation of Phi-Mamba. A MOHAWK-distilled model (Transformers to SSMs: Distilling Quadratic Knowledge to Subquadratic Mode…☆116Updated last year
- Official implementation of the transformer (TF) architecture suggested in a paper entitled "Looped Transformers as Programmable Computers…☆29Updated 2 years ago
- Official Code Repository for the paper "Key-value memory in the brain"☆29Updated 9 months ago
- Official PyTorch Implementation of the Longhorn Deep State Space Model☆56Updated 11 months ago
- ☆33Updated last year
- Parallelizing non-linear sequential models over the sequence length☆55Updated 5 months ago
- Official repo of paper LM2☆46Updated 9 months ago
- ☆33Updated last year
- Mixture of A Million Experts☆50Updated last year
- The official github repo for "Diffusion Language Models are Super Data Learners".☆205Updated 3 weeks ago
- ☆88Updated last year
- ☆36Updated 8 months ago
- Official Pytorch Implementation of "The Curse of Depth in Large Language Models" by Wenfang Sun, Xinyuan Song, Pengxiang Li, Lu Yin,Yefen…☆62Updated last month
- Code for NeurIPS 2024 Spotlight: "Scaling Laws and Compute-Optimal Training Beyond Fixed Training Durations"☆85Updated last year
- Stick-breaking attention☆61Updated 4 months ago
- ☆23Updated last year
- ☆52Updated 8 months ago
- ☆47Updated last month
- [NeurIPS 2024] Official Repository of The Mamba in the Llama: Distilling and Accelerating Hybrid Models☆231Updated last month
- Code for ICML 2024 paper☆34Updated 2 months ago
- [COLM 2025] Code for Paper: Learning Adaptive Parallel Reasoning with Language Models☆132Updated 3 months ago
- Griffin MQA + Hawk Linear RNN Hybrid☆89Updated last year
- ☆53Updated last year
- Implementation of CoLA: Compute-Efficient Pre-Training of LLMs via Low-Rank Activation☆24Updated 9 months ago
- 📄Small Batch Size Training for Language Models☆63Updated last month
- Code for reproducing our paper "Not All Language Model Features Are Linear"☆84Updated last year
- Code for "Reasoning to Learn from Latent Thoughts"☆122Updated 8 months ago
- ☆33Updated 10 months ago
- Some preliminary explorations of Mamba's context scaling.☆217Updated last year
- ☆13Updated 8 months ago