redotvideo / mamba-chat
Mamba-Chat: A chat LLM based on the state-space model architecture π
β911Updated 8 months ago
Related projects β
Alternatives and complementary repositories for mamba-chat
- Official implementation of "Samba: Simple Hybrid State Space Models for Efficient Unlimited Context Language Modeling"β804Updated 3 months ago
- [ICML'24 Spotlight] LLM Maybe LongLM: Self-Extend LLM Context Window Without Tuningβ615Updated 5 months ago
- A family of open-sourced Mixture-of-Experts (MoE) Large Language Modelsβ1,394Updated 8 months ago
- Implementation of the training framework proposed in Self-Rewarding Language Model, from MetaAIβ1,338Updated 7 months ago
- ReFT: Representation Finetuning for Language Modelsβ1,164Updated 2 weeks ago
- Extend existing LLMs way beyond the original training length with constant memory usage, without retraining