Long context evaluation for large language models
☆228Mar 3, 2025Updated last year
Alternatives and similar repositories for hash-hop
Users that are interested in hash-hop are comparing it to the libraries listed below
Sorting:
- ☆11Oct 11, 2023Updated 2 years ago
- Engineering the state of RNN language models (Mamba, RWKV, etc.)☆32May 25, 2024Updated last year
- Stick-breaking attention☆62Jul 1, 2025Updated 8 months ago
- [NeurIPS 2023] Sparse Modular Activation for Efficient Sequence Modeling☆40Dec 2, 2023Updated 2 years ago
- Some preliminary explorations of Mamba's context scaling.☆218Feb 8, 2024Updated 2 years ago
- Reference implementation of "Softmax Attention with Constant Cost per Token" (Heinsen, 2024)☆24Jun 6, 2024Updated last year
- Parallel Associative Scan for Language Models☆18Jan 8, 2024Updated 2 years ago
- Official PyTorch Implementation of the Longhorn Deep State Space Model☆56Dec 4, 2024Updated last year
- A repository for research on medium sized language models.☆78May 23, 2024Updated last year
- A fusion of a linear layer and a cross entropy loss, written for pytorch in triton.☆75Aug 2, 2024Updated last year
- ☆20Nov 28, 2024Updated last year
- ☆12Nov 13, 2024Updated last year
- Accelerated First Order Parallel Associative Scan☆195Jan 7, 2026Updated 2 months ago
- Code for ICLR 2025 Paper "What is Wrong with Perplexity for Long-context Language Modeling?"☆110Oct 11, 2025Updated 4 months ago
- FlexAttention w/ FlashAttention3 Support☆27Oct 5, 2024Updated last year
- A toolkit for scaling law research ⚖☆57Jan 27, 2025Updated last year
- ☆19Dec 4, 2025Updated 3 months ago
- Efficient PScan implementation in PyTorch☆17Jan 2, 2024Updated 2 years ago
- ☆36Feb 26, 2024Updated 2 years ago
- Long Context Extension and Generalization in LLMs☆63Sep 21, 2024Updated last year
- An efficient implementation of the NSA (Native Sparse Attention) kernel☆129Jun 24, 2025Updated 8 months ago
- PyTorch implementation of models from the Zamba2 series.☆187Jan 23, 2025Updated last year
- Awesome Triton Resources☆39Apr 27, 2025Updated 10 months ago
- A large-scale RWKV v7(World, PRWKV, Hybrid-RWKV) inference. Capable of inference by combining multiple states(Pseudo MoE). Easy to deploy…☆47Oct 21, 2025Updated 4 months ago
- Code for the paper: https://arxiv.org/pdf/2309.06979.pdf☆21Jul 29, 2024Updated last year
- Recursive Bayesian Networks☆11May 11, 2025Updated 9 months ago
- Neural Algorithmic Reasoning Tutorial☆12Dec 21, 2022Updated 3 years ago
- ☆13Dec 15, 2025Updated 2 months ago
- Code for the paper "Stack Attention: Improving the Ability of Transformers to Model Hierarchical Patterns"☆18Mar 15, 2024Updated last year
- The simplest implementation of recent Sparse Attention patterns for efficient LLM inference.☆91Jul 17, 2025Updated 7 months ago
- ☆58Jul 9, 2024Updated last year
- Ring attention implementation with flash attention☆987Sep 10, 2025Updated 5 months ago
- HGRN2: Gated Linear RNNs with State Expansion☆56Aug 20, 2024Updated last year
- ☆44Nov 1, 2025Updated 4 months ago
- ☆53May 20, 2024Updated last year
- Open-sourcing code associated with the AAAI-25 paper "On the Expressiveness and Length Generalization of Selective State-Space Models on …☆16Sep 18, 2025Updated 5 months ago
- [EMNLP 2023] Official implementation of the algorithm ETSC: Exact Toeplitz-to-SSM Conversion our EMNLP 2023 paper - Accelerating Toeplitz…☆14Oct 17, 2023Updated 2 years ago
- CUDA and Triton implementations of Flash Attention with SoftmaxN.☆73May 26, 2024Updated last year
- ☆124May 28, 2024Updated last year