todorkolev / titans-market-dataLinks
This project implements the Titans architecture from the paper "Titans: Learning to Memorize at Test Time" for market data prediction.
☆11Updated 11 months ago
Alternatives and similar repositories for titans-market-data
Users that are interested in titans-market-data are comparing it to the libraries listed below
Sorting:
- Titans - Learning to Memorize at Test Time☆55Updated 11 months ago
- Pytorch Implementation of the paper: "Learning to (Learn at Test Time): RNNs with Expressive Hidden States"☆25Updated 3 weeks ago
- [ICML 2024] When Linear Attention Meets Autoregressive Decoding: Towards More Effective and Efficient Linearized Large Language Models☆35Updated last year
- survery of small language models☆17Updated last year
- The official GitHub page for the survey paper "A Survey of RWKV".☆29Updated last year
- A simple implementation of [Mamba: Linear-Time Sequence Modeling with Selective State Spaces](https://arxiv.org/abs/2312.00752)☆22Updated last year
- [ICLR 2025] Drop-Upcycling: Training Sparse Mixture of Experts with Partial Re-initialization☆20Updated 3 months ago
- Implementation of a Hierarchical Mamba as described in the paper: "Hierarchical State Space Models for Continuous Sequence-to-Sequence Mo…☆14Updated last year
- Learning to Skip the Middle Layers of Transformers☆16Updated 5 months ago
- Enable everyone to develop, optimize and deploy AI models natively on everyone's devices.☆13Updated last year
- Official implementation of the ICML 2024 paper RoSA (Robust Adaptation)☆44Updated last year
- [ICLR 2025] Official Pytorch Implementation of "Mix-LN: Unleashing the Power of Deeper Layers by Combining Pre-LN and Post-LN" by Pengxia…☆27Updated 5 months ago
- This repo contains the source code for VB-LoRA: Extreme Parameter Efficient Fine-Tuning with Vector Banks (NeurIPS 2024).☆42Updated last year
- FBI-LLM: Scaling Up Fully Binarized LLMs from Scratch via Autoregressive Distillation☆50Updated 4 months ago
- [ICML 2025] Code for "R2-T2: Re-Routing in Test-Time for Multimodal Mixture-of-Experts"☆17Updated 10 months ago
- [ICLR 2024] This is the official PyTorch implementation of "QLLM: Accurate and Efficient Low-Bitwidth Quantization for Large Language Mod…☆30Updated last year
- This repository contains papers for a comprehensive survey on accelerated generation techniques in Large Language Models (LLMs).☆11Updated last year
- Implementation of MoE Mamba from the paper: "MoE-Mamba: Efficient Selective State Space Models with Mixture of Experts" in Pytorch and Ze…☆119Updated 2 months ago
- The first W4A4KV4 quantized + 50% sparse LLMs!☆20Updated last month
- A repository for DenseSSMs☆88Updated last year
- ☆78Updated 11 months ago
- BESA is a differentiable weight pruning technique for large language models.☆17Updated last year
- [ICLR 2025] Official Code Release for Explaining Modern Gated-Linear RNNs via a Unified Implicit Attention Formulation☆48Updated 10 months ago
- Official PyTorch implementation of CD-MOE☆12Updated 9 months ago
- KVTuner: Sensitivity-Aware Layer-wise Mixed Precision KV Cache Quantization for Efficient and Nearly Lossless LLM Inference☆25Updated 7 months ago
- State Space Models☆71Updated last year
- ☆48Updated last year
- Control LLM☆22Updated 9 months ago
- [CVPR 2023] Castling-ViT: Compressing Self-Attention via Switching Towards Linear-Angular Attention During Vision Transformer Inference☆29Updated last year
- [ICML24] Pruner-Zero: Evolving Symbolic Pruning Metric from scratch for LLMs☆98Updated last year