bojone / tiger
A Tight-fisted Optimizer
☆47Updated last year
Alternatives and similar repositories for tiger:
Users that are interested in tiger are comparing it to the libraries listed below
- Lion and Adam optimization comparison☆56Updated last year
- A Tight-fisted Optimizer (Tiger), implemented in PyTorch.☆11Updated 6 months ago
- ☆98Updated 10 months ago
- 基于Gated Attention Unit的Transformer模型(尝鲜版)☆97Updated last year
- This is a personal reimplementation of Google's Infini-transformer, utilizing a small 2b model. The project includes both model and train…☆55Updated 9 months ago
- Low-bit optimizers for PyTorch☆125Updated last year
- [EVA ICLR'23; LARA ICML'22] Efficient attention mechanisms via control variates, random features, and importance sampling☆80Updated last year
- [ICML'24] The official implementation of “Rethinking Optimization and Architecture for Tiny Language Models”☆119Updated this week
- [ICML'24 Oral] The official code of "DiJiang: Efficient Large Language Models through Compact Kernelization", a novel DCT-based linear at…☆99Updated 7 months ago
- [EMNLP 2022] Official implementation of Transnormer in our EMNLP 2022 paper - The Devil in Linear Transformer☆58Updated last year
- Converting Mixtral-8x7B to Mixtral-[1~7]x7B☆20Updated 10 months ago
- An Experiment on Dynamic NTK Scaling RoPE☆62Updated last year
- ☆13Updated last year
- # Unified Normalization (ACM MM'22) By Qiming Yang, Kai Zhang, Chaoxiang Lan, Zhi Yang, Zheyang Li, Wenming Tan, Jun Xiao, and Shiliang P…☆34Updated last year
- Lightning Attention-2: A Free Lunch for Handling Unlimited Sequence Lengths in Large Language Models☆226Updated 8 months ago
- ☆31Updated 7 months ago
- Research without Re-search: Maximal Update Parametrization Yields Accurate Loss Prediction across Scales☆31Updated last year
- Contextual Position Encoding but with some custom CUDA Kernels https://arxiv.org/abs/2405.18719☆22Updated 7 months ago
- [ICLR 2024] CLEX: Continuous Length Extrapolation for Large Language Models☆75Updated 10 months ago
- differentiable top-k operator☆21Updated 3 weeks ago
- 🔥 A minimal training framework for scaling FLA models☆24Updated this week
- Ladder Side-Tuning在CLUE上的简单尝试☆19Updated 2 years ago
- 32 times longer context window than vanilla Transformers and up to 4 times longer than memory efficient Transformers.☆44Updated last year
- [ICLR 2023] "Sparse MoE as the New Dropout: Scaling Dense and Self-Slimmable Transformers" by Tianlong Chen*, Zhenyu Zhang*, Ajay Jaiswal…☆48Updated last year
- Code for paper "Patch-Level Training for Large Language Models"☆75Updated 2 months ago
- Longitudinal Evaluation of LLMs via Data Compression☆30Updated 7 months ago
- ICLR2023 - Tailoring Language Generation Models under Total Variation Distance☆21Updated last year
- [ICLR 2024]EMO: Earth Mover Distance Optimization for Auto-Regressive Language Modeling(https://arxiv.org/abs/2310.04691)☆117Updated 10 months ago
- code for Scaling Laws of RoPE-based Extrapolation☆71Updated last year
- A small framework mimics PyTorch using CuPy or NumPy☆27Updated 2 years ago