corl-team / rebased
Official implementation of the paper "Linear Transformers with Learnable Kernel Functions are Better In-Context Models"
☆156Updated 8 months ago
Related projects ⓘ
Alternatives and complementary repositories for rebased
- Tree Attention: Topology-aware Decoding for Long-Context Attention on GPU clusters☆104Updated last month
- ☆20Updated 3 months ago
- Effective LLM Alignment Toolkit☆83Updated last week
- PyTorch implementation of models from the Zamba2 series.☆158Updated 2 months ago
- The simplest, fastest repository for training/finetuning medium-sized GPTs.☆83Updated last week
- ☆53Updated 9 months ago
- Modified Arena-Hard-Auto LLM evaluation toolkit with an emphasis on Russian language☆24Updated last week
- σ-GPT: A New Approach to Autoregressive Models☆59Updated 2 months ago
- Explorations into the proposal from the paper "Grokfast, Accelerated Grokking by Amplifying Slow Gradients"☆84Updated 2 months ago
- ☆40Updated this week
- Pytorch implementation of the PEER block from the paper, Mixture of A Million Experts, by Xu Owen He at Deepmind☆111Updated 2 months ago
- ☆26Updated last month
- Token Omission Via Attention☆119Updated 3 weeks ago
- ☆30Updated last week
- Understand and test language model architectures on synthetic tasks.☆161Updated 6 months ago
- Best practices & guides on how to write distributed pytorch training code☆278Updated this week
- ☆76Updated 6 months ago
- 2D Positional Embeddings for Webpage Structural Understanding 🦙👀☆93Updated 2 months ago
- A single repo with all scripts and utils to train / fine-tune the Mamba model with or without FIM☆49Updated 7 months ago
- Framework for processing and filtering datasets☆25Updated 3 months ago
- An efficent implementation of the method proposed in "The Era of 1-bit LLMs"☆154Updated 3 weeks ago
- ☆49Updated 7 months ago
- Video+code lecture on building nanoGPT from scratch☆64Updated 4 months ago
- Code for exploring Based models from "Simple linear attention language models balance the recall-throughput tradeoff"☆212Updated 2 months ago
- ☆292Updated 4 months ago
- Q-GaLore: Quantized GaLore with INT4 Projection and Layer-Adaptive Low-Rank Gradients.☆171Updated 3 months ago
- Fully fine-tune large models like Mistral, Llama-2-13B, or Qwen-14B completely for free☆219Updated last week
- Code for Adam-mini: Use Fewer Learning Rates To Gain More https://arxiv.org/abs/2406.16793☆322Updated last week