The repository for the code of the UltraFastBERT paper
☆519Mar 24, 2024Updated last year
Alternatives and similar repositories for UltraFastBERT
Users that are interested in UltraFastBERT are comparing it to the libraries listed below
Sorting:
- A repository for log-time feedforward networks☆224Apr 9, 2024Updated last year
- Code for the paper "QMoE: Practical Sub-1-Bit Compression of Trillion-Parameter Models".☆279Nov 3, 2023Updated 2 years ago
- Repo for "Monarch Mixer: A Simple Sub-Quadratic GEMM-Based Architecture"☆562Dec 28, 2024Updated last year
- ⚡ Build your chatbot within minutes on your favorite device; offer SOTA compression techniques for LLMs; run LLMs efficiently on Intel Pl…☆2,175Oct 8, 2024Updated last year
- The Truth Is In There: Improving Reasoning in Language Models with Layer-Selective Rank Reduction☆390Jul 9, 2024Updated last year
- Convolutions for Sequence Modeling☆912Jun 13, 2024Updated last year
- some common Huggingface transformers in maximal update parametrization (µP)☆87Mar 14, 2022Updated 3 years ago
- ☆50Mar 14, 2024Updated last year
- Scaling Data-Constrained Language Models☆342Jun 28, 2025Updated 8 months ago
- Official repository for the paper "Approximating Two-Layer Feedforward Networks for Efficient Transformers"☆39Jun 11, 2025Updated 8 months ago
- [ICML 2024] Break the Sequential Dependency of LLM Inference Using Lookahead Decoding☆1,315Mar 6, 2025Updated 11 months ago
- Official PyTorch implementation of QA-LoRA☆145Mar 13, 2024Updated last year
- Official repository for the paper "SwitchHead: Accelerating Transformers with Mixture-of-Experts Attention"☆102Sep 30, 2024Updated last year
- Serving multiple LoRA finetuned LLM as one☆1,145May 8, 2024Updated last year
- Official implementation of Half-Quadratic Quantization (HQQ)☆913Dec 18, 2025Updated 2 months ago
- Accessible large language models via k-bit quantization for PyTorch.☆7,997Updated this week
- Simple, minimal implementation of the Mamba SSM in one file of PyTorch.☆2,921Mar 8, 2024Updated last year
- ModuleFormer is a MoE-based architecture that includes two different types of experts: stick-breaking attention heads and feedforward exp…☆226Sep 18, 2025Updated 5 months ago
- Code for exploring Based models from "Simple linear attention language models balance the recall-throughput tradeoff"☆248Jun 6, 2025Updated 8 months ago
- YaRN: Efficient Context Window Extension of Large Language Models☆1,673Apr 17, 2024Updated last year
- [ICML 2023] Exploring the Benefits of Training Expert Language Models over Instruction Tuning☆98Apr 26, 2023Updated 2 years ago
- RWKV (pronounced RwaKuv) is an RNN with great LLM performance, which can also be directly trained like a GPT transformer (parallelizable)…☆14,375Feb 21, 2026Updated last week
- ☆83Apr 16, 2024Updated last year
- Tools for merging pretrained large language models.☆6,814Jan 26, 2026Updated last month
- Understand and test language model architectures on synthetic tasks.☆254Feb 24, 2026Updated last week
- Code repository for the paper "MrT5: Dynamic Token Merging for Efficient Byte-level Language Models."☆54Sep 25, 2025Updated 5 months ago
- PyTorch compiler that accelerates training and inference. Get built-in optimizations for performance, memory, parallelism, and easily wri…☆1,444Updated this week
- Implementation of the Mamba SSM with hf_integration.☆55Aug 31, 2024Updated last year
- [ICLR 2024] Efficient Streaming Language Models with Attention Sinks☆7,187Jul 11, 2024Updated last year
- Robust recipes to align language models with human and AI preferences☆5,506Sep 8, 2025Updated 5 months ago
- Simple and efficient pytorch-native transformer text generation in <1000 LOC of python.☆6,184Aug 22, 2025Updated 6 months ago
- Fast inference engine for Transformer models☆4,326Feb 4, 2026Updated 3 weeks ago
- Freeing data processing from scripting madness by providing a set of platform-agnostic customizable pipeline processing blocks.☆2,903Updated this week
- S-LoRA: Serving Thousands of Concurrent LoRA Adapters☆1,899Jan 21, 2024Updated 2 years ago
- Sparsity-aware deep learning inference runtime for CPUs☆3,159Jun 2, 2025Updated 9 months ago
- Code for the ICML 2023 paper "SparseGPT: Massive Language Models Can Be Accurately Pruned in One-Shot".☆870Aug 20, 2024Updated last year
- Extend existing LLMs way beyond the original training length with constant memory usage, without retraining☆737Apr 10, 2024Updated last year
- ☆316Jun 21, 2024Updated last year
- Code repository for Black Mamba☆263Feb 8, 2024Updated 2 years ago