youkaichao / fast_bpe_tokenizer
fast bpe tokenizer, simple to understand, easy to use
☆25Updated last year
Alternatives and similar repositories for fast_bpe_tokenizer:
Users that are interested in fast_bpe_tokenizer are comparing it to the libraries listed below
- Evaluating LLMs with Dynamic Data☆77Updated last month
- RWKV-7: Surpassing GPT☆80Updated 3 months ago
- An Experiment on Dynamic NTK Scaling RoPE☆62Updated last year
- Low-bit optimizers for PyTorch☆125Updated last year
- Longitudinal Evaluation of LLMs via Data Compression☆32Updated 9 months ago
- continous batching and parallel acceleration for RWKV6☆24Updated 8 months ago
- 32 times longer context window than vanilla Transformers and up to 4 times longer than memory efficient Transformers.☆46Updated last year
- Code for paper "Patch-Level Training for Large Language Models"☆81Updated 3 months ago
- Ouroboros: Speculative Decoding with Large Model Enhanced Drafting (EMNLP 2024 main)☆85Updated 5 months ago
- [ICLR 2024] CLEX: Continuous Length Extrapolation for Large Language Models☆76Updated last year
- SparseGPT + GPTQ Compression of LLMs like LLaMa, OPT, Pythia☆41Updated last year
- Implementation of the paper: "Leave No Context Behind: Efficient Infinite Context Transformers with Infini-attention" from Google in pyTO…☆53Updated last month
- ☆46Updated last year
- Large Scale Distributed Model Training strategy with Colossal AI and Lightning AI☆57Updated last year
- ☆30Updated 9 months ago
- [ICML'24 Oral] The official code of "DiJiang: Efficient Large Language Models through Compact Kernelization", a novel DCT-based linear at…☆99Updated 8 months ago
- Linear Attention Sequence Parallelism (LASP)☆79Updated 9 months ago
- Experiments on speculative sampling with Llama models☆125Updated last year
- ☆34Updated 7 months ago
- Structural Pruning for LLaMA☆54Updated last year
- RWKV in nanoGPT style☆187Updated 9 months ago
- ☆28Updated 11 months ago
- GPT2 implementation in C++ using Ort☆26Updated 4 years ago
- ☆42Updated last year
- Odysseus: Playground of LLM Sequence Parallelism☆66Updated 8 months ago
- BigKnow2022: Bringing Language Models Up to Speed☆14Updated last year
- Positional Skip-wise Training for Efficient Context Window Extension of LLMs to Extremely Length (ICLR 2024)☆204Updated 9 months ago
- QuIP quantization☆51Updated 11 months ago
- ☆115Updated 3 weeks ago