hscspring / bytepiece-rsLinks
The Bytepiece Tokenizer Implemented in Rust.
☆14Updated last year
Alternatives and similar repositories for bytepiece-rs
Users that are interested in bytepiece-rs are comparing it to the libraries listed below
Sorting:
- Langport is a language model inference service☆93Updated last year
- ☆20Updated last year
- ☆24Updated 5 months ago
- Longitudinal Evaluation of LLMs via Data Compression☆33Updated last year
- RWKV models and examples powered by candle.☆19Updated 7 months ago
- A single-binary, GPU-accelerated LLM server (HTTP and WebSocket API) written in Rust☆79Updated last year
- RWKV in nanoGPT style☆193Updated last year
- This is a text generation method which returns a generator, streaming out each token in real-time during inference, based on Huggingface/…☆97Updated last year
- implement llava using candle☆15Updated last year
- A high-performance constrained decoding engine based on context free grammar in Rust☆55Updated 4 months ago
- 3x Faster Inference; Unofficial implementation of EAGLE Speculative Decoding☆76Updated 3 months ago
- This is a personal reimplementation of Google's Infini-transformer, utilizing a small 2b model. The project includes both model and train…☆58Updated last year
- Fast LLM Training CodeBase With dynamic strategy choosing [Deepspeed+Megatron+FlashAttention+CudaFusionKernel+Compiler];☆41Updated last year
- Implementation of the RWKV language model in pure WebGPU/Rust.☆315Updated 3 weeks ago
- 本项目旨在对大量文本文件进行快速编码检测和转换以辅助mnbvc语料集项目的数据清洗工作☆65Updated 3 weeks ago
- Implementation of the LongRoPE: Extending LLM Context Window Beyond 2 Million Tokens Paper☆149Updated last year
- This repository provides an implementation of "A Simple yet Effective Training-free Prompt-free Approach to Chinese Spelling Correction B…☆78Updated 2 months ago
- 更纯粹、更高压缩率的Tokenizer in Rust☆13Updated 9 months ago
- Code for KaLM-Embedding models☆91Updated 3 months ago
- A relatively basic implementation of RWKV in Rust written by someone with very little math and ML knowledge. Supports 32, 8 and 4 bit eva…☆94Updated 2 years ago
- Efficent platform for inference and serving local LLMs including an OpenAI compatible API server.☆473Updated last week
- Rust crate for some audio utilities☆26Updated 6 months ago
- This repository has code for fine-tuning LLMs with GRPO specifically for Rust Programming using cargo as feedback☆105Updated 6 months ago
- Port of Andrej Karpathy's minbpe to Rust☆29Updated last year
- Inference of Mamba models in pure C☆192Updated last year
- Experiments on speculative sampling with Llama models☆128Updated 2 years ago
- Rust standalone inference of Namo-500M series models. Extremly tiny, runing VLM on CPU.☆24Updated 6 months ago
- Fast serverless LLM inference, in Rust.☆93Updated 7 months ago
- ☆32Updated 2 years ago
- ☆25Updated 2 years ago