wbrown / gpt_bpe
GPT2 Byte Pair Encoding implementation in Golang
☆24Updated last week
Related projects ⓘ
Alternatives and complementary repositories for gpt_bpe
- Demonstration that finetuning RoPE model on larger sequences than the pre-trained model adapts the model context limit☆63Updated last year
- Binding to transformers in ggml☆59Updated last week
- GGML implementation of BERT model with Python bindings and quantization.☆51Updated 8 months ago
- ☆34Updated last year
- ☆15Updated 7 months ago
- ☆26Updated last year
- ☆19Updated this week
- ☆40Updated last year
- ☆49Updated 8 months ago
- RWKV (Receptance Weighted Key Value) is a RNN with Transformer-level performance☆36Updated last year
- tinygrad port of the RWKV large language model.☆43Updated 4 months ago
- Full finetuning of large language models without large memory requirements☆93Updated 10 months ago
- ☆36Updated 2 years ago
- GPU accelerated client-side embeddings for vector search, RAG etc.☆63Updated 11 months ago
- Modified Stanford-Alpaca Trainer for Training Replit's Code Model☆40Updated last year
- The GeoV model is a large langauge model designed by Georges Harik and uses Rotary Positional Embeddings with Relative distances (RoPER).…☆122Updated last year
- Let us make Psychohistory (as in Asimov) a reality, and accessible to everyone. Useful for LLM grounding and games / fiction / business /…☆40Updated last year
- ☆55Updated 11 months ago
- Zeta implementation of a reusable and plug in and play feedforward from the paper "Exponentially Faster Language Modeling"☆15Updated this week
- Latent Diffusion Language Models☆67Updated last year
- A library for squeakily cleaning and filtering language datasets.☆45Updated last year
- ☆48Updated last year
- Fast inference of Instruct tuned LLaMa on your personal devices.☆22Updated last year
- RWKV-7: Surpassing GPT☆43Updated this week
- Tokun to can tokens☆15Updated last month
- Web browser version of StarCoder.cpp☆43Updated last year
- Trying to deconstruct RWKV in understandable terms☆14Updated last year
- SparseGPT + GPTQ Compression of LLMs like LLaMa, OPT, Pythia☆41Updated last year
- Here we collect trick questions and failed tasks for open source LLMs to improve them.☆32Updated last year