lucidrains / nim-tokenizer
Implementation of a simple BPE tokenizer, but in Nim
☆21Updated last year
Related projects ⓘ
Alternatives and complementary repositories for nim-tokenizer
- My explorations into editing the knowledge and memories of an attention network☆34Updated last year
- Implementation of N-Grammer, augmenting Transformers with latent n-grams, in Pytorch☆72Updated last year
- A python library for highly configurable transformers - easing model architecture search and experimentation.☆49Updated 2 years ago
- Implementation of the Kalman Filtering Attention proposed in "Kalman Filtering Attention for User Behavior Modeling in CTR Prediction"☆57Updated last year
- Various transformers for FSDP research☆33Updated 2 years ago
- A dashboard for exploring timm learning rate schedulers☆18Updated last year
- Local Attention - Flax module for Jax☆20Updated 3 years ago
- Implementation of Token Shift GPT - An autoregressive model that solely relies on shifting the sequence space for mixing☆47Updated 2 years ago
- **ARCHIVED** Filesystem interface to 🤗 Hub☆56Updated last year
- Repository for fine-tuning Transformers 🤗 based seq2seq speech models in JAX/Flax.☆34Updated last year
- Training and evaluation code for the paper "Headless Language Models: Learning without Predicting with Contrastive Weight Tying" (https:/…☆23Updated 7 months ago
- Implementation of a Transformer using ReLA (Rectified Linear Attention) from https://arxiv.org/abs/2104.07012☆49Updated 2 years ago
- Fast, Modern, Memory Efficient, and Low Precision PyTorch Optimizers☆59Updated 4 months ago
- Another attempt at a long-context / efficient transformer by me☆37Updated 2 years ago
- LTG-Bert☆29Updated 10 months ago
- An extension package of 🤗 Datasets that provides support for executing arbitrary SQL queries on HF datasets☆31Updated 9 months ago
- TorchServe+Streamlit for easily serving your HuggingFace NER models☆31Updated 2 years ago
- ☆29Updated 2 years ago
- Repo for training MLMs, CLMs, or T5-type models on the OLM pretraining data, but it should work with any hugging face text dataset.☆92Updated last year
- Tutorial to pretrain & fine-tune a 🤗 Flax T5 model on a TPUv3-8 with GCP☆58Updated 2 years ago
- Exploring an idea where one forgets about efficiency and carries out attention across each edge of the nodes (tokens)☆43Updated last month
- CUDA implementation of autoregressive linear attention, with all the latest research findings☆43Updated last year
- Speech in Flax/JAX☆15Updated 2 years ago
- QAmeleon introduces synthetic multilingual QA data using PaLM, a 540B large language model. This dataset was generated by prompt tuning P…☆34Updated last year
- Load any clip model with a standardized interface☆21Updated 6 months ago
- Contrastive Language-Audio Pretraining☆15Updated 3 years ago
- This is a new metric that can be used to evaluate faithfulness of text generated by LLMs. The work behind this repository can be found he…☆31Updated last year
- High performance pytorch modules☆18Updated last year