chandar-lab / NeoBERTLinks
☆81Updated 2 months ago
Alternatives and similar repositories for NeoBERT
Users that are interested in NeoBERT are comparing it to the libraries listed below
Sorting:
- Trully flash implementation of DeBERTa disentangled attention mechanism.☆63Updated 2 weeks ago
- ☆68Updated last month
- ☆51Updated 7 months ago
- minimal pytorch implementation of bm25 (with sparse tensors)☆104Updated last year
- A toolkit implementing advanced methods to transfer models and model knowledge across tokenizers.☆44Updated last month
- Official implementation of "GPT or BERT: why not both?"☆57Updated last month
- ☆49Updated 6 months ago
- Crispy reranking models by Mixedbread☆34Updated last month
- Fine-tune ModernBERT on a large Dataset with Custom Tokenizer Training☆67Updated 6 months ago
- ☆58Updated 3 months ago
- Pre-train Static Word Embeddings☆84Updated 2 months ago
- Truly flash T5 realization!☆70Updated last year
- Training and evaluation code for the paper "Headless Language Models: Learning without Predicting with Contrastive Weight Tying" (https:/…☆27Updated last year
- ☆63Updated 3 weeks ago
- A fast implementation of T5/UL2 in PyTorch using Flash Attention☆107Updated 5 months ago
- Fast, Modern, and Low Precision PyTorch Optimizers☆109Updated last week