chandar-lab / NeoBERTLinks
☆83Updated 3 months ago
Alternatives and similar repositories for NeoBERT
Users that are interested in NeoBERT are comparing it to the libraries listed below
Sorting:
- Trully flash implementation of DeBERTa disentangled attention mechanism.☆63Updated 2 weeks ago
- ☆69Updated 2 months ago
- minimal pytorch implementation of bm25 (with sparse tensors)☆104Updated last year
- Official implementation of "GPT or BERT: why not both?"☆58Updated last month
- A toolkit implementing advanced methods to transfer models and model knowledge across tokenizers.☆46Updated 2 months ago
- ☆49Updated 7 months ago
- Pre-train Static Word Embeddings☆85Updated last week
- Training and evaluation code for the paper "Headless Language Models: Learning without Predicting with Contrastive Weight Tying" (https:/…☆27Updated last year
- A massively multilingual modern encoder language model☆80Updated last week
- ☆51Updated 7 months ago
- Efficient few-shot learning with cross-encoders.☆58Updated last year
- Code for SaGe subword tokenizer (EACL 2023)☆26Updated 9 months ago
- Fine-tune ModernBERT on a large Dataset with Custom Tokenizer Training☆67Updated 7 months ago
- Using open source LLMs to build synthetic datasets for direct preference optimization☆65Updated last year
- A fast implementation of T5/UL2 in PyTorch using Flash Attention☆107Updated 6 months ago
- ☆39Updated last year
- Using short models to classify long texts☆21Updated 2 years ago
- ☆48Updated last year
- Fast, Modern, and Low Precision PyTorch Optimizers☆109Updated 2 weeks ago
- State-of-the-art paired encoder and decoder models (17M-1B params)☆45Updated last month
- QAmeleon introduces synthetic multilingual QA data using PaLM, a 540B large language model. This dataset was generated by prompt tuning P…☆34Updated 2 years ago
- Datamodels for hugging face tokenizers☆47Updated last week
- Crispy reranking models by Mixedbread☆35Updated this week
- Supercharge huggingface transformers with model parallelism.☆77Updated last month
- Improving Text Embedding of Language Models Using Contrastive Fine-tuning☆64Updated last year
- ☆58Updated 4 months ago
- Starbucks: Improved Training for 2D Matryoshka Embeddings☆21Updated 2 months ago
- ☆66Updated last month
- Datasets collection and preprocessings framework for NLP extreme multitask learning☆186Updated 2 months ago
- My NER Experiments with ModernBERT and Ettin☆22Updated 2 months ago