Gleghorn-Lab / Mixture-of-Experts-Sentence-SimilarityLinks
☆16Updated 9 months ago
Alternatives and similar repositories for Mixture-of-Experts-Sentence-Similarity
Users that are interested in Mixture-of-Experts-Sentence-Similarity are comparing it to the libraries listed below
Sorting:
- Efficient retrieval head analysis with triton flash attention that supports topK probability☆13Updated last year
- ☆27Updated 2 months ago
- Pre-trained Language Model for Scientific Text☆46Updated last year
- Text Diffusion Model with Encoder-Decoder Transformers for Sequence-to-Sequence Generation [NAACL 2024]☆98Updated 2 years ago
- Codebase for Instruction Following without Instruction Tuning☆36Updated last year
- Scaling Sparse Fine-Tuning to Large Language Models☆18Updated last year
- data collator for UL2 and U-PaLM☆29Updated 2 years ago
- [NeurIPS-2024] 📈 Scaling Laws with Vocabulary: Larger Models Deserve Larger Vocabularies https://arxiv.org/abs/2407.13623☆89Updated last year
- ☆51Updated last year
- [NAACL 2025] A Closer Look into Mixture-of-Experts in Large Language Models☆56Updated 10 months ago
- [COLM 2024] Early Weight Averaging meets High Learning Rates for LLM Pre-training☆18Updated last year
- Few-shot Learning with Auxiliary Data☆31Updated 2 years ago
- Unofficial PyTorch implementation of "Step-unrolled Denoising Autoencoders for Text Generation"☆24Updated 3 years ago
- ResiDual: Transformer with Dual Residual Connections, https://arxiv.org/abs/2304.14802☆96Updated 2 years ago
- ☆15Updated last year
- ☆20Updated 4 years ago
- ☆35Updated last year
- Offcial Repo of Paper "Eliminating Position Bias of Language Models: A Mechanistic Approach""☆19Updated 6 months ago
- Code for the arXiv preprint "The Unreasonable Effectiveness of Easy Training Data"☆48Updated last year
- Download, parse, and filter data PubMed, data-ready for The-Pile☆23Updated 4 years ago
- Code for ICML 25 paper "Metadata Conditioning Accelerates Language Model Pre-training (MeCo)"☆48Updated 5 months ago
- ☆11Updated last year
- An annotated implementation of the Hyena Hierarchy paper☆34Updated 2 years ago
- ☆20Updated 3 years ago
- ☆29Updated last year
- Embedding Recycling for Language models☆38Updated 2 years ago
- User-friendly implementation of the Mixture-of-Sparse-Attention (MoSA). MoSA selects distinct tokens for each head with expert choice rou…☆28Updated 7 months ago
- Transformers at any scale☆42Updated last year
- Interpretable unified language safety checking with large language models☆31Updated 2 years ago
- ☆18Updated last year