Di-Is / faiss-gpu-wheelsLinks
Unofficial faiss wheel builder for NVIDIA GPU
☆33Updated last month
Alternatives and similar repositories for faiss-gpu-wheels
Users that are interested in faiss-gpu-wheels are comparing it to the libraries listed below
Sorting:
- Evaluate your agent memory on real-world dialogues, not LLM-simulated dialogues.☆36Updated 7 months ago
- ☆58Updated 11 months ago
- Code for paper "Patch-Level Training for Large Language Models"☆97Updated 3 months ago
- ☆161Updated last year
- [ICLR 2025 Oral] "Your Mixture-of-Experts LLM Is Secretly an Embedding Model For Free"☆89Updated last year
- Code for Zero-Shot Tokenizer Transfer☆142Updated last year
- ☆51Updated last year
- ☆74Updated last year
- This is the repo for the paper "PANGEA: A FULLY OPEN MULTILINGUAL MULTIMODAL LLM FOR 39 LANGUAGES"☆118Updated 7 months ago
- ☆25Updated 10 months ago
- ☆84Updated 3 months ago
- Official implementation of "GPT or BERT: why not both?"☆61Updated 6 months ago
- Implementation of 🥥 Coconut, Chain of Continuous Thought, in Pytorch☆182Updated 7 months ago
- ☆75Updated last year
- [EMNLP 2025] Distill Visual Chart Reasoning Ability from LLMs to MLLMs☆59Updated 5 months ago
- ☆220Updated 3 months ago
- Organize the Web: Constructing Domains Enhances Pre-Training Data Curation☆77Updated 9 months ago
- ☆106Updated 8 months ago
- [TMLR 2026] When Attention Collapses: How Degenerate Layers in LLMs Enable Smaller, Stronger Models☆122Updated last year
- ☆53Updated last year
- Code for "SemDeDup", a simple method for identifying and removing semantic duplicates from a dataset (data pairs which are semantically s…☆151Updated 2 years ago
- Official code for paper "UniIR: Training and Benchmarking Universal Multimodal Information Retrievers" (ECCV 2024)☆177Updated last year
- Vocabulary Trimming (VT) is a model compression technique, which reduces a multilingual LM vocabulary to a target language by deleting ir…☆61Updated last year
- E5-V: Universal Embeddings with Multimodal Large Language Models☆273Updated 2 months ago
- A plug-and-play tool for visualizing attention-score heatmap in generative LLMs. Easy to customize for your own need.☆51Updated last year
- [NeurIPS 2024 Main Track] Code for the paper titled "Instruction Tuning With Loss Over Instructions"☆38Updated last year
- Implementation of the paper: "Leave No Context Behind: Efficient Infinite Context Transformers with Infini-attention" from Google in pyTO…☆58Updated last week
- This is the implementation of the paper AdaMix: Mixture-of-Adaptations for Parameter-efficient Model Tuning (https://arxiv.org/abs/2205.1…☆136Updated 2 years ago
- Easy modernBERT fine-tuning and multi-task learning☆63Updated 7 months ago
- [NeurIPS 2025] MergeBench: A Benchmark for Merging Domain-Specialized LLMs☆41Updated 2 weeks ago