takara-ai / SwarmFormerLinks
A pytorch implementation of SwarmFormer for text classification.
☆16Updated 5 months ago
Alternatives and similar repositories for SwarmFormer
Users that are interested in SwarmFormer are comparing it to the libraries listed below
Sorting:
- Set of scripts to finetune LLMs☆38Updated last year
- Official repo for the paper PHUDGE: Phi-3 as Scalable Judge. Evaluate your LLMs with or without custom rubric, reference answer, absolute…☆51Updated last year
- Code for NeurIPS LLM Efficiency Challenge☆59Updated last year
- Chunk your text using gpt4o-mini more accurately☆44Updated last year
- ☆51Updated 9 months ago
- ☆138Updated 3 months ago
- ☆124Updated last year
- Trully flash implementation of DeBERTa disentangled attention mechanism.☆67Updated 2 months ago
- Source code for the collaborative reasoner research project at Meta FAIR.☆106Updated 7 months ago
- ☆82Updated last year
- Code for ExploreTom☆87Updated 5 months ago
- Simple GRPO scripts and configurations.☆59Updated 9 months ago
- Source code of "How to Correctly do Semantic Backpropagation on Language-based Agentic Systems" 🤖☆76Updated 11 months ago
- Pre-train Static Word Embeddings☆91Updated 2 months ago
- Generalist and Lightweight Model for Text Classification☆165Updated 5 months ago
- Using open source LLMs to build synthetic datasets for direct preference optimization☆69Updated last year
- Luth is a state-of-the-art series of fine-tuned LLMs for French☆39Updated last month
- ☆86Updated 4 months ago
- ☆52Updated last year
- Anchored Preference Optimization and Contrastive Revisions: Addressing Underspecification in Alignment☆60Updated last year
- Collection of resources for RL and Reasoning☆26Updated 9 months ago
- Train LLM on Hugging Face infra☆67Updated 2 weeks ago
- An introduction to LLM Sampling☆79Updated 11 months ago
- ☆146Updated last year
- ☆80Updated last year
- ☆55Updated last year
- Experimental Code for StructuredRAG: JSON Response Formatting with Large Language Models☆113Updated 7 months ago
- Python library to use Pleias-RAG models☆67Updated 6 months ago
- State-of-the-art paired encoder and decoder models (17M-1B params)☆53Updated 3 months ago
- ☆59Updated last year