kevinwu23 / StanfordFineTuneBenchLinks
☆31Updated 8 months ago
Alternatives and similar repositories for StanfordFineTuneBench
Users that are interested in StanfordFineTuneBench are comparing it to the libraries listed below
Sorting:
- ☆49Updated 5 months ago
- An introduction to LLM Sampling☆79Updated 7 months ago
- Simple GRPO scripts and configurations.☆59Updated 6 months ago
- Storing long contexts in tiny caches with self-study☆121Updated this week
- ☆64Updated last month
- Matrix (Multi-Agent daTa geneRation Infra and eXperimentation framework) is a versatile engine for multi-agent conversational data genera…☆81Updated this week
- Trully flash implementation of DeBERTa disentangled attention mechanism.☆62Updated 2 months ago
- Training an LLM to use a calculator with multi-turn reinforcement learning, achieving a **62% absolute increase in evaluation accuracy**.☆45Updated 3 months ago
- Mixing Language Models with Self-Verification and Meta-Verification☆105Updated 7 months ago
- ☆87Updated last year
- Code for NeurIPS LLM Efficiency Challenge☆59Updated last year
- experiments with inference on llama☆104Updated last year
- minimal pytorch implementation of bm25 (with sparse tensors)☆104Updated last year
- ☆128Updated 3 months ago
- ☆73Updated 2 weeks ago
- Manage scalable open LLM inference endpoints in Slurm clusters☆268Updated last year
- Optimizing Causal LMs through GRPO with weighted reward functions and automated hyperparameter tuning using Optuna☆55Updated 6 months ago
- XTR/WARP (SIGIR'25) is an extremely fast and accurate retrieval engine based on Stanford's ColBERTv2/PLAID and Google DeepMind's XTR.☆152Updated 3 months ago
- LLM training in simple, raw C/CUDA☆15Updated 8 months ago
- Python library to use Pleias-RAG models☆61Updated 3 months ago
- A Python wrapper around HuggingFace's TGI (text-generation-inference) and TEI (text-embedding-inference) servers.☆33Updated 2 months ago
- Official repo for the paper PHUDGE: Phi-3 as Scalable Judge. Evaluate your LLMs with or without custom rubric, reference answer, absolute…☆49Updated last year
- ☆56Updated 2 months ago
- ☆47Updated last year
- An easy-to-understand framework for LLM samplers that rewind and revise generated tokens☆140Updated 5 months ago
- ☆23Updated 2 years ago
- Supervised instruction finetuning for LLM with HF trainer and Deepspeed☆35Updated 2 years ago
- ☆154Updated 8 months ago
- Small and Efficient Mathematical Reasoning LLMs☆71Updated last year
- Source code for the collaborative reasoner research project at Meta FAIR.☆99Updated 3 months ago