nexusflowai / NexusBench
Nexusflow function call, tool use, and agent benchmarks.
☆19Updated 4 months ago
Alternatives and similar repositories for NexusBench:
Users that are interested in NexusBench are comparing it to the libraries listed below
- Data preparation code for CrystalCoder 7B LLM☆44Updated 11 months ago
- Unleash the full potential of exascale LLMs on consumer-class GPUs, proven by extensive benchmarks, with no long-term adjustments and min…☆26Updated 5 months ago
- Repo hosting codes and materials related to speeding LLMs' inference using token merging.☆35Updated 11 months ago
- Optimizing Causal LMs through GRPO with weighted reward functions and automated hyperparameter tuning using Optuna☆39Updated 2 months ago
- Modified Beam Search with periodical restart☆12Updated 7 months ago
- The Benefits of a Concise Chain of Thought on Problem Solving in Large Language Models☆21Updated 4 months ago
- ☆16Updated last month
- Easy to use, High Performant Knowledge Distillation for LLMs☆58Updated this week
- ☆15Updated this week
- ☆18Updated 3 weeks ago
- ☆48Updated 5 months ago
- Radiantloom Email Assist 7B is an email-assistant large language model fine-tuned from Zephyr-7B-Beta, over a custom-curated dataset of 1…☆14Updated last year
- ☆53Updated 10 months ago
- XmodelLM☆39Updated 4 months ago
- ☆36Updated 2 years ago
- ☆13Updated 3 months ago
- A fast, local, and secure approach for training LLMs for coding tasks using GRPO with WebAssembly and interpreter feedback.☆21Updated last week
- Plug in and Play implementation of "Certified Reasoning with Language Models" that elevates model reasoning by 40%☆17Updated last year
- Public reports detailing responses to sets of prompts by Large Language Models.☆30Updated 3 months ago
- Training hybrid models for dummies.☆20Updated 2 months ago
- AgentParse is a high-performance parsing library designed to map various structured data formats (such as Pydantic models, JSON, YAML, an…☆13Updated last week
- ☆11Updated last week
- Parameter-Efficient Sparsity Crafting From Dense to Mixture-of-Experts for Instruction Tuning on General Tasks☆31Updated 10 months ago
- ☆46Updated 8 months ago
- Official homepage for "Self-Harmonized Chain of Thought" (NAACL 2025)☆91Updated 2 months ago
- The Swarm Ecosystem☆19Updated 8 months ago
- Using modal.com to process FineWeb-edu data☆20Updated last week
- Using multiple LLMs for ensemble Forecasting☆16Updated last year
- entropix style sampling + GUI☆25Updated 5 months ago
- A new way to generate large quantities of high quality synthetic data (on par with GPT-4), with better controllability, at a fraction of …☆22Updated 6 months ago