Watchfulio / dataset-generator
A new way to generate large quantities of high quality synthetic data (on par with GPT-4), with better controllability, at a fraction of the cost of prompting LLMs directly.
☆22Updated 6 months ago
Alternatives and similar repositories for dataset-generator:
Users that are interested in dataset-generator are comparing it to the libraries listed below
- ☆48Updated 5 months ago
- Script for processing OpenAI's PRM800K process supervision dataset into an Alpaca-style instruction-response format☆27Updated last year
- Implementation of the paper: "AssistantBench: Can Web Agents Solve Realistic and Time-Consuming Tasks?"☆53Updated 4 months ago
- A fast, local, and secure approach for training LLMs for coding tasks using GRPO with WebAssembly and interpreter feedback.☆22Updated last week
- Anchored Preference Optimization and Contrastive Revisions: Addressing Underspecification in Alignment☆55Updated 7 months ago
- Optimizing Causal LMs through GRPO with weighted reward functions and automated hyperparameter tuning using Optuna☆39Updated 2 months ago
- Simple GRPO scripts and configurations.☆58Updated 2 months ago
- Training hybrid models for dummies.☆20Updated 3 months ago
- Nexusflow function call, tool use, and agent benchmarks.☆19Updated 4 months ago
- ☆15Updated last week
- Code for the paper: CodeTree: Agent-guided Tree Search for Code Generation with Large Language Models☆17Updated 2 weeks ago
- Plug in and Play implementation of "Certified Reasoning with Language Models" that elevates model reasoning by 40%☆17Updated last year
- Lightweight tools for quick and easy LLM demo's☆26Updated 6 months ago
- Repository for "I am a Strange Dataset: Metalinguistic Tests for Language Models"☆42Updated last year
- A repository for research on medium sized language models.☆76Updated 10 months ago
- Based on the tree of thoughts paper☆48Updated last year
- Zeus LLM Trainer is a rewrite of Stanford Alpaca aiming to be the trainer for all Large Language Models☆69Updated last year
- Code for RATIONALYST: Pre-training Process-Supervision for Improving Reasoning https://arxiv.org/pdf/2410.01044☆32Updated 6 months ago
- Repository containing the SPIN experiments on the DIBT 10k ranked prompts☆24Updated last year
- Mixture of Expert (MoE) techniques for enhancing LLM performance through expert-driven prompt mapping and adapter combinations.☆12Updated last year
- ☆16Updated 6 months ago
- Small, simple agent task environments for training and evaluation☆18Updated 5 months ago
- Official repo for NAACL 2024 Findings paper "LeTI: Learning to Generate from Textual Interactions."☆63Updated last year
- Functional Benchmarks and the Reasoning Gap☆85Updated 6 months ago
- ☆21Updated 6 months ago
- ☆27Updated 3 weeks ago
- Measuring and Controlling Persona Drift in Language Model Dialogs☆17Updated last year
- Minimum Description Length probing for neural network representations☆19Updated 2 months ago
- A framework for pitting LLMs against each other in an evolving library of games ⚔☆33Updated 2 weeks ago
- ☆33Updated 9 months ago