Oxen-AI / GRPO-With-Cargo-FeedbackLinks
This repository has code for fine-tuning LLMs with GRPO specifically for Rust Programming using cargo as feedback
☆111Updated 8 months ago
Alternatives and similar repositories for GRPO-With-Cargo-Feedback
Users that are interested in GRPO-With-Cargo-Feedback are comparing it to the libraries listed below
Sorting:
- High-performance MinHash implementation in Rust with Python bindings for efficient similarity estimation and deduplication of large datas…☆214Updated last month
- Fast serverless LLM inference, in Rust.☆105Updated last week
- Train your own SOTA deductive reasoning model☆108Updated 8 months ago
- ☆135Updated last year
- ☆68Updated 5 months ago
- A high-performance constrained decoding engine based on context free grammar in Rust☆55Updated 5 months ago
- implement llava using candle☆15Updated last year
- Inference Llama 2 in one file of zero-dependency, zero-unsafe Rust☆39Updated 2 years ago
- Faster structured generation☆257Updated last week
- Inference engine for GLiNER models, in Rust☆76Updated last week
- Optimizing Causal LMs through GRPO with weighted reward functions and automated hyperparameter tuning using Optuna☆58Updated 3 weeks ago
- ☆139Updated last year
- TensorRT-LLM server with Structured Outputs (JSON) built with Rust☆60Updated 6 months ago
- Performance centered DSPy rewrite to(not port) Rust☆174Updated last week
- j1-micro (1.7B) & j1-nano (600M) are absurdly tiny but mighty reward models.☆97Updated 3 months ago
- Rust implementation of Surya☆63Updated 8 months ago
- Unofficial Rust bindings to Apple's mlx framework☆206Updated last month
- A tree-based prefix cache library that allows rapid creation of looms: hierarchal branching pathways of LLM generations.☆72Updated 9 months ago
- The Easiest Rust Interface for Local LLMs and an Interface for Deterministic Signals from Probabilistic LLM Vibes☆237Updated 3 months ago
- A single-binary, GPU-accelerated LLM server (HTTP and WebSocket API) written in Rust☆79Updated last year
- Storing long contexts in tiny caches with self-study☆213Updated 3 weeks ago
- Pivotal Token Search☆131Updated 4 months ago
- Low rank adaptation (LoRA) for Candle.☆166Updated 6 months ago
- Super basic implementation (gist-like) of RLMs with REPL environments.☆248Updated 3 weeks ago
- Training an LLM to use a calculator with multi-turn reinforcement learning, achieving a **62% absolute increase in evaluation accuracy**.☆58Updated 6 months ago
- an open source reproduction of NVIDIA's nGPT (Normalized Transformer with Representation Learning on the Hypersphere)☆107Updated 8 months ago
- ☆13Updated 9 months ago
- OpenCoconut implements a latent reasoning paradigm where we generate thoughts before decoding.☆173Updated 9 months ago
- NanoGPT (124M) quality in 2.67B tokens☆28Updated last month
- Formatron empowers everyone to control the format of language models' output with minimal overhead.☆228Updated 5 months ago