sam-paech / slop-forensicsLinks
☆304Updated 2 months ago
Alternatives and similar repositories for slop-forensics
Users that are interested in slop-forensics are comparing it to the libraries listed below
Sorting:
- Conduct in-depth research with AI-driven insights : DeepDive is a command-line tool that leverages web searches and AI models to generate…☆44Updated last year
- explore token trajectory trees on instruct and base models☆150Updated 7 months ago
- AI management tool☆119Updated last year
- ☆334Updated 5 months ago
- Hallucinations (Confabulations) Document-Based Benchmark for RAG. Includes human-verified questions and answers.☆241Updated 5 months ago
- A simple tool that let's you explore different possible paths that an LLM might sample.☆199Updated 8 months ago
- ☆107Updated 2 months ago
- ☆134Updated 8 months ago
- An easy-to-understand framework for LLM samplers that rewind and revise generated tokens☆150Updated last week
- Benchmark that evaluates LLMs using 759 NYT Connections puzzles extended with extra trick words☆190Updated last month
- klmbr - a prompt pre-processing technique to break through the barrier of entropy while generating text with LLMs☆86Updated last year
- Transplants vocabulary between language models, enabling the creation of draft models for speculative decoding WITHOUT retraining.☆47Updated 2 months ago
- ☆158Updated 9 months ago
- ☆210Updated last week
- Train an adapter for any embedding model in under a minute☆130Updated 9 months ago
- Easily view and modify JSON datasets for large language models☆86Updated 8 months ago
- A very fast, very minimal prompt optimizer☆299Updated last year
- Code action agent with local execution sandbox and first-class support for programmatic tool calling☆117Updated this week
- Enhancing LLMs with LoRA☆205Updated 2 months ago
- Command-line personal assistant using your favorite proprietary or local models with access to over 30+ tools☆111Updated 6 months ago
- Verify Precision of all Kimi K2 API Vendor☆494Updated last week
- Thematic Generalization Benchmark: measures how effectively various LLMs can infer a narrow or specific "theme" (category/rule) from a sm…☆63Updated 3 months ago
- Multi-Agent Step Race Benchmark: Assessing LLM Collaboration and Deception Under Pressure. A multi-player “step-race” that challenges LLM…☆81Updated last month
- Fast parallel LLM inference for MLX☆241Updated last year
- A simple MLX implementation for pretraining LLMs on Apple Silicon.☆85Updated 4 months ago
- Public Goods Game (PGG) Benchmark: Contribute & Punish is a multi-agent benchmark that tests cooperative and self-interested strategies a…☆39Updated 9 months ago
- Serving LLMs in the HF-Transformers format via a PyFlask API☆72Updated last year
- This benchmark tests how well LLMs incorporate a set of 10 mandatory story elements (characters, objects, core concepts, attributes, moti…☆329Updated last month
- A prompting library☆189Updated 6 months ago
- A user interface for DSPy☆208Updated 3 months ago