phunterlau / code-in-blogLinks
all code examples in the blog posts
☆22Updated 5 months ago
Alternatives and similar repositories for code-in-blog
Users that are interested in code-in-blog are comparing it to the libraries listed below
Sorting:
- A framework for fine-tuning retrieval-augmented generation (RAG) systems.☆122Updated this week
- Official repo for the paper PHUDGE: Phi-3 as Scalable Judge. Evaluate your LLMs with or without custom rubric, reference answer, absolute…☆49Updated last year
- ☆145Updated 11 months ago
- ☆94Updated 3 months ago
- ☆77Updated last year
- Low latency, High Accuracy, Custom Query routers for Humans and Agents. Built by Prithivi Da☆105Updated 3 months ago
- Experimental Code for StructuredRAG: JSON Response Formatting with Large Language Models☆108Updated 3 months ago
- Lean implementation of various multi-agent LLM methods, including Iteration of Thought (IoT)☆115Updated 5 months ago
- Collection of resources for RL and Reasoning☆25Updated 5 months ago
- Fine-tune an LLM to perform batch inference and online serving.☆112Updated last month
- Codebase accompanying the Summary of a Haystack paper.☆79Updated 9 months ago
- Source code of "How to Correctly do Semantic Backpropagation on Language-based Agentic Systems" 🤖☆71Updated 7 months ago
- ☆71Updated 4 months ago
- ☆29Updated last year
- Banishing LLM Hallucinations Requires Rethinking Generalization☆276Updated last year
- LangFair is a Python library for conducting use-case level LLM bias and fairness assessments☆219Updated this week
- LLM reads a paper and produce a working prototype☆58Updated 3 months ago
- SynthGenAI - Package for Generating Synthetic Datasets using LLMs.☆37Updated 5 months ago
- Code for evaluating with Flow-Judge-v0.1 - an open-source, lightweight (3.8B) language model optimized for LLM system evaluations. Crafte…☆73Updated 8 months ago
- A comprehensive guide to LLM evaluation methods designed to assist in identifying the most suitable evaluation techniques for various use…☆123Updated last week
- A framework for standardizing evaluations of large foundation models, beyond single-score reporting and rankings.☆163Updated this week
- Large Language Model (LLM) powered evaluator for Retrieval Augmented Generation (RAG) pipelines.☆29Updated last year
- An overview of GRPO & DeepSeek-R1 Training with Open Source GRPO Model Fine Tuning☆34Updated last month
- The first dense retrieval model that can be prompted like an LM☆81Updated 2 months ago
- Initiative to evaluate and rank the most popular LLMs across common task types based on their propensity to hallucinate.☆111Updated 10 months ago
- Official Implementation of "Affordable AI Assistants with Knowledge Graph of Thoughts"☆131Updated 3 weeks ago
- Code for LitLLMs, LLMs for Literature Review: Are we there yet? (TMLR 2025)☆33Updated 2 months ago
- Benchmark various LLM Structured Output frameworks: Instructor, Mirascope, Langchain, LlamaIndex, Fructose, Marvin, Outlines, etc on task…☆173Updated 9 months ago
- ☆118Updated 10 months ago
- Simple examples using Argilla tools to build AI☆53Updated 7 months ago