ibm-granite / granite-guardian
The Granite Guardian models are designed to detect risks in prompts and responses.
☆79Updated last month
Alternatives and similar repositories for granite-guardian:
Users that are interested in granite-guardian are comparing it to the libraries listed below
- Code repo for "Agent Instructs Large Language Models to be General Zero-Shot Reasoners"☆107Updated 7 months ago
- ☆254Updated 5 months ago
- Code for our paper PAPILLON: PrivAcy Preservation from Internet-based and Local Language MOdel ENsembles☆30Updated this week
- A better way of testing, inspecting, and analyzing AI Agent traces.☆35Updated this week
- This is the official code for the paper "Virus: Harmful Fine-tuning Attack for Large Language Models Bypassing Guardrail Moderation"☆46Updated 3 months ago
- Scaling Data for SWE-agents☆101Updated this week
- This is the reproduction repository for my 🤗 Hugging Face blog post on synthetic data☆68Updated last year
- This repository contains popular code generation frameworks such as MapCoder, CodeSIM.☆44Updated 2 weeks ago
- ☆93Updated 7 months ago
- The first dense retrieval model that can be prompted like an LM☆71Updated 7 months ago
- Beating the GAIA benchmark with Transformers Agents. 🚀☆113Updated 2 months ago
- Simple examples using Argilla tools to build AI☆52Updated 5 months ago
- Accompanying code and SEP dataset for the "Can LLMs Separate Instructions From Data? And What Do We Even Mean By That?" paper.☆51Updated last month
- Pre-training code for CrystalCoder 7B LLM☆54Updated 11 months ago
- ☆74Updated 3 months ago
- MiniCheck: Efficient Fact-Checking of LLMs on Grounding Documents [EMNLP 2024]☆155Updated 4 months ago
- Official homepage for "Self-Harmonized Chain of Thought" (NAACL 2025)☆90Updated 3 months ago
- ☆55Updated this week
- Functional Benchmarks and the Reasoning Gap☆85Updated 7 months ago
- ☆73Updated this week
- Official repo for the paper PHUDGE: Phi-3 as Scalable Judge. Evaluate your LLMs with or without custom rubric, reference answer, absolute…☆49Updated 9 months ago
- Systematic evaluation framework that automatically rates overthinking behavior in large language models.☆88Updated last month
- Lean implementation of various multi-agent LLM methods, including Iteration of Thought (IoT)☆109Updated 2 months ago
- ☆86Updated this week
- 🦄 Unitxt is a Python library for enterprise-grade evaluation of AI performance, offering the world's largest catalog of tools and data …☆191Updated this week
- EvolKit is an innovative framework designed to automatically enhance the complexity of instructions used for fine-tuning Large Language M…☆215Updated 6 months ago
- Mixing Language Models with Self-Verification and Meta-Verification☆104Updated 4 months ago
- [ACL 2024] Do Large Language Models Latently Perform Multi-Hop Reasoning?☆65Updated last month
- Language Model for Mainframe Modernization☆52Updated 8 months ago
- ☆114Updated 2 months ago