AI safety evaluation framework testing LLM epistemic robustness under adversarial self-history manipulation
-
Updated
Dec 18, 2025 - Python
AI safety evaluation framework testing LLM epistemic robustness under adversarial self-history manipulation
Benchmark LLM jailbreak resilience across providers with standardized tests, adversarial mode, rich analytics, and a clean Web UI.
A multi-agent safety engineering framework that subjects systems to adversarial audit. Orchestrates specialized agents (Engineer, Psychologist, Physicist) to find process risks and human factors.
LLM-powered fuzzing and adversarial testing framework for Solana programs. Generates intelligent attack scenarios, builds real transactions, and reports vulnerabilities with CWE classifications.
A dependency-aware Bayesian belief gate that resists correlated evidence and yields only under true independent verification.
Red team toolkit for stress-testing MCP security scanners — find detection gaps before attackers do
Generate adversarial pytest tests using LLM. Tries to find edge cases in your Python code.
Add a description, image, and links to the adversarial-testing topic page so that developers can more easily learn about it.
To associate your repository with the adversarial-testing topic, visit your repo's landing page and select "manage topics."