REST API Support
Works with any LLM accessible via HTTP REST calls—OpenAI, Anthropic, or self-hosted models.
Mithra is an LLM security scanner that attacks REST-accessible models for risks like prompt injection, data leakage, jailbreaks, toxicity, misinformation, and more.
Probes language models with a curated library of adversarial prompts to surface weaknesses: hallucination, data leakage, prompt injection, misinformation, toxicity, jailbreaks, and more.
Tests for input-driven behavior changes, context bleeding, and access control bypass via crafted prompts.
Evaluates data exfiltration, hallucinations, and harmful content including toxicity and policy bypass.
Research-backed prompts (e.g., DAN jailbreaks, encoding bypasses, exfiltration) mapped to known categories.
Pre-defined attacks cover documented weaknesses including jailbreaks, encoding bypasses, and exfiltration techniques. Scenarios are curated rather than dynamically generated.
Works with any LLM accessible via HTTP REST calls—OpenAI, Anthropic, or self-hosted models.
Curated, research-backed adversarial prompts (e.g., jailbreaks, encoding bypasses) mapped to vulnerability categories.
Target any LLM with an HTTP endpoint. Mithra generates REST requests, applies curated attacks, detects exploit signals, and produces human- and machine-readable reports.
Uses a REST generator to articulate prompts and parameters over HTTP to your target model.
Runs a variety of adversarial prompts to test leakage, jailbreaks, policy bypass, and more.
Detectors analyze responses to assess whether an exploit succeeded or behavior is undesirable.
Get concise summaries and machine-readable outputs for automation and auditing.