StealthGraph: Exposing Domain-Specific Risks in LLMs through Knowledge-Graph-Guided Harmful Prompt Generation
About
Large language models (LLMs) are increasingly applied in specialized domains such as finance and healthcare, where they introduce unique safety risks. Domain-specific datasets of harmful prompts remain scarce and still largely rely on manual construction; public datasets mainly focus on explicit harmful prompts, which modern LLM defenses can often detect and refuse. In contrast, implicit harmful prompts-expressed through indirect domain knowledge-are harder to detect and better reflect real-world threats. We identify two challenges: transforming domain knowledge into actionable constraints and increasing the implicitness of generated harmful prompts. To address them, we propose an end-to-end framework that first performs knowledge-graph-guided harmful prompt generation to systematically produce domain-relevant prompts, and then applies dual-path obfuscation rewriting to convert explicit harmful prompts into implicit variants via direct and context-enhanced rewriting. This framework yields high-quality datasets combining strong domain relevance with implicitness, enabling more realistic red-teaming and advancing LLM safety research. We release our code and datasets at GitHub.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Jailbreak Attack Evaluation | StealthGraph SG-Implicit | -- | 12 | |
| Jailbreak Attack Evaluation | AdvBench | -- | 8 | |
| Jailbreak Attack Evaluation | Do-Not-Answer | -- | 6 | |
| Jailbreak Attack Evaluation | HARMFULQA | -- | 6 | |
| Jailbreak Attack Evaluation | StealthGraph SG-Origin | -- | 6 | |
| Language Modeling | SG-Implicit | PPL79.87 | 2 | |
| Language Modeling | SG Origin | Perplexity29.37 | 1 | |
| Language Modeling | AdvBench | -- | 1 | |
| Language Modeling | Do-Not-Answer | -- | 1 | |
| Language Modeling | HARMFULQA | -- | 1 |