survey 2025

Guardians and Offenders: A Survey on Harmful Content Generation and Safety Mitigation of LLM

Chi Zhang 1, Changjia Zhu 1, Junjie Xiong 2, Xiaoran Xu 1, Lingyao Li 1, Yao Liu 1, Zhuo Lu 1

0 citations

α

Published on arXiv

2508.05775

Prompt Injection

OWASP LLM Top 10 — LLM01

Key Finding

Identifies a dual trajectory where LLMs are simultaneously sources of harmful content and promising tools for safety, while highlighting critical limitations in current evaluation methodologies for LLM safety.


Large Language Models (LLMs) have revolutionized content creation across digital platforms, offering unprecedented capabilities in natural language generation and understanding. These models enable beneficial applications such as content generation, question and answering (Q&A), programming, and code reasoning. Meanwhile, they also pose serious risks by inadvertently or intentionally producing toxic, offensive, or biased content. This dual role of LLMs, both as powerful tools for solving real-world problems and as potential sources of harmful language, presents a pressing sociotechnical challenge. In this survey, we systematically review recent studies spanning unintentional toxicity, adversarial jailbreaking attacks, and content moderation techniques. We propose a unified taxonomy of LLM-related harms and defenses, analyze emerging multimodal and LLM-assisted jailbreak strategies, and assess mitigation efforts, including reinforcement learning with human feedback (RLHF), prompt engineering, and safety alignment. Our synthesis highlights the evolving landscape of LLM safety, identifies limitations in current evaluation methodologies, and outlines future research directions to guide the development of robust and ethically aligned language technologies.


Key Contributions

  • Unified taxonomy of LLM-related harms and defenses spanning unintentional toxicity, adversarial jailbreaks, and content moderation
  • Analysis of emerging multimodal and LLM-assisted jailbreak strategies
  • Assessment of mitigation efforts including RLHF, prompt engineering, and safety alignment, with identified gaps in current evaluation methodologies

🛡️ Threat Analysis


Details

Domains
nlpmultimodal
Model Types
llmvlm
Threat Tags
inference_timetargeteddigital
Applications
content moderationchatbottext generation