AI Security Research

2,077+ academic papers on AI security, attacks, and defenses

Total
2,077
Attack
809
Benchmark
603
Defense
272
Tool
226
Survey
113

Showing 741–760 of 971 papers

Clear filters
Attack MEDIUM

ShadowLogic: Backdoors in Any Whitebox LLM

Kasimir Schulz, Amelia Kawasaki, Leo Ring

Large language models (LLMs) are widely deployed across various applications, often with safeguards to prevent the generation of harmful or...

4 months ago cs.CR cs.AI PDF
Defense MEDIUM

Reimagining Safety Alignment with An Image

Yifan Xia, Guorui Chen, Wenqian Yu +3 more

Large language models (LLMs) excel in diverse applications but face dual challenges: generating harmful content under jailbreak attacks and...

4 months ago cs.AI cs.CR PDF
Attack MEDIUM

Diffusion LLMs are Natural Adversaries for any LLM

David Lüdke, Tom Wollschläger, Paul Ungermann +2 more

We introduce a novel framework that transforms the resource-intensive (adversarial) prompt optimization problem into an \emph{efficient, amortized...

4 months ago cs.LG stat.ML PDF
Benchmark MEDIUM

Self-HarmLLM: Can Large Language Model Harm Itself?

Heehwan Kim, Sungjune Park, Daeseon Choi

Large Language Models (LLMs) are generally equipped with guardrails to block the generation of harmful responses. However, existing defenses always...

4 months ago cs.CL cs.AI PDF

Track AI security vulnerabilities in real time

Get breaking CVE alerts, compliance reports (ISO 42001, EU AI Act), and CISO risk assessments for your AI/ML stack.

Start 14-Day Free Trial