AI Security Research

2,077+ academic papers on AI security, attacks, and defenses

Total
2,077
Attack
809
Benchmark
603
Defense
272
Tool
226
Survey
113

Showing 141–160 of 272 papers

Clear filters
Defense MEDIUM

Safety Alignment of LMs via Non-cooperative Games

Anselm Paulus, Ilia Kulikov, Brandon Amos +4 more

Ensuring the safety of language models (LMs) while maintaining their usefulness remains a critical challenge in AI alignment. Current approaches rely...

3 months ago cs.AI PDF
Defense LOW

Distributional AGI Safety

Nenad Tomašev, Matija Franklin, Julian Jacobs +2 more

AI safety and alignment research has predominantly been focused on methods for safeguarding individual AI systems, resting on the assumption of an...

3 months ago cs.AI PDF
Defense MEDIUM

Challenges of Evaluating LLM Safety for User Welfare

Manon Kempermann, Sai Suresh Macharla Vasu, Mahalakshmi Raveenthiran +2 more

Safety evaluations of large language models (LLMs) typically focus on universal risks like dangerous capabilities or undesirable propensities....

3 months ago cs.AI cs.CY PDF
Defense MEDIUM

Phishing Email Detection Using Large Language Models

Najmul Hasan, Prashanth BusiReddyGari, Haitao Zhao +3 more

Email phishing is one of the most prevalent and globally consequential vectors of cyber intrusion. As systems increasingly deploy Large Language...

3 months ago cs.CR cs.IR PDF

Track AI security vulnerabilities in real time

Get breaking CVE alerts, compliance reports (ISO 42001, EU AI Act), and CISO risk assessments for your AI/ML stack.

Start 14-Day Free Trial