AI Security Research

2,077+ academic papers on AI security, attacks, and defenses

Total
2,077
Attack
809
Benchmark
603
Defense
272
Tool
226
Survey
113

Showing 681–700 of 986 papers

Clear filters
Attack MEDIUM

LLM Reinforcement in Context

Thomas Rivasseau

Current Large Language Model alignment research mostly focuses on improving model robustness against adversarial attacks and misbehavior by training...

4 months ago cs.CL cs.CR PDF
Tool MEDIUM

ICX360: In-Context eXplainability 360 Toolkit

Dennis Wei, Ronny Luss, Xiaomeng Hu +6 more

Large Language Models (LLMs) have become ubiquitous in everyday life and are entering higher-stakes applications ranging from summarizing meeting...

4 months ago cs.CL cs.LG PDF

Track AI security vulnerabilities in real time

Get breaking CVE alerts, compliance reports (ISO 42001, EU AI Act), and CISO risk assessments for your AI/ML stack.

Start 14-Day Free Trial