Benchmark MEDIUM
Abdullah Caglar Oksuz, Anisa Halimi, Erman Ayday
Membership inference attacks (MIAs) threaten the privacy of machine learning models by revealing whether a specific data point was used during...
1 months ago cs.LG cs.CR
PDF
Benchmark LOW
Martin Bertran, Riccardo Fogliato, Zhiwei Steven Wu
Empirical conclusions depend not only on data but on analytic decisions made throughout the research process. Many-analyst studies have quantified...
1 months ago cs.AI cs.LG
PDF
Benchmark HIGH
Mirae Kim, Seonghun Jeong, Youngjun Kwak
Jailbreaking poses a significant risk to the deployment of Large Language Models (LLMs) and Vision Language Models (VLMs). VLMs are particularly...
1 months ago cs.CL cs.AI cs.DB
PDF
Benchmark LOW
Anna Babarczy, Andras Lukacs, Peter Vedres +1 more
The study explores whether current Large Language Models (LLMs) exhibit Theory of Mind (ToM) capabilities -- specifically, the ability to infer...
1 months ago cs.CL cs.AI
PDF
Benchmark MEDIUM
Zachary Coalson, Bo Fang, Sanghyun Hong
Multi-turn interaction length is a dominant factor in the operational costs of conversational LLMs. In this work, we present a new failure mode in...
1 months ago cs.LG cs.CR
PDF
Benchmark MEDIUM
Gelei Deng, Yi Liu, Yuekang Li +5 more
LLM-based agents show promise for automating penetration testing, yet reported performance varies widely across systems and benchmarks. We analyze 28...
1 months ago cs.CR cs.SE
PDF
Benchmark LOW
Takyoung Kim, Jinseok Nam, Chandrayee Basu +5 more
Conversational agents powered by large language models (LLMs) with tool integration achieve strong performance on fixed task-oriented dialogue...
1 months ago cs.CL cs.AI
PDF
Benchmark HIGH
Priyaranjan Pattnayak, Sanchari Chowdhuri
Safety alignment of large language models (LLMs) is mostly evaluated in English and contract-bound, leaving multilingual vulnerabilities...
1 months ago cs.AI cs.CL
PDF
Benchmark MEDIUM
Simon Lermen, Daniel Paleka, Joshua Swanson +3 more
We show that large language models can be used to perform at-scale deanonymization. With full Internet access, our agent can re-identify Hacker News...
1 months ago cs.CR cs.AI cs.LG
PDF
Benchmark LOW
Stephan Rabanser, Sayash Kapoor, Peter Kirgis +3 more
AI agents are increasingly deployed to execute important tasks. While rising accuracy scores on standard benchmarks suggest rapid progress, many...
1 months ago cs.AI cs.CY cs.LG
PDF
Benchmark MEDIUM
Michael Cunningham
We present a practical system for privacy-aware large language model (LLM) inference that splits a transformer between a trusted local GPU and an...
1 months ago cs.CR cs.DC
PDF
Benchmark MEDIUM
Nivya Talokar, Ayush K Tarun, Murari Mandal +2 more
LLM-based agents execute real-world workflows via tools and memory. These affordances enable ill-intended adversaries to also use these agents to...
1 months ago cs.CL cs.LG
PDF
Benchmark MEDIUM
Johannes Bertram, Jonas Geiping
We introduce NESSiE, the NEceSsary SafEty benchmark for large language models (LLMs). With minimal test cases of information and access security,...
1 months ago cs.CR cs.SE
PDF
Benchmark MEDIUM
Shahriar Golchin, Marc Wetter
We systematically evaluate the quality of widely used AI safety datasets from two perspectives: in isolation and in practice. In isolation, we...
1 months ago cs.CR cs.AI cs.CL
PDF
Benchmark MEDIUM
Haodong Zhao, Jinming Hu, Gongshen Liu
Federated learning security research has predominantly focused on backdoor threats from a minority of malicious clients that intentionally corrupt...
Benchmark LOW
Aditi Prabakaran, Priyesh Shukla
Transient objects in casual multi-view captures cause ghosting artifacts in 3D Gaussian Splatting (3DGS) reconstruction. Existing solutions relied on...
Benchmark LOW
Udbhav Prasad, Aniesh Chawla
Cryptographic digests (e.g., MD5, SHA-256) are designed to provide exact identity. Any single-bit change in the input produces a completely different...
1 months ago cs.CR cs.AI
PDF
Benchmark MEDIUM
Max Fomin
Detecting prompt injection and jailbreak attacks is critical for deploying LLM-based agents safely. As agents increasingly process untrusted data...
Benchmark LOW
Edibe Yilmaz, Kahraman Kostas
The integration of large language models (LLMs) into educational processes introduces significant constraints regarding data privacy and reliability,...
1 months ago cs.CL cs.AI cs.CR
PDF
Benchmark HIGH
Haoyu Li, Xijia Che, Yanhao Wang +2 more
Proof-of-Vulnerability (PoV) generation is a critical task in software security, serving as a cornerstone for vulnerability validation, false...
1 months ago cs.SE cs.CR
PDF
Track AI security vulnerabilities in real time
Get breaking CVE alerts, compliance reports (ISO 42001, EU AI Act),
and CISO risk assessments for your AI/ML stack.
Start 14-Day Free Trial