Attack HIGH
Yige Liu, Yiwei Lou, Che Wang +2 more
As a distributed collaborative machine learning paradigm, vertical federated learning (VFL) allows multiple passive parties with distinct features...
4 weeks ago cs.LG cs.CR
PDF
Attack HIGH
David Schmotz, Luca Beurer-Kellner, Sahar Abdelnabi +1 more
LLM agents are evolving rapidly, powered by code execution, tools, and the recently introduced agent skills feature. Skills allow users to extend LLM...
1 months ago cs.CR cs.LG
PDF
Tool HIGH
Ian Steenstra, Paola Pedrelli, Weiyan Shi +2 more
Large Language Models (LLMs) are increasingly utilized for mental health support; however, current safety benchmarks often fail to detect the...
1 months ago cs.CL cs.AI cs.CY
PDF
Attack HIGH
Nadav Kadvil, Malak Fares, Ayellet Tal
Large Vision-Language Models (LVLMs) can be vulnerable to adversarial images that subtly bias their outputs toward plausible yet incorrect responses....
Attack HIGH
Xiaochong Jiang, Shiqi Yang, Wenting Yang +2 more
Agentic systems built on large language models (LLMs) extend beyond text generation to autonomously retrieve information and invoke tools. This...
1 months ago cs.CR cs.AI
PDF
Tool HIGH
Xingyu Shen, Tommy Duong, Xiaodong An +6 more
Age estimation systems are increasingly deployed as gatekeepers for age-restricted online content, yet their robustness to cosmetic modifications has...
1 months ago cs.CV cs.CR cs.LG
PDF
Survey HIGH
Kunal Mukherjee
Trusted Execution Environments (TEEs) (e.g., Intel SGX and ArmTrustZone) aim to protect sensitive computation from a compromised operating system,...
1 months ago cs.CR cs.AI
PDF
Attack HIGH
Amirhossein Farzam, Majid Behabahani, Mani Malek +2 more
Large language models (LLMs) remain vulnerable to jailbreak prompts that are fluent and semantically coherent, and therefore difficult to detect with...
Attack HIGH
Charles Ye, Jasmine Cui, Dylan Hadfield-Menell
Language models remain vulnerable to prompt injection attacks despite extensive safety training. We trace this failure to role confusion: models...
1 months ago cs.CL cs.AI cs.CR
PDF
Attack HIGH
Charles Ye, Jasmine Cui, Dylan Hadfield-Menell
Language models remain vulnerable to prompt injection attacks despite extensive safety training. We trace this failure to role confusion: models...
1 months ago cs.CL cs.AI cs.CR
PDF
Attack HIGH
Sieun Kim, Yeeun Jo, Sungmin Na +5 more
Red-teaming, where adversarial prompts are crafted to expose harmful behaviors and assess risks, offers a dynamic approach to surfacing underlying...
Attack HIGH
Shenyang Chen, Liuwan Zhu
Standard evaluations of backdoor attacks on text-to-image (T2I) models primarily measure trigger activation and visual fidelity. We challenge this...
1 months ago cs.CR cs.AI
PDF
Attack HIGH
Zafir Shamsi, Nikhil Chekuru, Zachary Guzman +1 more
Large Language Models (LLMs) are increasingly integrated into high-stakes applications, making robust safety guarantees a central practical and...
1 months ago cs.CL cs.AI
PDF
Benchmark HIGH
Mirae Kim, Seonghun Jeong, Youngjun Kwak
Jailbreaking poses a significant risk to the deployment of Large Language Models (LLMs) and Vision Language Models (VLMs). VLMs are particularly...
1 months ago cs.CL cs.AI cs.DB
PDF
Tool HIGH
Phan The Duy, Nghi Hoang Khoa, Nguyen Tran Anh Quan +3 more
The increasing deployment of Federated Learning (FL) in Intrusion Detection Systems (IDS) introduces new challenges related to data privacy,...
1 months ago cs.CR cs.AI
PDF
Attack HIGH
Jingkai Guo, Chaitali Chakrabarti, Deliang Fan
Large language models (LLMs) are increasingly deployed in safety and security critical applications, raising concerns about their robustness to model...
1 months ago cs.CR cs.CL cs.LG
PDF
Attack HIGH
Manuel Wirth
As Large Language Models (LLMs) are increasingly integrated into automated decision-making pipelines, specifically within Human Resources (HR), the...
1 months ago cs.CR cs.AI
PDF
Attack HIGH
Xinhao Deng, Jiaqing Wu, Miao Chen +3 more
Agent hijacking, highlighted by OWASP as a critical threat to the Large Language Model (LLM) ecosystem, enables adversaries to manipulate execution...
1 months ago cs.AI cs.LG
PDF
Benchmark HIGH
Priyaranjan Pattnayak, Sanchari Chowdhuri
Safety alignment of large language models (LLMs) is mostly evaluated in English and contract-bound, leaving multilingual vulnerabilities...
1 months ago cs.AI cs.CL
PDF
Attack HIGH
Thomas Michel, Debabrota Basu, Emilie Kaufmann
Modern AI models are not static. They go through multiple updates in their lifecycles. Thus, exploiting the model dynamics to create stronger...
1 months ago cs.LG cs.CR math.ST
PDF
Track AI security vulnerabilities in real time
Get breaking CVE alerts, compliance reports (ISO 42001, EU AI Act),
and CISO risk assessments for your AI/ML stack.
Start 14-Day Free Trial