Attack HIGH
Kun Wang, Meng Chen, Junhao Wang +6 more
With the widespread deployment of deep-learning-based speech models in security-critical applications, backdoor attacks have emerged as a serious...
1 weeks ago cs.CR cs.LG cs.SD
PDF
Attack MEDIUM
Saikat Maiti
Autonomous AI agents powered by large language models are being deployed in production with capabilities including shell execution, file system...
1 weeks ago cs.CR cs.AI
PDF
Attack HIGH
Zhihua Wei, Qiang Li, Jian Ruan +4 more
Large vision-language models (VLMs) often exhibit weakened safety alignment with the integration of the visual modality. Even when text prompts...
1 weeks ago cs.CV cs.AI
PDF
Attack HIGH
Hammad Atta, Ken Huang, Kyriakos Rock Lambros +11 more
Agentic LLM systems equipped with persistent memory, RAG pipelines, and external tool connectors face a class of attacks - Logic-layer Prompt Control...
Attack MEDIUM
Patrick Levi
Retrieval augmented generation systems have become an integral part of everyday life. Whether in internet search engines, email systems, or service...
1 weeks ago cs.CR cs.AI
PDF
Attack HIGH
Shenao Yan, Shimaa Ahmed, Shan Jin +4 more
Code generation large language models (LLMs) are increasingly integrated into modern software development workflows. Recent work has shown that these...
1 weeks ago cs.CR cs.AI cs.SE
PDF
Attack MEDIUM
Kushankur Ghosh, Mehar Klair, Kian Kyars +2 more
Provenance graphs model causal system-level interactions from logs, enabling anomaly detectors to learn normal behavior and detect deviations as...
1 weeks ago cs.CR cs.LG
PDF
Attack HIGH
Yong Zou, Haoran Li, Fanxiao Li +5 more
Recent progress in image generation models (IGMs) enables high-fidelity content creation but also amplifies risks, including the reproduction of...
1 weeks ago cs.CV cs.AI cs.CR
PDF
Attack HIGH
Guangsheng Zhang, Huan Tian, Leo Zhang +4 more
Semantic segmentation models are widely deployed in safety-critical applications such as autonomous driving, yet their vulnerability to backdoor...
Attack HIGH
Deng Liu, Song Chen
Hardware faults, specifically bit-flips in quantized weights, pose a severe reliability threat to Large Language Models (LLMs), often triggering...
Attack HIGH
Xiaobing Sun, Perry Lam, Shaohua Li +4 more
Modern LLMs employ safety mechanisms that extend beyond surface-level input filtering to latent semantic representations and generation-time...
Attack MEDIUM
Amira Guesmi, Muhammad Shafique
Vision-language models (VLMs) have recently shown remarkable capabilities in visual understanding and generation, but remain vulnerable to...
1 weeks ago cs.CR cs.CV
PDF
Attack HIGH
Mateusz Dziemian, Maxwell Lin, Xiaohan Fu +28 more
LLM based agents are increasingly deployed in high stakes settings where they process external data sources such as emails, documents, and code...
1 weeks ago cs.CR cs.AI
PDF
Attack HIGH
Zhenlin Xu, Xiaogang Zhu, Yu Yao +2 more
Modern agentic systems allow Large Language Model (LLM) agents to tackle complex tasks through extensive tool usage, forming structured control flows...
Attack MEDIUM
Ruyi Zhang, Heng Gao, Songlei Jian +2 more
Backdoor attacks compromise model reliability by using triggers to manipulate outputs. Trigger inversion can accurately locate these triggers via a...
1 weeks ago cs.CR cs.AI
PDF
Attack HIGH
Maël Jenny, Jérémie Dentan, Sonia Vanier +1 more
Most jailbreak techniques for Large Language Models (LLMs) primarily rely on prompt modifications, including paraphrasing, obfuscation, or...
Attack HIGH
Chongxin Li, Hanzhang Wang, Lian Duan
Safety prompts constitute an interpretable layer of defense against jailbreak attacks in vision-language models (VLMs); however, their efficacy is...
Attack HIGH
Yiling Tao, Xinran Zheng, Shuo Yang +2 more
While large language model-based agents demonstrate great potential in collaborative tasks, their interactivity also introduces security...
Attack HIGH
Zijian Ling, Pingyi Hu, Xiuyong Gao +6 more
Speech-driven large language models (LLMs) are increasingly accessed through speech interfaces, introducing new security risks via open acoustic...
1 weeks ago cs.CR cs.AI cs.SD
PDF
Attack MEDIUM
Md. Abdul Awal, Mrigank Rochan, Chanchal K. Roy
Large language models for code have achieved strong performance across diverse software analytics tasks, yet their real-world adoption remains...
Track AI security vulnerabilities in real time
Get breaking CVE alerts, compliance reports (ISO 42001, EU AI Act),
and CISO risk assessments for your AI/ML stack.
Start 14-Day Free Trial