Benchmark HIGH
Rishika Bhagwatkar, Kevin Kasa, Abhay Puri +5 more
AI agents are vulnerable to indirect prompt injection attacks, where malicious instructions embedded in external content or tool outputs cause...
Attack HIGH
Weiliang Zhao, Jinjun Peng, Daniel Ben-Levi +2 more
The proliferation of powerful large language models (LLMs) has necessitated robust safety alignment, yet these models remain vulnerable to evolving...
5 months ago cs.CR cs.CL
PDF
Attack HIGH
Kuofeng Gao, Yiming Li, Chao Du +4 more
Jailbreaking attacks on the vision modality typically rely on imperceptible adversarial perturbations, whereas attacks on the textual modality are...
5 months ago cs.CL cs.AI cs.CR
PDF
Benchmark MEDIUM
Punya Syon Pandey, Hai Son Le, Devansh Bhardwaj +2 more
Large language models (LLMs) are increasingly deployed in contexts where their failures can have direct sociopolitical consequences. Yet, existing...
5 months ago cs.CL cs.AI cs.LG
PDF
Attack HIGH
Yuxin Wen, Arman Zharmagambetov, Ivan Evtimov +4 more
Prompt injection poses a serious threat to the reliability and safety of LLM agents. Recent defenses against prompt injection, such as Instruction...
5 months ago cs.CR cs.LG
PDF
Defense LOW
Siwei Han, Kaiwen Xiong, Jiaqi Liu +9 more
As Large Language Model (LLM) agents increasingly gain self-evolutionary capabilities to adapt and refine their strategies through real-world...
5 months ago cs.LG cs.AI
PDF
Attack HIGH
Santhosh KumarRavindran
The rapid adoption of large language models (LLMs) in enterprise systems exposes vulnerabilities to prompt injection attacks, strategic deception,...
5 months ago cs.CR cs.AI
PDF
Defense MEDIUM
Shuai Zhao, Xinyi Wu, Shiqian Zhao +4 more
During fine-tuning, large language models (LLMs) are increasingly vulnerable to data-poisoning backdoor attacks, which compromise their reliability...
5 months ago cs.CR cs.AI cs.CL
PDF
Attack HIGH
Buyun Liang, Liangzu Peng, Jinqi Luo +3 more
Large Language Models (LLMs) are increasingly deployed in high-risk domains. However, state-of-the-art LLMs often exhibit hallucinations, raising...
5 months ago cs.CL cs.AI cs.CR
PDF
Defense MEDIUM
Anindya Sundar Das, Kangjie Chen, Monowar Bhuyan
Pre-trained language models have achieved remarkable success across a wide range of natural language processing (NLP) tasks, particularly when...
5 months ago cs.CL cs.LG
PDF
Defense MEDIUM
Rui Wu, Yihao Quan, Zeru Shi +3 more
Safety-aligned Large Language Models (LLMs) still show two dominant failure modes: they are easily jailbroken, or they over-refuse harmless inputs...
5 months ago cs.CL cs.LG
PDF
Attack HIGH
Yu Cui, Sicheng Pan, Yifei Liu +2 more
Large language models (LLMs) have been widely deployed in Conversational AIs (CAIs), while exposing privacy and security threats. Recent research...
Attack HIGH
Yanjie Li, Yiming Cao, Dong Wang +1 more
Multimodal agents built on large vision-language models (LVLMs) are increasingly deployed in open-world settings but remain highly vulnerable to...
5 months ago cs.CR cs.AI
PDF
Attack HIGH
Xiangxiang Chen, Peixin Zhang, Jun Sun +2 more
Model quantization is a popular technique for deploying deep learning models on resource-constrained environments. However, it may also introduce...
5 months ago cs.CR cs.AI cs.LG
PDF
Benchmark LOW
Peichao Lai, Jinhui Zhuang, Kexuan Zhang +6 more
Automating the conversion of UI images into web code is a critical task for front-end development and rapid prototyping. Advances in multimodal large...
Tool MEDIUM
Rijha Safdar, Danyail Mateen, Syed Taha Ali +1 more
Artificial Intelligence (AI) and more specifically Large Language Models (LLMs) have demonstrated exceptional progress in multiple areas including...
Survey LOW
Mizanur Rahman, Amran Bhuiyan, Mohammed Saidul Islam +5 more
Recent advances in large language models (LLMs) have enabled a new class of AI agents that automate multiple stages of the data science workflow by...
5 months ago cs.AI cs.CL
PDF
Attack MEDIUM
Guangyu Shen, Siyuan Cheng, Xiangzhe Xu +4 more
Large Language Models (LLMs) can acquire deceptive behaviors through backdoor attacks, where the model executes prohibited actions whenever secret...
5 months ago cs.CR cs.AI
PDF
Benchmark MEDIUM
Jehyeok Yeon, Isha Chaudhary, Gagandeep Singh
Large language models (LLMs) are increasingly deployed in agentic systems where they map user intents to relevant external tools to fulfill a task. A...
5 months ago cs.CR cs.AI
PDF
Benchmark MEDIUM
Chengxiao Wang, Isha Chaudhary, Qian Hu +3 more
Large Language Models (LLMs) can produce catastrophic responses in conversational settings that pose serious risks to public safety and security....
5 months ago cs.AI cs.CR cs.LG
PDF
Track AI security vulnerabilities in real time
Get breaking CVE alerts, compliance reports (ISO 42001, EU AI Act),
and CISO risk assessments for your AI/ML stack.
Start 14-Day Free Trial