Attack MEDIUM
Zaixi Zhang, Souradip Chakraborty, Amrit Singh Bedi +16 more
The rapid adoption of generative artificial intelligence (GenAI) in the biosciences is transforming biotechnology, medicine, and synthetic biology....
5 months ago cs.CR q-bio.BM
PDF
Attack MEDIUM
Tiarnaigh Downey-Webb, Olamide Jogunola, Oluwaseun Ajao
This paper presents a systematic security assessment of four prominent Large Language Models (LLMs) against diverse adversarial attack vectors. We...
5 months ago cs.CR cs.AI cs.CY
PDF
Attack MEDIUM
Brandon Lit, Edward Crowder, Daniel Vogel +1 more
AI chatbots are an emerging security attack vector, vulnerable to threats such as prompt injection, and rogue chatbot creation. When deployed in...
Attack MEDIUM
Abhishek K. Mishra, Antoine Boutet, Lucas Magnana
Large Language Models (LLMs) are increasingly deployed across multilingual applications that handle sensitive data, yet their scale and linguistic...
5 months ago cs.CL cs.CR
PDF
Attack MEDIUM
Aofan Liu, Lulu Tang
Vision-Language Models (VLMs) have garnered significant attention for their remarkable ability to interpret and generate multimodal content. However,...
5 months ago cs.CR cs.AI
PDF
Attack MEDIUM
Jiyang Qiu, Xinbei Ma, Yunqing Xu +2 more
The rapid deployment of large language model (LLM)-based agents in real-world applications has raised serious concerns about their trustworthiness....
Attack MEDIUM
Tavish McDonald, Bo Lei, Stanislav Fort +2 more
Models are susceptible to adversarially out-of-distribution (OOD) data despite large training-compute investments into their robustification. Zaremba...
Attack MEDIUM
Tiancheng Xing, Jerry Li, Yixuan Du +1 more
Large language models (LLMs) are increasingly used as rerankers in information retrieval, yet their ranking behavior can be steered by small,...
5 months ago cs.CL cs.AI cs.IR
PDF
Attack MEDIUM
Zizhao Wang, Dingcheng Li, Vaishakh Keshava +4 more
Large Language Model (LLM) agents can leverage tools such as Google Search to complete complex tasks. However, this tool usage introduces the risk of...
5 months ago cs.LG cs.AI cs.CL
PDF
Attack MEDIUM
Guangyu Shen, Siyuan Cheng, Xiangzhe Xu +4 more
Large Language Models (LLMs) can acquire deceptive behaviors through backdoor attacks, where the model executes prohibited actions whenever secret...
5 months ago cs.CR cs.AI
PDF
Attack MEDIUM
Tanqiu Jiang, Min Bai, Nikolaos Pappas +2 more
Vision-language model (VLM)-based web agents increasingly power high-stakes selection tasks like content recommendation or product ranking by...
5 months ago cs.AI cs.CR
PDF
Attack MEDIUM
Fatmazohra Rezkellah, Ramzi Dakhmouche
With the increasing adoption of Large Language Models (LLMs), more customization is needed to ensure privacy-preserving and safe generation. We...
5 months ago cs.LG cs.CL cs.CR
PDF
Attack MEDIUM
Abrar Shahid, Ibteeker Mahir Ishum, AKM Tahmidul Haque +2 more
This paper presents a controlled study of adversarial reinforcement learning in network security through a custom OpenAI Gym environment that models...
5 months ago cs.LG cs.AI cs.CR
PDF
Attack MEDIUM
Davide Gabrielli, Simone Sestito, Iacopo Masi
The current landscape of defensive mechanisms for LLMs is fragmented and underdeveloped, unlike prior work on classifiers. To further promote...
Attack MEDIUM
Youwei Bao, Shuhan Yang, Hyunsoo Yang
Deterministic pseudo random number generators (PRNGs) used in generative artificial intelligence (GAI) models produce predictable patterns vulnerable...
5 months ago cs.LG cond-mat.mtrl-sci physics.data-an
PDF
Attack MEDIUM
Zhenyu Pan, Yiting Zhang, Zhuo Liu +13 more
LLM-based multi-agent systems excel at planning, tool use, and role coordination, but their openness and interaction complexity also expose them to...
Attack MEDIUM
Jaiden Fairoze, Sanjam Garg, Keewoo Lee +1 more
As large language models (LLMs) advance, ensuring AI safety and alignment is paramount. One popular approach is prompt guards, lightweight mechanisms...
5 months ago cs.LG cs.CR
PDF
Attack MEDIUM
Yen-Shan Chen, Sian-Yao Huang, Cheng-Lin Yang +1 more
Existing data poisoning attacks on retrieval-augmented generation (RAG) systems scale poorly because they require costly optimization of poisoned...
5 months ago cs.LG cs.CL cs.CR
PDF
Attack MEDIUM
Tsubasa Takahashi, Shojiro Yamabe, Futa Waseda +1 more
Differential Attention (DA) has been proposed as a refinement to standard attention, suppressing redundant or noisy context through a subtractive...
5 months ago cs.LG cs.CR
PDF
Attack MEDIUM
Yu Yan, Siqi Lu, Yang Gao +4 more
Recently, Bit-Flip Attack (BFA) has garnered widespread attention for its ability to compromise software system integrity remotely through hardware...
Track AI security vulnerabilities in real time
Get breaking CVE alerts, compliance reports (ISO 42001, EU AI Act),
and CISO risk assessments for your AI/ML stack.
Start 14-Day Free Trial