Reimagining Safety Alignment with An Image
Yifan Xia, Guorui Chen, Wenqian Yu +3 more
Large language models (LLMs) excel in diverse applications but face dual challenges: generating harmful content under jailbreak attacks and...
2,077+ academic papers on AI security, attacks, and defenses
Showing 1561–1580 of 2,027 papers
Clear filtersYifan Xia, Guorui Chen, Wenqian Yu +3 more
Large language models (LLMs) excel in diverse applications but face dual challenges: generating harmful content under jailbreak attacks and...
Mohammed N. Swileh, Shengli Zhang
Centralized Software-Defined Networking (cSDN) offers flexible and programmable control of networks but suffers from scalability and reliability...
Ruofan Liu, Yun Lin, Zhiyong Huang +1 more
Large language models (LLMs) are increasingly integrated into IT infrastructures, where they process user data according to predefined instructions....
Xin Yao, Haiyang Zhao, Yimin Chen +3 more
The Contrastive Language-Image Pretraining (CLIP) model has significantly advanced vision-language modeling by aligning image-text pairs from...
Kayua Oleques Paim, Rodrigo Brandao Mansilha, Diego Kreutz +2 more
The rapid proliferation of Large Language Models (LLMs) has raised significant concerns about their security against adversarial attacks. In this...
David Lüdke, Tom Wollschläger, Paul Ungermann +2 more
We introduce a novel framework that transforms the resource-intensive (adversarial) prompt optimization problem into an \emph{efficient, amortized...
David Farr, Lynnette Hui Xian Ng, Stephen Prochaska +2 more
Disinformation campaigns can distort public perception and destabilize institutions. Understanding how different populations respond to information...
Md Abdul Hannan, Ronghao Ni, Chi Zhang +3 more
Large language models (LLMs) have demonstrated impressive capabilities across a wide range of coding tasks, including summarization, translation,...
Kathrin Grosse, Nico Ebert
Recent improvement gains in large language models (LLMs) have lead to everyday usage of AI-based Conversational Agents (CAs). At the same time, LLMs...
Chenghao Du, Quanfeng Huang, Tingxuan Tang +3 more
Large Language Models (LLMs) have transformed software development, enabling AI-powered applications known as LLM-based agents that promise to...
Heehwan Kim, Sungjune Park, Daeseon Choi
Large Language Models (LLMs) are generally equipped with guardrails to block the generation of harmful responses. However, existing defenses always...
Arnabh Borah, Md Tanvirul Alam, Nidhi Rastogi
Security applications are increasingly relying on large language models (LLMs) for cyber threat detection; however, their opaque reasoning often...
Alex Irpan, Alexander Matt Turner, Mark Kurzeja +2 more
An LLM's factuality and refusal training can be compromised by simple changes to a prompt. Models often adopt user beliefs (sycophancy) or satisfy...
Zishuo Zheng, Vidhisha Balachandran, Chan Young Park +2 more
As large language model (LLM) based systems take on high-stakes roles in real-world decision-making, they must reconcile competing instructions from...
Seif Ikbarieh, Maanak Gupta, Elmahedi Mahalal
The Internet of Things has expanded rapidly, transforming communication and operations across industries but also increasing the attack surface and...
William Overman, Mohsen Bayati
As increasingly capable agents are deployed, a central safety challenge is how to retain meaningful human control without modifying the underlying...
Aylton Almeida, Laerte Xavier, Marco Tulio Valente
Keeping software systems up to date is essential to avoid technical debt, security vulnerabilities, and the rigidity typical of legacy systems....
Shaked Zychlinski, Yuval Kainan
Large Language Models (LLMs) are susceptible to jailbreak attacks where malicious prompts are disguised using ciphers and character-level encodings...
Yingjia Wang, Ting Qiao, Xing Liu +3 more
The rapid advancement of deep neural networks (DNNs) heavily relies on large-scale, high-quality datasets. However, unauthorized commercial use of...
David Schmotz, Sahar Abdelnabi, Maksym Andriushchenko
Enabling continual learning in LLMs remains a key unresolved research challenge. In a recent announcement, a frontier LLM company made a step towards...
Get breaking CVE alerts, compliance reports (ISO 42001, EU AI Act), and CISO risk assessments for your AI/ML stack.
Start 14-Day Free Trial