Collaborative penetration testing suite for emerging generative AI algorithms
Petar Radanliev
Problem Space: AI Vulnerabilities and Quantum Threats Generative AI vulnerabilities: model inversion, data poisoning, adversarial inputs. Quantum...
2,077+ academic papers on AI security, attacks, and defenses
Showing 801–820 of 973 papers
Clear filtersPetar Radanliev
Problem Space: AI Vulnerabilities and Quantum Threats Generative AI vulnerabilities: model inversion, data poisoning, adversarial inputs. Quantum...
Thomas Wang, Haowen Li
As large language models (LLMs) are increasingly integrated into real-world applications, ensuring their safety, robustness, and privacy compliance...
Alexander Nemecek, Zebin Yun, Zahra Rahmani +4 more
As large language models (LLMs) become progressively more embedded in clinical decision-support, documentation, and patient-information systems,...
Marco Alecci, Jordan Samhi, Tegawendé F. Bissyandé +1 more
Mobile apps often embed authentication secrets, such as API keys, tokens, and client IDs, to integrate with cloud services. However, developers often...
Giovanni De Muri, Mark Vero, Robin Staab +1 more
LLMs are often used by downstream users as teacher models for knowledge distillation, compressing their capabilities into memory-efficient models....
Oleksandr Adamov, Anders Carlsson
This paper explores the challenges of cyberattack attribution, specifically APTs, applying the case study approach for the WhisperGate cyber...
Yixuan Liu, Xinlei Li, Yi Li
Phishing attacks in Web3 ecosystems are increasingly sophisticated, exploiting deceptive contract logic, malicious frontend scripts, and token...
Yushi Yang, Shreyansh Padarha, Andrew Lee +1 more
Agentic reinforcement learning (RL) trains large language models to autonomously call tools during reasoning, with search as the most common...
Rishi Jha, Harold Triedman, Justin Wagle +1 more
Control-flow hijacking attacks manipulate orchestration mechanisms in multi-agent systems into performing unsafe actions that compromise the system...
Runlin Lei, Lu Yi, Mingguo He +4 more
While Graph Neural Networks (GNNs) and Large Language Models (LLMs) are powerful approaches for learning on Text-Attributed Graphs (TAGs), a...
Elias Hossain, Swayamjit Saha, Somshubhra Roy +1 more
Even when prompts and parameters are secured, transformer language models remain vulnerable because their key-value (KV) cache during inference...
Qiusi Zhan, Angeline Budiman-Chan, Abdelrahman Zayed +3 more
Large language model (LLM) based search agents iteratively generate queries, retrieve external information, and reason to answer open-domain...
Qiusi Zhan, Angeline Budiman-Chan, Abdelrahman Zayed +3 more
Large language model (LLM) based search agents iteratively generate queries, retrieve external information, and reason to answer open-domain...
Bo-Han Feng, Chien-Feng Liu, Yu-Hsuan Li Liang +9 more
Large audio-language models (LALMs) extend text-based LLMs with auditory understanding, offering new opportunities for multimodal applications. While...
Yue Liu, Zhenchang Xing, Shidong Pan +1 more
In recent years, the AI wave has grown rapidly in software development. Even novice developers can now design and generate complex...
Jie Zhang, Meng Ding, Yang Liu +2 more
We present a novel approach for attacking black-box large language models (LLMs) by exploiting their ability to express confidence in natural...
Asmita Mohanty, Gezheng Kang, Lei Gao +1 more
Large Language Models (LLMs) have demonstrated strong performance across diverse tasks, but fine-tuning them typically relies on cloud-based,...
Shivam Ratnakar, Sanjay Raghavendra
Integration of Large Language Models with search/retrieval engines has become ubiquitous, yet these systems harbor a critical vulnerability that...
Xiaofan Li, Xing Gao
The Model Context Protocol (MCP) is an emerging open standard that enables AI-powered applications to interact with external tools through structured...
David Peer, Sebastian Stabinger
Large Language Models (LLMs) have demonstrated impressive capabilities, yet their deployment in high-stakes domains is hindered by inherent...
Get breaking CVE alerts, compliance reports (ISO 42001, EU AI Act), and CISO risk assessments for your AI/ML stack.
Start 14-Day Free Trial