Defense LOW
Daniel Fein, Max Lamparth, Violet Xiang +2 more
Reward Models (RMs) are crucial for online alignment of language models (LMs) with human preferences. However, RM-based preference-tuning is...
1 months ago cs.CL cs.AI
PDF
Benchmark HIGH
Junhyeok Lee, Han Jang, Kyu Sung Choi
Large Language Models (LLMs) and Retrieval-Augmented Generation (RAG) systems are increasingly integrated into clinical workflows; however, prompt...
1 months ago cs.CL cs.LG
PDF
Benchmark MEDIUM
Navita Goyal, Hal Daumé
Model steering, which involves intervening on hidden representations at inference time, has emerged as a lightweight alternative to finetuning for...
1 months ago cs.LG cs.AI cs.CL
PDF
Benchmark MEDIUM
José Ramón Pareja Monturiol, Juliette Sinnott, Roger G. Melko +1 more
Machine learning in clinical settings must balance predictive accuracy, interpretability, and privacy. Models such as logistic regression (LR) offer...
1 months ago cs.LG cs.CR quant-ph
PDF
Attack HIGH
Xin Chen, Jie Zhang, Florian Tramèr
Prompt injection is one of the most critical vulnerabilities in LLM agents; yet, effective automated attacks remain largely unexplored from an...
1 months ago cs.LG cs.AI
PDF
Benchmark LOW
Rui Jia, Ruiyi Lan, Fengrui Liu +7 more
Large language models (LLMs) have advanced the development of personalized learning in education. However, their inherent generation mechanisms often...
Attack MEDIUM
Tao Huang, Rui Wang, Xiaofei Liu +3 more
%Large vision-language models (LVLMs) have shown substantial advances in multimodal understanding and generation. However, when presented with...
Defense MEDIUM
Rohan Subramanian Thomas, Shikhar Shiromani, Abdullah Chaudhry +4 more
Prompt design significantly impacts the moral competence and safety alignment of large language models (LLMs), yet empirical comparisons remain...
1 months ago cs.AI cs.CL
PDF
Attack HIGH
Takashi Koide, Hiroki Nakano, Daiki Chiba
Phishing sites continue to grow in volume and sophistication. Recent work leverages large language models (LLMs) to analyze URLs, HTML, and rendered...
Attack HIGH
Yao Zhou, Zeen Song, Wenwen Qiang +4 more
Safety alignment mechanisms in Large Language Models (LLMs) often operate as latent internal states, obscuring the model's inherent capabilities....
Benchmark LOW
Nelu D. Radpour
Contemporary benchmarks for agentic artificial intelligence (AI) frequently evaluate safety through isolated task-level accuracy thresholds,...
1 months ago cs.CY cs.AI cs.HC
PDF
Attack HIGH
Zihan Wang, Hongwei Li, Rui Zhang +2 more
Chat template is a common technique used in the training and inference stages of Large Language Models (LLMs). It can transform input and output data...
Defense MEDIUM
Zhenxiong Yu, Zhi Yang, Zhiheng Jin +19 more
As large language models (LLMs) evolve into autonomous agents, their real-world applicability has expanded significantly, accompanied by new security...
1 months ago cs.CR cs.AI
PDF
Attack HIGH
Ziyou Jiang, Lin Shi, Guowei Yang +3 more
Cyber attacks have become a serious threat to the security of software systems. Many organizations have built their security knowledge bases to...
Tool MEDIUM
Guangwei Zhang, Jianing Zhu, Cheng Qian +12 more
We present Copyright Detective, the first interactive forensic system for detecting, analyzing, and visualizing potential copyright risks in LLM...
Attack HIGH
Yunbei Zhang, Yingqiang Ge, Weijie Xu +3 more
Current multimodal red teaming treats images as wrappers for malicious payloads via typography or adversarial noise. These attacks are structurally...
1 months ago cs.CR cs.CV cs.LG
PDF
Attack HIGH
Ethan Rathbun, Wo Wei Lin, Alina Oprea +1 more
Simulated environments are a key piece in the success of Reinforcement Learning (RL), allowing practitioners and researchers to train decision making...
1 months ago cs.CR cs.LG cs.RO
PDF
Attack HIGH
Jafar Isbarov, Murat Kantarcioglu
As AI agents automate critical workloads, they remain vulnerable to indirect prompt injection (IPI) attacks. Current defenses rely on monitoring...
1 months ago cs.CR cs.AI
PDF
Benchmark MEDIUM
Ruixin Yang, Ethan Mendes, Arthur Wang +4 more
Vision-language models (VLMs) have demonstrated strong performance in image geolocation, a capability further sharpened by frontier multimodal large...
1 months ago cs.CR cs.AI
PDF
Attack MEDIUM
Vishruti Kakkad, Paul Chung, Hanan Hibshi +1 more
An exponential growth of Machine Learning and its Generative AI applications brings with it significant security challenges, often referred to as...
1 months ago cs.CR cs.AI
PDF
Track AI security vulnerabilities in real time
Get breaking CVE alerts, compliance reports (ISO 42001, EU AI Act),
and CISO risk assessments for your AI/ML stack.
Start 14-Day Free Trial