Skip to main content

Showing 1–14 of 14 results for author: Shayegani, E

Searching in archive cs. Search in all archives.
.
  1. arXiv:2604.02486  [pdf, ps, other

    cs.CV cs.CL

    VLMs Need Words: Vision Language Models Ignore Visual Detail In Favor of Semantic Anchors

    Authors: Haz Sameen Shahgir, Xiaofu Chen, Yu Fu, Erfan Shayegani, Nael Abu-Ghazaleh, Yova Kementchedjhieva, Yue Dong

    Abstract: Vision Language Models (VLMs) achieve impressive performance across a wide range of multimodal tasks. However, on some tasks that demand fine-grained visual perception, they often fail even when the required information is present in their internal representations. In this work, we demonstrate that this gap arises from their narrow training pipeline which focuses on moving visual information to th… ▽ More

    Submitted 2 April, 2026; originally announced April 2026.

  2. arXiv:2511.03143  [pdf, ps, other

    cs.HC cs.AI cs.CL cs.CY cs.LG

    From Measurement to Expertise: Empathetic Expert Adapters for Context-Based Empathy in Conversational AI Agents

    Authors: Erfan Shayegani, Jina Suh, Andy Wilson, Nagu Rangan, Javier Hernandez

    Abstract: Empathy is a critical factor in fostering positive user experiences in conversational AI. While models can display empathy, it is often generic rather than tailored to specific tasks and contexts. In this work, we introduce a novel framework for developing and evaluating context-specific empathetic large language models (LLMs). We first analyze a real-world conversational dataset consisting of 672… ▽ More

    Submitted 4 November, 2025; originally announced November 2025.

  3. arXiv:2510.22437  [pdf, ps, other

    cs.AI cs.CL

    Modeling Hierarchical Thinking in Large Reasoning Models

    Authors: G M Shahariar, Ali Nazari, Erfan Shayegani, Nael Abu-Ghazaleh

    Abstract: Large Language Models (LLMs) have demonstrated remarkable reasoning abilities when they generate step-by-step solutions, known as chain-of-thought (CoT) reasoning. When trained to using chain-of-thought reasoning examples, the resulting models (called Large Reasoning Models, or LRMs) appear to learn hierarchical thinking strategies similar to those used by humans. However, understanding LRMs emerg… ▽ More

    Submitted 25 October, 2025; originally announced October 2025.

  4. arXiv:2510.01670  [pdf, ps, other

    cs.AI cs.CL cs.CR cs.CY cs.LG

    Just Do It!? Computer-Use Agents Exhibit Blind Goal-Directedness

    Authors: Erfan Shayegani, Keegan Hines, Yue Dong, Nael Abu-Ghazaleh, Roman Lutz, Spencer Whitehead, Vidhisha Balachandran, Besmira Nushi, Vibhav Vineet

    Abstract: Computer-Use Agents (CUAs) are an increasingly deployed class of agents that take actions on GUIs to accomplish user goals. In this paper, we show that CUAs consistently exhibit Blind Goal-Directedness (BGD): a bias to pursue goals regardless of feasibility, safety, reliability, or context. We characterize three prevalent patterns of BGD: (i) lack of contextual reasoning, (ii) assumptions and deci… ▽ More

    Submitted 2 October, 2025; originally announced October 2025.

  5. arXiv:2509.16437  [pdf, ps, other

    cs.HC cs.AI

    SENSE-7: Taxonomy and Dataset for Measuring User Perceptions of Empathy in Sustained Human-AI Conversations

    Authors: Jina Suh, Lindy Le, Erfan Shayegani, Gonzalo Ramos, Judith Amores, Desmond C. Ong, Mary Czerwinski, Javier Hernandez

    Abstract: Empathy is increasingly recognized as a key factor in human-AI communication, yet conventional approaches to "digital empathy" often focus on simulating internal, human-like emotional states while overlooking the inherently subjective, contextual, and relational facets of empathy as perceived by users. In this work, we propose a human-centered taxonomy that emphasizes observable empathic behaviors… ▽ More

    Submitted 19 September, 2025; originally announced September 2025.

  6. arXiv:2509.15213  [pdf, ps, other

    cs.CR

    Evil Vizier: Vulnerabilities of LLM-Integrated XR Systems

    Authors: Yicheng Zhang, Zijian Huang, Sophie Chen, Erfan Shayegani, Jiasi Chen, Nael Abu-Ghazaleh

    Abstract: Extended reality (XR) applications increasingly integrate Large Language Models (LLMs) to enhance user experience, scene understanding, and even generate executable XR content, and are often called "AI glasses". Despite these potential benefits, the integrated XR-LLM pipeline makes XR applications vulnerable to new forms of attacks. In this paper, we analyze LLM-Integated XR systems in the literat… ▽ More

    Submitted 27 September, 2025; v1 submitted 18 September, 2025; originally announced September 2025.

  7. arXiv:2504.03735  [pdf, other

    cs.CR cs.AI cs.CL cs.CY cs.LG

    Misaligned Roles, Misplaced Images: Structural Input Perturbations Expose Multimodal Alignment Blind Spots

    Authors: Erfan Shayegani, G M Shahariar, Sara Abdali, Lei Yu, Nael Abu-Ghazaleh, Yue Dong

    Abstract: Multimodal Language Models (MMLMs) typically undergo post-training alignment to prevent harmful content generation. However, these alignment stages focus primarily on the assistant role, leaving the user role unaligned, and stick to a fixed input prompt structure of special tokens, leaving the model vulnerable when inputs deviate from these expectations. We introduce Role-Modality Attacks (RMA), a… ▽ More

    Submitted 31 March, 2025; originally announced April 2025.

  8. arXiv:2411.04291  [pdf, ps, other

    cs.CL cs.CV

    Layer-wise Alignment: Examining Safety Alignment Across Image Encoder Layers in Vision Language Models

    Authors: Saketh Bachu, Erfan Shayegani, Rohit Lal, Trishna Chakraborty, Arindam Dutta, Chengyu Song, Yue Dong, Nael Abu-Ghazaleh, Amit K. Roy-Chowdhury

    Abstract: Vision-language models (VLMs) have improved significantly in their capabilities, but their complex architecture makes their safety alignment challenging. In this paper, we reveal an uneven distribution of harmful information across the intermediate layers of the image encoder and show that skipping a certain set of layers and exiting early can increase the chance of the VLM generating harmful resp… ▽ More

    Submitted 19 June, 2025; v1 submitted 6 November, 2024; originally announced November 2024.

    Comments: Accepted by ICML 2025 as a spotlight poster

  9. arXiv:2406.02575  [pdf, ps, other

    cs.CL cs.CR cs.LG

    Cross-Modal Safety Alignment: Is textual unlearning all you need?

    Authors: Trishna Chakraborty, Erfan Shayegani, Zikui Cai, Nael Abu-Ghazaleh, M. Salman Asif, Yue Dong, Amit K. Roy-Chowdhury, Chengyu Song

    Abstract: Recent studies reveal that integrating new modalities into Large Language Models (LLMs), such as Vision-Language Models (VLMs), creates a new attack surface that bypasses existing safety training techniques like Supervised Fine-tuning (SFT) and Reinforcement Learning with Human Feedback (RLHF). While further SFT and RLHF-based safety training can be conducted in multi-modal settings, collecting mu… ▽ More

    Submitted 14 October, 2025; v1 submitted 27 May, 2024; originally announced June 2024.

    Comments: Accepted by EMNLP 2024 Findings

  10. arXiv:2403.12503  [pdf, ps, other

    cs.CR cs.AI cs.LG

    Securing Large Language Models: Threats, Vulnerabilities and Responsible Practices

    Authors: Sara Abdali, Richard Anarfi, CJ Barberan, Jia He, Erfan Shayegani

    Abstract: Large language models (LLMs) have significantly transformed the landscape of Natural Language Processing (NLP). Their impact extends across a diverse spectrum of tasks, revolutionizing how we approach language understanding and generations. Nevertheless, alongside their remarkable utility, LLMs introduce critical security and risk considerations. These challenges warrant careful examination to ens… ▽ More

    Submitted 11 June, 2025; v1 submitted 19 March, 2024; originally announced March 2024.

  11. arXiv:2310.10844  [pdf, other

    cs.CL cs.CR cs.LG

    Survey of Vulnerabilities in Large Language Models Revealed by Adversarial Attacks

    Authors: Erfan Shayegani, Md Abdullah Al Mamun, Yu Fu, Pedram Zaree, Yue Dong, Nael Abu-Ghazaleh

    Abstract: Large Language Models (LLMs) are swiftly advancing in architecture and capability, and as they integrate more deeply into complex systems, the urgency to scrutinize their security properties grows. This paper surveys research in the emerging interdisciplinary field of adversarial attacks on LLMs, a subfield of trustworthy ML, combining the perspectives of Natural Language Processing and Security.… ▽ More

    Submitted 16 October, 2023; originally announced October 2023.

  12. arXiv:2308.09146  [pdf, other

    cs.CR

    That Doesn't Go There: Attacks on Shared State in Multi-User Augmented Reality Applications

    Authors: Carter Slocum, Yicheng Zhang, Erfan Shayegani, Pedram Zaree, Nael Abu-Ghazaleh, Jiasi Chen

    Abstract: Augmented Reality (AR) is expected to become a pervasive component in enabling shared virtual experiences. In order to facilitate collaboration among multiple users, it is crucial for multi-user AR applications to establish a consensus on the "shared state" of the virtual world and its augmentations, through which they interact within augmented reality spaces. Current methods to create and access… ▽ More

    Submitted 8 March, 2024; v1 submitted 17 August, 2023; originally announced August 2023.

    Comments: Accepted by USENIX Security 2024

  13. arXiv:2307.14539  [pdf, other

    cs.CR cs.CL

    Jailbreak in pieces: Compositional Adversarial Attacks on Multi-Modal Language Models

    Authors: Erfan Shayegani, Yue Dong, Nael Abu-Ghazaleh

    Abstract: We introduce new jailbreak attacks on vision language models (VLMs), which use aligned LLMs and are resilient to text-only jailbreak attacks. Specifically, we develop cross-modality attacks on alignment where we pair adversarial images going through the vision encoder with textual prompts to break the alignment of the language model. Our attacks employ a novel compositional strategy that combines… ▽ More

    Submitted 10 October, 2023; v1 submitted 26 July, 2023; originally announced July 2023.

  14. arXiv:2307.08811  [pdf, other

    cs.LG cs.IT

    Co(ve)rtex: ML Models as storage channels and their (mis-)applications

    Authors: Md Abdullah Al Mamun, Quazi Mishkatul Alam, Erfan Shayegani, Pedram Zaree, Ihsen Alouani, Nael Abu-Ghazaleh

    Abstract: Machine learning (ML) models are overparameterized to support generality and avoid overfitting. The state of these parameters is essentially a "don't-care" with respect to the primary model provided that this state does not interfere with the primary model. In both hardware and software systems, don't-care states and undefined behavior have been shown to be sources of significant vulnerabilities.… ▽ More

    Submitted 11 May, 2024; v1 submitted 17 July, 2023; originally announced July 2023.