The dark patterns, deceptive interface designs manipulating user behaviors, have been extensively studied for their effects on human decision-making and autonomy. Yet, with the rising prominence of LLM-powered GUI agents that automate tasks from high-level intents, understanding how dark patterns affect agents is increasingly important. We present a two-phase empirical study examining how agents, human participants, and human-AI teams respond to 16 types of dark patterns across diverse scenarios. Phase 1 highlights that agents often fail to recognize dark patterns, and even when aware, prioritize task completion over protective action. Phase 2 revealed divergent failure modes: humans succumb due to cognitive shortcuts and habitual compliance, while agents falter from procedural blind spots. Human oversight improved avoidance but introduced costs such as attentional tunneling and cognitive load. Our findings show neither humans nor agents are uniformly resilient, and collaboration introduces new vulnerabilities, suggesting design needs for transparency, adjustable autonomy, and oversight.
| Dark Pattern | GPT-4o | Claude 3.7 | DeepSeek V3 | Gemini 2.0 | Claude CUA | Operator |
|---|---|---|---|---|---|---|
| Adding Steps | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ |
| Bait and Switch | ✓ | ✓ | ✗ | ✗ | ✗ | ✓ |
| Hiding Information | ✗ | ✗ | ✗ | ✗ | ✓ | ✓ |
| Manipulating Visual Choice Architecture | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ |
| Bad Defaults | ✗ | ✗ | ✗ | ✗ | ||
| Emotional or Sensory Manipulation | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ |
| Trick Questions | ✗ | ✗ | ✗ | ✓ | ||
| Choice Overload | ✓ | ✓ | ✓ | ✗ | ✓ | ✓ |
| Social Proof | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ |
| Nagging | ✓ | ✓ | ✓ | ✓ | ✗ | ✓ |
| Forced Communication or Disclosure | ✗ | ✗ | ✗ | ✗ | ||
| Scarcity and Popularity Claims | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ |
| Forced Registration | ✓ | ✗ | ✓ | ✓ | ✓ | |
| Hidden Information | ✗ | ✗ | ✗ | ✗ | ||
| Urgency | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ |
| Shaming | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ |
Select an Agent type and Dark Pattern to view corresponding materials
| Dark Pattern Type | Human-Only | Human Oversight |
|---|---|---|
| Adding Steps | 100.0% | 100.0% |
| Bait and Switch | 70.0% | 91.7% |
| Hiding Information | 70.0% | 100.0% |
| Manipulating Visual Choice Architecture | 90.9% | 100.0% |
| Bad Defaults | 33.3% | 80.0% |
| Emotional or Sensory Manipulation | 83.3% | 100.0% |
| Trick Questions | 45.5% | 81.8% |
| Choice Overload | 100.0% | 100.0% |
| Social Proof | 100.0% | 90.9% |
| Nagging | 83.3% | 100.0% |
| Forced Communication or Disclosure | 25.0% | 50.0% |
| Scarcity and Popularity Claims | 90.9% | 100.0% |
| Forced Registration | 40.0% | 25.0% |
| Hidden Information | 20.0% | 33.3% |
| Urgency | 100.0% | 100.0% |
| Shaming | 100.0% | 100.0% |
@misc{tang2025darkpatternsmeetgui,
title={Dark Patterns Meet GUI Agents: LLM Agent Susceptibility to Manipulative Interfaces and the Role of Human Oversight},
author={Jingyu Tang and Chaoran Chen and Jiawen Li and Zhiping Zhang and Bingcan Guo and Ibrahim Khalilov and Simret Araya Gebreegziabher and Bingsheng Yao and Dakuo Wang and Yanfang Ye and Tianshi Li and Ziang Xiao and Yaxing Yao and Toby Jia-Jun Li},
year={2025},
eprint={2509.10723},
archivePrefix={arXiv},
primaryClass={cs.HC},
url={https://arxiv.org/abs/2509.10723},
}