AgentA/B: Automated and Scalable Web A/BTesting with Interactive LLM Agents
Wang, Dakuo, Hsu, Ting-Yao, Lu, Yuxuan, Gu, Hansu, Cui, Limeng, Xie, Yaochen, Headean, William, Yao, Bingsheng, Veeragouni, Akash, Liu, Jiapeng, Nag, Sreyashi, Wang, Jessie
–arXiv.org Artificial Intelligence
A/B testing experiment is a widely adopted method for evaluating UI/UX design decisions in modern web applications. Yet, traditional A/B testing remains constrained by its dependence on the large-scale and live traffic of human participants, and the long time of waiting for the testing result. Through formative interviews with six experienced industry practitioners, we identified critical bottlenecks in current A/B testing workflows. In response, we present AgentA/B, a novel system that leverages Large Language Model-based autonomous agents (LLM Agents) to automatically simulate user interaction behaviors with real webpages. AgentA/B enables scalable deployment of LLM agents with diverse personas, each capable of navigating the dynamic webpage and interactively executing multi-step interactions like search, clicking, filtering, and purchasing. In a demonstrative controlled experiment, we employ AgentA/B to simulate a between-subject A/B testing with 1,000 LLM agents Amazon.com, and compare agent behaviors with real human shopping behaviors at a scale. Our findings suggest AgentA/B can emulate human-like behavior patterns.
arXiv.org Artificial Intelligence
Sep-22-2025
- Country:
- North America > United States
- Florida > Palm Beach County
- Boca Raton (0.04)
- Massachusetts > Suffolk County
- Boston (0.40)
- New York > New York County
- New York City (0.04)
- Pennsylvania > Centre County
- State College (0.04)
- Texas > Travis County
- Austin (0.04)
- Florida > Palm Beach County
- North America > United States
- Genre:
- Research Report
- Experimental Study (1.00)
- New Finding (1.00)
- Research Report
- Industry:
- Information Technology > Services (0.46)
- Retail > Online (0.34)
- Technology: