The Digital Twins Lab explores digital twin technology and its real-world use cases with a commitment to transparency, independence, interdisciplinary collaboration, rigor, and accessibility. We develop and evaluate methods, datasets, and tools that help researchers and practitioners understand when digital twins work, when they fail, and how they can be deployed responsibly.
Digital twin technology leverages large language models to simulate human behavior and complex systems, enabling rapid testing, experimentation, and iteration. It has the potential to transform marketing, social science, behavioral research, and business operations, among many other domains. More broadly, digital twins enable high-precision personalization, efficient synthetic data generation, and scalable reinforcement-learning environments—forming a foundation for world models and effective human–AI collaboration.
Our research spans measurement and evaluation, human–AI alignment, domain-specific applications, and reproducible research practices. We emphasize impact beyond the lab by producing open resources, clear reporting standards, and public-facing artifacts that support broad adoption, critical scrutiny, and responsible use.
We are hiring predocs and postdocs. For more information, please contact aib@gsb.columbia.edu.

Publications
November 2025
E-GEO: A Testbed for Generative Engine Optimization in E-Commerce
Puneet S Bagga, Vivek F Farias, Tamar Korkotashvili, Tianyi Peng, Yuhang Wu
September 2025
Leveraging LLMs to Improve Experimental Design: A Generative Stratification Approach
arXiv
September 2025
A Mega-Study of Digital Twins Reveals Strengths, Weaknesses, and Opportunities for Further Improvement
Tianyi Peng et al.
2025
Twin-2K-500: A Data Set for Building Digital Twins of over 2,000 People Based on Their Answers to over 500 Questions
Olivier Toubia et al. · Marketing Science
2025
Prompt Architecture Induces Methodological Artifacts in Large Language Models
Melanie Brucks & Olivier Toubia · PLOS ONE
2025
LLM Generated Persona Is a Promise with a Catch
Ang Li, Haozhe Chen, Hongseok Namkoong & Tianyi Peng · NeurIPS 2025 (Position Paper)
2025
AI Agents for Web Testing: A Case Study in the Wild
Naimeng Ye, Xiao Yu, Ruize Xu, Tianyi Peng, Zhou Yu · NeurIPS 2025 Workshop on Bridging Language, Agent, and World Models
December 2023
The Challenge of Using LLMs to Simulate Human Behavior: A Causal Inference Perspective
arXiv