Model-Free Approximate Bayesian Learning for Large-Scale Conversion Funnel Optimization

Author:

Iyengar Garud1,Singal Raghav2ORCID

Affiliation:

1. IEOR, Columbia University, New York, NY, USA

2. Tuck School of Business, Dartmouth College, Hanover, NH, USA

Abstract

The flexibility of choosing the ad action as a function of the consumer state is critical for modern-day marketing campaigns. We study the problem of identifying the optimal sequential personalized interventions that maximize the adoption probability for a new product. We model consumer behavior by a conversion funnel that captures the state of each consumer (e.g., interaction history with the firm) and allows the consumer behavior to vary as a function of both her state and firm’s sequential interventions. We show our model captures consumer behavior with very high accuracy (out-of-sample area under the curve of over 0.95) in a real-world email marketing dataset. However, it results in a very large-scale learning problem, where the firm must learn the state-specific effects of various interventions from consumer interactions. We propose a novel attribution-based decision-making algorithm for this problem that we call model-free approximate Bayesian learning. Our algorithm inherits the interpretability and scalability of Thompson sampling for bandits and maintains an approximate belief over the value of each state-specific intervention. The belief is updated as the algorithm interacts with the consumers. Despite being an approximation to the Bayes update, we prove the asymptotic optimality of our algorithm and analyze its convergence rate. We show that our algorithm significantly outperforms traditional approaches on extensive simulations calibrated to a real-world email marketing dataset.

Publisher

SAGE Publications

Reference41 articles.

1. Abe N, Pednault E, Wang H, et al. (2002) Empirical comparison of various reinforcement learning strategies for sequential targeted marketing. In: IEEE international conference on data mining (eds V Kumar, S Tsurnoto, N Zhong, et al.), Maebashi City, Japan, 9–12 December, pp.3–10. New York, NY: IEEE.

2. Agrawal S, Goyal N (2012) Analysis of Thompson sampling for the multi-armed bandit problem. In: Conference on learning theory, Vol. 23, (eds S Mannor, N Srebro and RC Williamson), Edinburgh, Scotland, 25–27 June, pp.39.1–39.26. PMLR.

3. Agrawal S, Goyal N (2013) Thompson sampling for contextual bandits with linear payoffs. In: International conference on machine learning (eds S Dasgupta and D McAllester), Atlanta, USA, 16–21 June, pp.127–135. PMLR.

4. Learning the Minimal Representation of a Dynamic System from Transition Data

5. Beyond the Last Touch: Attribution in Online Advertising

同舟云学术

1.学者识别学者识别

2.学术分析学术分析

3.人才评估人才评估

"同舟云学术"是以全球学者为主线,采集、加工和组织学术论文而形成的新型学术文献查询和分析系统,可以对全球学者进行文献检索和人才价值评估。用户可以通过关注某些学科领域的顶尖人物而持续追踪该领域的学科进展和研究前沿。经过近期的数据扩容,当前同舟云学术共收录了国内外主流学术期刊6万余种,收集的期刊论文及会议论文总量共计约1.5亿篇,并以每天添加12000余篇中外论文的速度递增。我们也可以为用户提供个性化、定制化的学者数据。欢迎来电咨询!咨询电话:010-8811{复制后删除}0370

www.globalauthorid.com

TOP

Copyright © 2019-2024 北京同舟云网络信息技术有限公司
京公网安备11010802033243号  京ICP备18003416号-3