2020
DOI: 10.48550/arxiv.2012.03800
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Revenue Maximization and Learning in Products Ranking

Abstract: We consider the revenue maximization problem for an online retailer who plans to display a set of products di ering in their prices and qualities and rank them in order. The consumers have random attention spans and view the products sequentially before purchasing a "satis cing" product or leaving the platform emptyhanded when the attention span gets exhausted. Our framework extends the cascade model in two directions: the consumers have random attention spans instead of xed ones and the rm maximizes revenues … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2021
2021
2021
2021

Publication Types

Select...
2

Relationship

1
1

Authors

Journals

citations
Cited by 2 publications
(2 citation statements)
references
References 49 publications
(53 reference statements)
0
2
0
Order By: Relevance
“…Furthermore, OIM falls in the Combinatorial Multi-Armed Bandit (CMAB) regime, where each arm is a combination of multiple selected items. CMAB has found widespread applications in the real-world, such as assortment optimization (Agrawal et al 2019, Oh and Iyengar 2021, and product ranking (Chen et al 2020). Most CMAB models assume semi-bandit feedback, where the agent can observe the outcome of every item within the selected arm (Gai et al 2012, Chen et al 2013, Kveton et al 2015c, Chen et al 2016b, Wang and Chen 2017, or partial-feedback, where the agent can only observe a subset of the selected items (Kveton et al 2015b, Combes et al 2015, Katariya et al 2016, Zong et al 2016, Cheung et al 2019.…”
Section: Literature Reviewsmentioning
confidence: 99%
“…Furthermore, OIM falls in the Combinatorial Multi-Armed Bandit (CMAB) regime, where each arm is a combination of multiple selected items. CMAB has found widespread applications in the real-world, such as assortment optimization (Agrawal et al 2019, Oh and Iyengar 2021, and product ranking (Chen et al 2020). Most CMAB models assume semi-bandit feedback, where the agent can observe the outcome of every item within the selected arm (Gai et al 2012, Chen et al 2013, Kveton et al 2015c, Chen et al 2016b, Wang and Chen 2017, or partial-feedback, where the agent can only observe a subset of the selected items (Kveton et al 2015b, Combes et al 2015, Katariya et al 2016, Zong et al 2016, Cheung et al 2019.…”
Section: Literature Reviewsmentioning
confidence: 99%
“…The work in (Kveton, Szepesvari, Wen, & Ashkan, 2015;Cao, Sun, & Shen, 2019;Chen, Li, & Yang, 2020) studies variants of sequential choice bandit model without feedback. Unlike our setting, the sequence of action is pre-determined at the arrival of each user, independently from the user's feedback.…”
Section: Related Workmentioning
confidence: 99%