2018
DOI: 10.48550/arxiv.1805.09360
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Deep Reinforcement Learning of Marked Temporal Point Processes

Utkarsh Upadhyay,
Abir De,
Manuel Gomez-Rodriguez

Abstract: In a wide variety of applications, humans interact with a complex environment by means of asynchronous stochastic discrete events in continuous time. Can we design online interventions that will help humans achieve certain goals in such asynchronous setting? In this paper, we address the above problem from the perspective of deep reinforcement learning of marked temporal point processes, where both the actions taken by an agent and the feedback it receives from the environment are asynchronous stochastic discr… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2

Citation Types

0
4
0

Year Published

2023
2023
2023
2023

Publication Types

Select...
2
1
1

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(4 citation statements)
references
References 19 publications
0
4
0
Order By: Relevance
“…Instead of directly summing over all past events, as in models based on the Hawkes process (Hawkes, 1971), a fixed length vector representation of the past is learnt and updated at each new time step. Forecasts can then be made by modelling the conditional intensity function on this vector (Xiao et al, 2017;Li et al, 2018;Upadhyay et al, 2018;Huang et al, 2019;Omi et al, 2019), or instead through directly modelling the probability of the next event (Shchur et al, 2019). We direct the reader to Shchur et al (2021) for a review of neural point process models.…”
Section: Introductionmentioning
confidence: 99%
“…Instead of directly summing over all past events, as in models based on the Hawkes process (Hawkes, 1971), a fixed length vector representation of the past is learnt and updated at each new time step. Forecasts can then be made by modelling the conditional intensity function on this vector (Xiao et al, 2017;Li et al, 2018;Upadhyay et al, 2018;Huang et al, 2019;Omi et al, 2019), or instead through directly modelling the probability of the next event (Shchur et al, 2019). We direct the reader to Shchur et al (2021) for a review of neural point process models.…”
Section: Introductionmentioning
confidence: 99%
“…A classic approach to model event sequences as TPPs is through the Hawkes process, in which a simple parametric form is used to capture temporal dependence among events (Hawkes 1971). In the past few years, many researchers have developed neural TPP models that have achieved fruitful results on standard benchmarks for predictive tasks, because neural networks are ca-pable of capturing more complex dependencies (Du et al 2016;Mei and Eisner 2016;Xiao et al 2017;Upadhyay, De, and Gomez-Rodriguez 2018;Omi, Ueda, and Aihara 2019;Shchur, Biloš, and Günnemann 2019;Zuo et al 2020;Zhang et al 2020a;Shchur et al 2020;Boyd et al 2020;Gao et al 2020;Gu 2021).…”
Section: Introductionmentioning
confidence: 99%
“…Instead of directly summing over all past events, as in models based on the Hawkes process (Hawkes, 1971), a fixed length vector representation of the past is learned and updated at each new time step. Forecasts can then be made by modeling the conditional intensity function on this vector (Huang et al, 2019;Li et al, 2018;Omi et al, 2019;Upadhyay et al, 2018;Xiao et al, 2017), or instead through directly modeling the probability of the next event (Shchur et al, 2019). We direct the reader to Shchur et al (2021) for a review of neural point process models.…”
mentioning
confidence: 99%