2020
DOI: 10.3905/jfds.2020.1.052
|View full text |Cite
|
Sign up to set email alerts
|

Deep Hedging of Derivatives Using Reinforcement Learning

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

2
40
0

Year Published

2023
2023
2025
2025

Publication Types

Select...
8

Relationship

1
7

Authors

Journals

citations
Cited by 48 publications
(42 citation statements)
references
References 24 publications
2
40
0
Order By: Relevance
“…To the best of our knowledge, this paper is the first work to implement the RL algorithms with online learning to hedge contingent claims, particularly variable annuities. Contrary to Xu (2020) and Carbonneau (2021), in which both adapted the state-of-the-art DH approach in Bühler et al (2019), this paper is in line with the recent works by Kolm & Ritter (2019) and Cao et al (2021), while extends with actuarial components. We shall outline the differences between the RL and DH approaches throughout sections 3 and 4, as well as Appendices A and B.…”
Section: Introductionsupporting
confidence: 68%
“…To the best of our knowledge, this paper is the first work to implement the RL algorithms with online learning to hedge contingent claims, particularly variable annuities. Contrary to Xu (2020) and Carbonneau (2021), in which both adapted the state-of-the-art DH approach in Bühler et al (2019), this paper is in line with the recent works by Kolm & Ritter (2019) and Cao et al (2021), while extends with actuarial components. We shall outline the differences between the RL and DH approaches throughout sections 3 and 4, as well as Appendices A and B.…”
Section: Introductionsupporting
confidence: 68%
“…Tamar et al (2016) showed how both the first and second moments (and possibly higher moments) of the distribution of G t can be updated using more than one Q-function. Cao et al (2021) used this approach and produced results for an objective function involving the mean and standard deviation of G t , where one Q-function approximates the mean of the terminal return and a second Q-function approximates the expected value of the square of the terminal return. However, the approach is less than ideal.…”
Section: The Rl Modelmentioning
confidence: 99%
“…Previous studies concerned with the application of RL to hedging decisions include Halperin (2017), Buehler et al (2019), Kolm and Ritter (2019), and Cao et al (2021). These authors consider how RL can be used to hedge a single call or put option using a position in the underlying asset.…”
Section: Introductionmentioning
confidence: 99%
“…As a remark of a concept in deep hedging, continuing from the proposition that the approach is model-free, we can include market frictions [3]. Considering the applications of deep reinforcement machine learning methods [6] as typical example models with the presence of market frictions, only generator, loss function and trading instruments are additionally needed [9] that depend on no market dynamics. And Shi et al [3] have developed an algorithm that yields optimum even in high dimensions which primarily depends on the number of hedge instruments [5].…”
Section: Related Workmentioning
confidence: 99%
“…As mentioned by Shi et al [3], the FBSDE solver's algorithm has a problem with the time horizon and discrete-time when scaling. Their approach is targeting the utility function, training and learning the optimal trading strategy directly with reinforcement learning algorithms [6].…”
Section: Introductionmentioning
confidence: 99%