2018
DOI: 10.1177/1748301818779059
|View full text |Cite
|
Sign up to set email alerts
|

An actor-critic-based portfolio investment method inspired by benefit-risk optimization

Abstract: How to get maximal benefit within a range of risk in securities market is a very interesting and widely concerned issue. Meanwhile, as there are many complex factors that affect securities' activity, such as the risk and uncertainty of the benefit, it is very difficult to establish an appropriate model for investment. Aiming at solving the curse of dimension and model disaster caused by the problem, we use the approximate dynamic programming to set up a Markov decision model for the multi-time segment portfoli… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
4
0

Year Published

2018
2018
2023
2023

Publication Types

Select...
4
4

Relationship

0
8

Authors

Journals

citations
Cited by 9 publications
(4 citation statements)
references
References 22 publications
0
4
0
Order By: Relevance
“…refined log-optimal strategy and combined it with reinforcement learning [12]. Lili Tang proposed a model-based actor-critic algorithm under uncertain environment is proposed, where the optimal value function is obtained by iteration on the basis of the constrained risk range and a limited number of funds [13]. David W. Lu implemented in Long Short Term Memory (LSTM) recurrent structures with Reinforcement Learning or Evolution Strategies acting as agents The robustness and feasibility of the system is verified on GBPUSD trading [14].…”
Section: Discussionmentioning
confidence: 99%
“…refined log-optimal strategy and combined it with reinforcement learning [12]. Lili Tang proposed a model-based actor-critic algorithm under uncertain environment is proposed, where the optimal value function is obtained by iteration on the basis of the constrained risk range and a limited number of funds [13]. David W. Lu implemented in Long Short Term Memory (LSTM) recurrent structures with Reinforcement Learning or Evolution Strategies acting as agents The robustness and feasibility of the system is verified on GBPUSD trading [14].…”
Section: Discussionmentioning
confidence: 99%
“…However, the lack of cathepsin C cannot completely eliminate the production of IL‐1β (Campden & Zhang, 2019). Cathepsin L may compensate for cathepsin B in some cases and play an indirect role in the activation of the NLRP3 inflammasome (Tang, 2018). Cathepsin S can compensate for the loss of other cathepsins in the activation of NLRP3 inflammasome.…”
Section: Nlrp3 Inflammasome Activation Pathwaysmentioning
confidence: 99%
“…That is, RL can directly output trade positions, bypassing the explicit forecasting step by the trial-and-error interaction with the financial environment. Many existing RL methods get promising results by focusing on various technologies to extract richer representation, e.g., by modelbased learning (Tang 2018;Yu et al 2019), by adversarial learning (Liang et al 2018), or by state augmentation (Ye et al 2020). However, these RL algorithms assume that portfolio weights can change immediately at the last price once an order is placed.…”
Section: Introductionmentioning
confidence: 99%