2022
DOI: 10.21203/rs.3.rs-2229642/v1
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Box Jenkins Neural Network Keynecian Reinforcement Learning Based Financial Big Data Analysis for Optimal Prediction

Abstract: In the digital era, Economic and Finance information is huge and attainable. The financial industry is a key to stimulating the evolution of the national economy, and with the immense availability of huge data makes the process of financial data analysis both a time consuming and tedious decision making in the financial market. By analyzing this Economic and Finance information significantly and easily, valuable perceptions can be provided. For rational economic growth, information pertaining to economic affai… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
5
0

Year Published

2023
2023
2023
2023

Publication Types

Select...
2

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(5 citation statements)
references
References 20 publications
(47 reference statements)
0
5
0
Order By: Relevance
“…However, it introduces extra overheads and leads to suboptimal regret. More recently, Wei and Luo (2021) propose a general approach that is applicable to various RL settings (including both episodic MDPs and infinite-horizon MDPs) and achieves optimal dynamic regret without any prior knowledge of the degree of nonstationarity.…”
Section: Nonstationary Episodic Rlmentioning
confidence: 99%
“…However, it introduces extra overheads and leads to suboptimal regret. More recently, Wei and Luo (2021) propose a general approach that is applicable to various RL settings (including both episodic MDPs and infinite-horizon MDPs) and achieves optimal dynamic regret without any prior knowledge of the degree of nonstationarity.…”
Section: Nonstationary Episodic Rlmentioning
confidence: 99%
“…To achieve a nearly-optimal regret without the prior knowledge of the variation budget, (Auer, Gajane, and Ortner 2019) and (Chen et al 2019) maintain a distribution over bandit arms with properly controlled variance for all reward estimators. For RL problems, the seminar work (Wei and Luo 2021) proposes a black-box reduction approach that turns a certain RL algorithm with optimal regret in a (near-)stationary environment into another algorithm with optimal dynamic regret in a non-stationary environment. However, the above works only consider risk-neutral RL and may not apply to the more general risk-sensitive RL problems.…”
Section: Related Workmentioning
confidence: 99%
“…To overcome this limitation, we propose a meta-algorithm that adaptively detects the non-stationarity without the knowledge of B, while still achieving the similar dynamic regret as in Theorems 3.1 and 3.2. In particular, we generalize the black-box approach (Wei and Luo 2021) to the risk-sensitive RL setting and design a non-stationarity detection based on the exponential Bellman equations (3).…”
Section: Adaptive Algorithm Without the Knowledge Of Variation Budgetmentioning
confidence: 99%
See 2 more Smart Citations