2007
DOI: 10.1109/tsmca.2007.904825
|View full text |Cite
|
Sign up to set email alerts
|

A Multiagent Approach to $Q$-Learning for Daily Stock Trading

Abstract: Abstract-The portfolio management for trading in the stock market poses a challenging stochastic control problem of significant commercial interests to finance industry. To date, many researchers have proposed various methods to build an intelligent portfolio management system that can recommend financial decisions for daily stock trading. Many promising results have been reported from the supervised learning community on the possibility of building a profitable trading system. More recently, several studies h… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

3
35
0
4

Year Published

2009
2009
2024
2024

Publication Types

Select...
4
4
1

Relationship

0
9

Authors

Journals

citations
Cited by 96 publications
(42 citation statements)
references
References 13 publications
3
35
0
4
Order By: Relevance
“…Data not present in Table 2 showed that the average return per trade is 0.06537% for the long trades and 0.2691% for short trades, which means that the OneR classifier was extremely profitable in downwards periods for this particular experiment. For the other classifiers, while prediction accuracies may be similar to those observed in the literature (Chen & Shih, 2006;Eng et al, 2008;Kim, 2003;Lee et al, 2007; S. Li & Kuo, 2008;Tenti, 1996), the average return per trade does not allow the models to have a positive return at the end of the trading period. These initial results suggest that it is possible to generate positive return with modest middle range accuracy.…”
Section: Single Training Experimentssupporting
confidence: 52%
See 1 more Smart Citation
“…Data not present in Table 2 showed that the average return per trade is 0.06537% for the long trades and 0.2691% for short trades, which means that the OneR classifier was extremely profitable in downwards periods for this particular experiment. For the other classifiers, while prediction accuracies may be similar to those observed in the literature (Chen & Shih, 2006;Eng et al, 2008;Kim, 2003;Lee et al, 2007; S. Li & Kuo, 2008;Tenti, 1996), the average return per trade does not allow the models to have a positive return at the end of the trading period. These initial results suggest that it is possible to generate positive return with modest middle range accuracy.…”
Section: Single Training Experimentssupporting
confidence: 52%
“…Maggini et al (Maggini, Giles, & Horne, 1997) had pointed out that there is an inherent difficulty in generating statistically reliable technical indicators, due to the fact that the rules inferred to produce accurate predictions are changing continually in financial time series, and that it is even possible to evidence the presence of a high number of contradictory instances in the training sets due to the fact that market data exhibit statistical characteristics found in other types of time series. This situation is reflected in the large volume of papers (Chen & Shih, 2006;Eng, Li, Wang, & Lee, 2008;Kim, 2003;Lee, Park, O, Lee, & Hong, 2007;S. Li & Kuo, 2008;Tenti, 1996) that have reported accuracies under 60% with ML models which have shown impressive performance in areas other than financial prediction.…”
Section: Machine Learning In Financial Forecastingmentioning
confidence: 99%
“…It was conducted using several selected shares on the Italian stock market over an investment period of 30 stock market years with the algorithm making buy and sell decisions on the six selected shares over the period [13]. Deep Q-learning have also been applied to the foreign exchange market against the baseline buy and hold strategy and an expert trader [14] as well as to a stock market index [15]. Sornmayura in 2019 applied this methodology and compared its performance against the expert trade and baseline buy and hold strategy using the currency pairs EUR/USD and USD/JPY within 15 years of foreign exchange market data [14].…”
Section: On and Off Policy Reinforcement Learningmentioning
confidence: 99%
“…The comparison for the performance of Q‐Learning and RRL to optimize the portfolios is provided in (Du, Zhai, and Lv ()), and the authors concluded that RRL delivers superior results. The authors (Lee, Park, Jangmin, Lee, and Hong ()) proposed a model that incorporates multiple Q‐Learning agents to provide adequate decision support for the daily stock trading problem. Moody and Saffell () introduced a trading system based on direct RL to optimize stock trading.…”
Section: Introductionmentioning
confidence: 99%