Expert Systems With Applications
Expert Systems With Applications
a r t i c l e i n f o a b s t r a c t
Article history: In this paper, two hybrid models are used for timing of the stock markets on the basis of the technical
Received 22 May 2017 analysis of Japanese Candlestick by Support Vector Machine (SVM) and Heuristic Algorithms of Imperi-
Revised 16 August 2017
alist Competition and Genetic. In the first model, SVM and Imperialist Competition Algorithm (ICA) are
Accepted 9 October 2017
developed for stock market timing in which ICA is used to optimize the SVM parameters. In the second
Available online 11 October 2017
model, SVM is used with Genetic Algorithm (GA) where GA is used for feature selection in addition to
Keywords: SVM parameters optimization. Here the two approaches, Raw-based and Signal-based are devised on the
Finance basis of the literature to generate the input data of the model. For a comparison, the Hit Rate is con-
Stock market forecasting sidered as the percentage of correct predictions for periods of 1–6 day. The results show that SVM-ICA
Candlestick technical analysis performance is better than SVM-GA and most importantly the feed-forward static neural network of the
Support Vector Machine literature as the standard one.
Imperialist Competition Algorithm
© 2017 Elsevier Ltd. All rights reserved.
Genetic Algorithm
https://doi.org/10.1016/j.eswa.2017.10.023
0957-4174/© 2017 Elsevier Ltd. All rights reserved.
22 E. Ahmadi et al. / Expert Systems With Applications 94 (2018) 21–31
Table 1
Raw approach.
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15
C2 C3 C4 C5 C6 C7 O5 H5 L5 O6 H6 L6 O7 H7 L7
C1 C1 C1 C1 C1 C1 C1 C1 C1 C1 C1 C1 C1 C1 C1
Table 2
Signal approach.
1 2 3 4 5 6 7 8
C2 C3 C4 C5 C6 C7 O5 O6
C1 C1 C1 C1 C1 C1 C5 C6
9 10 11 12 13 14 15 16
O7 H7 Min (O7 ,C7 ) Max(O7 ,C7 ) Min (O7 ,C7 ) O7 L6 C7
C7 Max(O7 ,C7 ) L7 Max(O6 ,C6 ) Min (O6 ,C6 ) H6 O7 O6
17 18 19 20 21 22 23 24
Max(O6 ,C6 ) Max(O7 ,C7 ) Min (O7 ,C7 ) Min (O6 ,C6 ) H7 H7 L6 L7
Min (O5 ,C5 ) Max(O5 ,C5 ) Min (O5 ,C5 ) Max(O5 ,C5 ) H6 H5 L5 L7
Neural Networks has few disadvantages in the learning process In this part, the new insights that are brought with the pro-
based on the strong likelihood that results in the lack of repro- posed methods to overcome the limitations of previous studies, are
ducibility of the process. For this reason, many researchers prefer discussed.
choosing new approaches based on robust statistical principles like
SVM (Fernandez-Lozano et al., 2013). Recently, the SVM method, 3.1. SVM
one of supervised learning methods, has gain popularity and has
been known as one of the most advanced applications of regres- SVM is actually a binary classifier in which two classes are sep-
sion and classification methods. SVM formulation minimizes the arated using a linear boundary. In this method by using an op-
structural risk and more importantly it has high performance in timization algorithm, samples that make up the boundary classes
the practical applications (Huang, Nakamori, & Wang, 2005). can be achieved. They are called Support Vectors. In Fig. 1, two
Given the above advantages, since SVM was introduced based classes and their associated support vectors are shown. Input fea-
on Vapnik’s statistical learning theory, many studies focus on the ture space that is a vector, consists of two classes and classes
theory and its applications. Several studies use the SVM to pre- hold xi educational points while i = 1, …N. These two classes are
dict the time series (Tay & Cao, 2001; Huang et al., 2005) The tagged with yi = ±1. To calculate the decision boundary of two ab-
SVM developed by Vapnik (1995, 1998), is a machine learning solutely separate classes, the method of Optimal Margin is used
technique and due to its interesting features and excellent per- (Fernandez-Lozano et al., 2013; Huang, Yang, & Chuang, 2008; Tay
formance in various problems, has been known widely on issues & Cao, 2001). In general, a boundary-line decisions can be written
of non-linear prediction and it’s been very successful (Wang, Xu, as follow:
& Weizhen Lu, 2003). Although, recently it is not one of the first w.x + b = 0 (1)
choices, a significant number of researchers report a variety of per-
formance applications in pattern recognition, regression and time Where x is a point on the decision boundary and w is an n-
series forecasting. For example, Tay and Cao (2001) try to use this dimensional vector that is perpendicular to the decision boundary.
type of neural network to predict time series; they have concluded It is to be noted that, wb is the distance between the origin and
that the SVM is better than multi-layer propagation neural net- the decision boundary, and w.x is the inner product of the two vec-
work in forecasting of financial time series. Wang et al. (2003) use tors.
SVM to predict air quality and they came to the conclusion that In the situation that classes overlap separating the classes by
the functions of neural networks which are based on the ra- boundary linear decision-making will always have errors to solve
dius are very effective. The experimental results and review of this problem, you can start using initial data from the Rn dimen-
the literature have shown that the kernel parameters, C and σ sion using a nonlinear transformation ∅, moved to the Rm dimen-
have a significant effect on the accuracy and precision of SVM sion, in the dimensions that classes has less interference with each
(Cherkassky & Ma, 2004). Heuristic methods have not been used other. In this case, finding the optimal decision boundary to solve
successfully for determining the value of these parameters. For the optimization problem becomes the following:
this reason, researchers used meta-heuristic methods to obtain the
−1
N N N
right number of variables. Pai and Hong (20 05, 20 06) used GA Max : α1 , . . . , αN αi α j yi y j ∅(xi ).∅ x j + αi
and Gradual Annealing Algorithm, respectively. In another case, 2
i=1 j=1 i=1
Hong, Dong, Zheng, and Lai (2011) and Hong, Dong, Chen, and
N
Wei (2011) used continuous Ant Colony Algorithm and GA respec- ×0 ≤ αi ≤ C i = 1, . . . , N αi yi = 0 (2)
tively, to achieve SVR parameters. Note that for numerical opti-
i=1
24 E. Ahmadi et al. / Expert Systems With Applications 94 (2018) 21–31
Table 3
Details of training and testing of the applied datasets.
No. Training period Test period No. Training period Test period No. Training period Test period No. Training period Test period
In this optimization problem α i .α is Lagrange multipliers and c Polynomial Kernel with parameter k
is a constant. In the formula (2) Instead of using ∅ it’s preferred to k
use a core function which is defined as follows: K (a, b) = (a.b + 1 ) (5)
3.2. SVM and ICA In this model as is shown in Fig. 3, GA for feature selection and
selection of optimal parameters of SVM is used. The purpose of
In this section, we aim to develop a model for timing the stock the feature set is a subset of useful and unique features. By ex-
market with SVM and ICA based on the business strategy techni- tracting the most essential attributes that have the least number of
cal analysis with the help of Japanese Candlesticks. As is shown by features, we can greatly reduce the computational cost. Variables
Fig. 2, the turning point of this model is using the Imperialist com- related to the SVM are not only features, but they are also the
petitive algorithm to optimize the SVM parameters. The main con- Kernel function parameters. In this model, we use the SVM with
tribution of this model is the following: 1) the combination of the a combination of feature selection and parameter optimization by
Imperialist competitive algorithm and multi-class SVM algorithm GA. As typically used in data preprocessing, feature selection, se-
for determination of the amount of SVM parameters and present- lects a subset of attributes that make up the model and explains
ing a new algorithm for stock market timing. 2) Creating a model the data, by eliminating the redundant and irrelevant or confus-
based on the SVM concepts and Japanese candlesticks charts in a ing features, feature selection or the classifier can predict predic-
quite different way from previous studies. tion accuracy and the ability to understand and improve. How to
determine the methods of feature selection and feature selection
3.2.1. Imperialist competitive algorithm and optimization of Kernel indicator to build models that can better read interpret data. Fea-
parameters ture selection on one hand can decrease the cost of computing by
Select a Kernel; selecting the appropriate Kernel for the SVM reducing the dataset dimension; and on the other hand it can im-
has a major impact. Selecting the kernel that determines our data prove the model performance by eliminating redundant and irrel-
from the Initial space to the space that we want to do mapping to evant features. It should be noted that, feature selection is differ-
and the new space where we can make a linear classifier is very ent from methods like feature conversion or feature extraction in
impressive. Generally, the Gaussian Kernel or polynomial Kernel, which new features are developed out of the main ones.
have been used. If you have a basic understanding of the spatial
data, it can help in selecting the appropriate Kernel. Another way 3.3.1. GA
is to select a different Kernel and classify the data, and then se- GA optimal response which is based on the principle of Dar-
lect the appropriate Kernel based on the test data. In addition to win’s “survival of the qualified”, is obtained after a series of itera-
selecting the appropriate Kernel, it is also very important to get tive calculations. In GA, successive populations of diverse solutions
the Kernel parameters. Among Kernels that are used, the polyno- are offered by a chromosome until acceptable results are achieved.
mial, Gaussian and sigmoid Kernel are used more often (Cristianini The answer of the quality fitness function is examined in the eval-
& Shawe-Taylor, 20 0 0). uation step. The functions of Mutation and crossover are the main
E. Ahmadi et al. / Expert Systems With Applications 94 (2018) 21–31 25
4. Methodology
The input dataset used in this study, is based on the two ap-
proaches introduced for the first time by Jasemi et al. (2011b). In
these two approaches, the daily stock prices including Low, High,
Open and Close prices turn into 15 and 24 indicators based on
what is shown in Tables 1 and 2, respectively for the first and sec-
ond approaches. It is to be noted that in the tables Oi , Hi , Li , and
Ci respectively denotes open, high, low and close prices on the ith
day while 7th day is today (last day), 6th day is yesterday and so
on. The output is stock performance that is given in the form of
buy, sell or no-action signal.
Table 3 describes the 48 datasets that are used on a daily basis
for model training and testing. Each dataset is divided into two
groups of training and testing sets and each set contains daily
stock prices. For example in dataset 1, data of year 20 0 0 is used
for training and data of year 2001 is used for testing. In dataset
2, time-distance between training and test data are increased and
data of year 2002 are used for testing. In other datasets the num-
ber of training data are also increased; for example in dataset 9,
data of year 20 0 0 and 2001 are used together as a single training
data.
4.2.1. SVM-ICA
The following parameters or notations will be used in the rest
of the paper:
operators that randomly affect the fitness value. Chromosomes se-
lection for reproduction is done by evaluating the amount of the NPop : the number of the primary population
merit. The more suitable chromosomes are more likely to be cho- Nimp : the number of the imperialists
sen for reproduction by roulette wheel method or the other ones. NCol : the number of the colonies
Crossover is the main operator that seeks the new answer in the Zeta: the impact coefficient of the colonies’ cost on the empires’
search space while a random mechanism for genes exchange be- cost
tween two chromosomes takes place by one-point, two-point or Prevolution: the probability of the revolution
multiple-point crossover. The evolutionary process is repeated till Imp Cost: Imperialist’s cost
the stop condition is satisfied. Imp Fitness: Imperialist’s fitness
26 E. Ahmadi et al. / Expert Systems With Applications 94 (2018) 21–31
Fig. 8. ICA performance during the 100 courses in the first dataset.
Fig. 11. The total number of signals obtained from the model in both approaches.
Table 5 is taken place in the search space. During the absorption and rev-
Hit rate for 1 and 6 day periods by SVM-ICA model in both approaches.
olution, it is possible that a colony reaches a better position and
Raw approach Signal approach has the chance of controlling the whole empire and substitutes
No. 1d (%) 6d (%) Sig. no. No. 1d (%) 6d (%) Sig. no. the current imperialist of the empire. In addition, if any of the
colonies has more power than the respective imperialist, their po-
1 0.20 0.91 86 1 0.23 0.87 30
2 0.25 0.94 72 2 0.09 0.77 35
sition is changed. In this case the MisClass criterion is used as the
3 0.12 0.54 84 3 0.31 0.71 35 cost function in the ICA. After training the SVM, The SVM’s out-
4 0.07 0.75 59 4 0.16 0.65 51 put for the training data are calculated with the formula (8). After
… … … … … … … … the training phase, test dataset is used for checking the SVM, and
45 0.33 0.33 3 45 0.26 0.79 114
the Misclass test is also calculated. The cost of each empire is de-
46 0.17 0.63 52 46 0.37 0.93 30
47 0.16 0.68 74 47 0.32 0.86 107 pended on the cost of its imperialist and colonies which is calcu-
48 0.09 0.73 11 48 0.00 1.00 2 lated by the formula (12).
Fig. 12. Prediction accuracy in different dataset by changing the number of features with the approach 1.
Table 6
The complete list of the results.
No. 1 2 3 4 5 6 7 8 No. 1 2 3 4 5 6 7 8
1 17 13 17 11 8 12 78 86 1 7 7 7 2 2 1 26 30
2 18 8 15 12 9 6 68 72 2 3 7 6 5 4 2 27 35
3 10 7 8 8 5 7 45 84 3 11 5 3 2 1 3 25 35
4 4 7 6 13 7 7 44 59 4 8 7 6 5 4 3 33 51
… … … … … … … … … … … … … … … … … …
45 1 0 0 0 0 0 1 3 45 30 17 18 13 9 3 90 114
46 9 8 5 6 2 3 33 52 46 11 5 3 4 3 2 28 30
47 12 10 9 11 3 5 50 74 47 34 19 12 19 5 3 92 107
48 1 3 1 1 1 1 8 11 48 0 0 1 1 0 0 2 2
Table 7
Optimal features and parameters with predicted accuracy in each data sets.
This algorithm is repeated until we reach to the stopping con- 7. Calculating the hit rate
dition and at last the best answer is selected among then and the
hit rate is calculated. 4.2.2. SVM-GA
Semi code for the model is like the following: While GA can search locally for various combinations of feature
sets so that the optimization criterion is enhanced, it is expected
1. Uploading data that this class of algorithms can perform better in feature selection
2. Determining training and testing set problem. In addition different studies explain using GA in feature
3. ICA setting selection. Therefor in this study the GA is used for searching an
4. Determining kernel function and its parameters optimized subset of features for the stock market prediction model.
5. Initializing ICA Fig. 6 depicts the evolution cycle in GA.
6. ICA process Apart from selecting feature to choose the optimal RBF pa-
- Creating the primary population randomly rameters we use this algorithm, because it can be used for high-
- Defining the MisClass function as the cost function dimensional data analysis. In this case, parameters C and σ , must
- Absorbing colonies be created for SVM model.
- Revolution of colonies The model presented by SVM and GA, includes four main steps
- Exchange with the best colony that are described with more details. In this model, after identify-
- Calculating the total cost ing the number of features (in the initial procedure the number of
- Imperialism competitive features starts from 3) and loading data and other settings of the
- Extracting empires and developed colonies model, such as initial population and stop condition parameters
- Specifying the best empire, best empire position, best em- (number of generations) GA starts working. GA to calculate chro-
pire cost mosomal uses the probability of the likelihood of a mistake by the
- Determining the outputs SVM which is the same as Misclass function of SVM-ICA and uses
E. Ahmadi et al. / Expert Systems With Applications 94 (2018) 21–31 29
Fig. 13. Prediction accuracy in different data set by changing the number of features with the approach 2.
160
Total Number Of Signals
140
120
100
80
60
40
20
0 5 10 15 20 25 30 35 40 45 50
DataSet
Fig. 14. Compare the approaches 1 and 2 in terms of the total number of signals obtained in the SVM-GA model.
Table 8 features and kernel parameters). Each chromosome contains all the
Hit rate for 1 and 6 day periods with total number of buy and sell signals by
information needed to select the optimal features and parameters.
SVM-GA.
The length of each chromosome is 2 + n where n is the number of
Raw approach Signal approach features and 2 is the number of parameters of the RBF. In fact, this
No. 1d (%) 6d (%) Sig. no. No. 1d (%) 6d (%) Sig. no. model has a chromosome consists of 3 parts, the first part consists
1 0.36 0.61 102 1 0.25 0.41 150
of features, the second part is the parameter C and the third one
2 0.18 0.37 113 2 0.17 0.36 148 is parameter σ . Chromosome structure is shown in Fig. 7.
3 0.57 0.96 72 3 0.36 0.56 131 Step 2: Training
4 0.65 0.89 81 4 0.70 0.83 100 After creating the initial population, system uses SVM with help
… … … … … … … …
process parameter values in the chromosome and calculates the
45 0.24 0.67 73 45 0.15 0.44 120
46 0.22 0.55 112 46 0.08 0.23 162 performance of each chromosome. The performance of each chro-
47 0.53 0.83 72 47 0.10 0.28 164 mosome with the possibility of wrong classification is calculated
48 0.20 0.63 43 48 0.06 0.18 153 by Eq. (8).
In this study, the main goal is to find optimal or near-optimal
parameters so that predictions be more accurate. Then we set fit-
ness function for testing the accuracy of prediction.
it as the penalty functions. After running the algorithm and get to Step 3: Genetic operation
the final stop condition the chromosome that performed better is Here, a new generation of the population by using genetic op-
selected as the best answer. The algorithm for a higher number of erators such as selection, crossover and mutation is created. Based
properties up to 15 features in the first approach and 24 features in on the fitness values of each chromosome, the best chromosomes
the second approach is repeated and an optimal solution for each are selected and applied for the crossover as well as the mutation
is calculated. Finally, among the sub-optimal subset that works the operator with a very small mutation rate. After the new genera-
best model is chosen as the final answer, and the results antici- tion creation, the step 2 is run again along with calculation of fit-
pated by these features and optimal parameters are examined and ness values. This process is repeated until the stop condition (the
blow rate is calculated. pre-determined number of generations) is achieved. Now the chro-
Step 1: Start mosome that shows the best performance in the last selected pop-
In the first step, the system generates the initial population (the ulation is considered as the final answer.
number of features 3), uses it to find the optimal parameters (i.e.
30 E. Ahmadi et al. / Expert Systems With Applications 94 (2018) 21–31
Table 9
Results with the complete details by SVM-GA.
No. 1 2 3 4 5 6 7 8 No. 1 2 3 4 5 6 7 8
1 37 15 6 2 2 0 62 102 1 52 21 9 4 4 2 92 150
2 21 8 6 3 3 1 42 113 2 40 17 11 6 5 4 83 148
3 41 15 6 2 2 2 69 72 3 62 22 9 3 3 3 102 131
4 53 10 3 3 1 1 72 81 4 85 16 5 4 2 2 114 100
… … … … … … … … … … … … … … … … … …
45 18 11 8 5 4 3 49 73 45 33 18 11 7 7 6 82 120
46 25 12 6 2 1 0 62 112 46 27 14 9 7 6 5 68 162
47 38 13 5 2 1 1 60 72 47 31 15 10 8 6 6 76 164
48 9 6 4 4 3 2 27 43 48 24 12 7 6 5 4 58 153
Table 10
Comparing the models hit ratios in percentage.
Total Raw approach Signal approach 1-day raw approach 1-day signal approach
Step 4: Reaching the optimal solution As a sample, ICA performance in the first data set is shown in
Then algorithms for higher features are performed and chromo- Fig. 8 and the overall results-obtained from the model in the series,
some that has the best performance is achieved for different num- is shown in Fig. 9. In this matrix, prediction accuracy and MisClass
ber of features. Then from all the answers among all optimal chro- of the model is shown. The diagonal of the matrix output shows
mosomes, the chromosome with the best performance is selected the number of correct signals and other elements of the matrix,
and the hit rate is calculated. show the number of target signals that have been predicted by
mistake. There are 3 classes of ascending, neutral and descending
4.3. Calculate the total number of signals and hit rate signals in this matrix which have been predicted by the model. To-
tal of rows 1, 2 and 3 elements, shows the number of ascending,
Performance measures can be categorized into two groups of neutral and descending signals respectively. It is to be noted that
statistical and non-statistical ones. Non-statistical measures cover the matrix is achieved for every data set.
the economic aspects. In the area of this paper, the statistical Fig. 10 shows the prediction accuracy by SVM with the two ap-
ones are more common while the most popular one is Hit Rate proaches. It is clearly can be seen that the accuracy of SVM by sig-
(Atsalakis & Valavanis, 2009). Hit rate is defined as (number of nal approach is higher than raw approach for most of the datasets.
success) / (total signals). If the hit rate is higher than 51%, it is Also, the mean accuracy of 48 datasets in the first and the second
considered as a useful model (Lee, 2009). approaches are 70% and 79% respectively. For better depiction of
At this stage, with model outputs, sell and buy signals and total the results, Fig. 11 shows the total number of signals in both ap-
number of signals are figured out and the number of correct sig- proaches.
nals during a 6-day period are calculated. Since the base or stan- Table 5 shows the hit rate for periods of 1 and 6 day as well
dard study is Jasemi et al. (2011a), every details are set according as the total number of buying and selling signals. Table 6 presents
to that study and reading that paper is recommended for better the complete list of results while columns 1– 6 represent the per-
understanding. centages of correct signals on one, two, three, four, five and six-day
periods, respectively. Column 7 shows the total number of correct
signals and column 8 relates to the total number of signals emitted
5. Results and discussion by the model.
As is presented by Tables 8 and 9, the 6-day hit rate for the Dahal, K., Almejalli, K., Hossain, M. A., & Chen, W. (2015). GA-based learning for rule
approaches 1 and 2 is 59% and 61% respectively while the 1-day identification in fuzzy neural networks. Applied Soft Computing, 35, 605–617.
de Campos, L. M. L., de Oliveira, R. C. L., & Roisenberg, M. (2016). Optimization of
hit rate is almost the same at 33%. neural networks through grammatical evolution and a genetic algorithm. Expert
Systems with Applications, 56, 368–384.
6. Conclusion Fernandez-Lozano, C., Canto, C., Gestal, M., Andrade-Garda, J. M., Rabuñal, J. R., Do-
rado, J., et al. (2013). Hybrid model based on genetic algorithms and SVM ap-
plied to variable selection within fruit juice classification. The Scientific World
This study proposed two SVM-ICA and SVM-GA models for Journal. Article ID 982438, 13 pages.
stock market timing in the way that SVM is a classifier and ICA Hong, W. C., Dong, Y., Zheng, F., & Lai, C. Y. (2011a). Forecasting urban traffic flow
by SVR with continuous ACO. Applied Mathematical Modelling, 35(3), 1282–1291.
and GA are used for optimization of the SVM parameters. GA here
Hong, W. C., Dong, Y., Chen, L. Y., & Wei, S. Y. (2011b). SVR with hybrid chaotic ge-
also selects the optimum features for better prediction. To make netic algorithms for tourism demand forecasting. Applied Soft Computing, 11(2),
the comparison fair, all the details are set according to the base 1881–1890.
Huang, C. J., Yang, D. X., & Chuang, Y. T. (2008). Application of wrapper approach
study. So a 6-day long time period is considered for evaluation of
and composite classifier to the stock trend prediction. Expert Systems with Ap-
the proposed new models. plications, 34(4), 2870–2878.
The overall comparison between the three models (the base Huang, W., Nakamori, Y., & Wang, S. Y. (2005). Forecasting stock market movement
study and the two new proposed models of this study) are shown direction with support vector machine. Computers & Operations Research, 32(10),
2513–2522.
in Table 10. The results show that both of the new models are re- Jasemi, M., Kimiagari, A. M., & Memariani, A. (2011a). A conceptual model for port-
liable in 6-day periods while the SVM-ICA is better than the other folio management sensitive to mass psychology of market. International Journal
one and even the base study. The SVM-ICA hit rate is 76% while of Industrial Engineering-Theory Application and Practice, 18(1), 1–15.
Jasemi, M., Kimiagari, A. M., & Memariani, A. (2011b). A modern neural net-
the hit rate for the base study and the SVM-GA are 74.2% and 60% work model to do stock market timing on the basis of the ancient invest-
respectively. It is to be noted that the SVM-ICA got the fantastic ment technique of Japanese Candlestick. Expert Systems with Applications, 38(4),
average hit rate of 79% when the second approach is applied. 3884–3890.
Kuo, S. C., Lin, C. J., & Liao, J. R. (2011). 3D reconstruction and face recognition using
It is surprising that the 1-day hit rate of the base study is still kernel-based ICA and neural networks. Expert Systems with Applications, 38(5),
dominant to the others under any condition that suggests a good 5406–5415.
reason for a deeper research in the field. Finally it should be noted Lan, Q., Zhang, D., & Xiong, L. (2011). Reversal pattern discovery in financial time
series based on fuzzy candlestick lines. Systems Engineering Procedia, 2, 182–190.
that application of other methods on the basis of the conceptual
Lee, K. H., & Jo, G. S. (1999). Expert system for predicting stock market timing using
model presented by Jasemi et al. (2011b) to get better results is a candlestick chart. Expert systems with applications, 16(4), 357–364.
always a promising area for future research. Lee, M. C. (2009). Using support vector machine with a hybrid feature selection
method to the stock trend prediction. Expert Systems with Applications, 36(8),
10896–10904.
References
Majhi, B., Rout, M., & Baghel, V. (2014). On the development and performance eval-
uation of a multiobjective GA-based RBF adaptive model for the prediction of
Ahmadi, E., Abooie, M. H., Jasemi, M., & Zare Mehrjardi, Y. (2016). A nonlinear au- stock indices. Journal of King Saud University-Computer and Information Sciences,
toregressive model with exogenous variables neural network for stock market 26(3), 319–331.
timing: The candlestick technical analysis. International Journal of Engineering, Pai, P. F., & Hong, W. C. (2005). An improved neural network model in forecasting
29(12), 1717–1725. arrivals. Annals of Tourism Research, 32(4), 1138–1141.
Anbalagan, T., & Maheswari, S. U. (2015). Classification and prediction of stock mar- Pai, P. F., & Hong, W. C. (2006). Software reliability forecasting by support vector
ket index based on fuzzy metagraph. Procedia Computer Science, 47, 214–221. machines with simulated annealing algorithms. Journal of Systems and Software,
Ardjani, F., & Sadouni, K. (2010). Optimization of SVM multiclass by particle swarm 79(6), 747–755.
(PSO-SVM). I.J.Modern Education and Computer Science, 2, 32–38. Sahin, U., & Ozbayoglu, A. M. (2014). TN-RSI: Trend-normalized RSI indicator for
Atsalakis, G. S., & Valavanis, K. P. (2009). Surveying stock market forecasting tech- stock trading systems with evolutionary computation. Procedia Computer Sci-
niques–Part II: Soft computing methods. Expert Systems with Applications, 36(3), ence, 36, 240–245.
5932–5941. Sankar, C. P., Vidyaraj, R., & Kumar, K. S. (2015). Trust based stock recommenda-
Barak, S., Heidary, J., & Dahooie, T. T. (2015). Wrapper ANFIS-ICA method to do stock tion system–A social network analysis approach. Procedia Computer Science, 46,
market timing and feature selection on the basis of Japanese Candlestick. Expert 299–305.
Systems with Applications, 42(23), 9221–9235. Tay, F. E., & Cao, L. (2001). Application of support vector machines in financial time
Barak, S., Arjmand, A., & Ortobelli, S. (2017). Fusion of multiple diverse predictors in series forecasting. Omega, 29(4), 309–317.
stock market. Journal of Information Fusion, 36, 90–102. Vapnik, V. (1995). The nature of statistical learning theory. New York:
Bartz-Beielstein, T. (2010). SPOT: An R package for automatic and interactive tuning Springer_Verlag.
of optimization algorithms by sequential parameter optimization. arXiv preprint Vapnik, V. N. (1998). Statistical learning theory. New York: John Wiley & Sons.
arXiv:1006.4645. Wang, W., Xu, Z., & Weizhen Lu, J. (2003). Three improved neural network models
Boutte, D., & Santhanam, B. (2009). A hybrid ICA-SVM approach to continuous phase for air quality forecasting. Engineering Computations, 20(2), 192–210.
modulation recognition. IEEE Signal Processing Letters, 16(5), 402–405. Xie, H., Zhao, X., & Wang, S. (2012). A comprehensive look at the predictive infor-
Cherkassky, V., & Ma, Y. (2004). Practical selection of SVM parameters and noise mation in Japanese candlestick. Procedia Computer Science, 9, 1219–1227.
estimation for SVM regression. Neural Networks, 17(1), 113–126.
Cristianini, N., & Shawe-Taylor, J. (20 0 0). An introduction to support vector machines.
Cambridge University Press.