Classification of Customer Reviews Using Machine Learning Algorithms (1)
Classification of Customer Reviews Using Machine Learning Algorithms (1)
An International Journal
Behrooz Noori
To cite this article: Behrooz Noori (2021) Classification of Customer Reviews Using
Machine Learning Algorithms, Applied Artificial Intelligence, 35:8, 567-588, DOI:
10.1080/08839514.2021.1922843
ABSTRACT
The information resulting from the use of the organization's
products and services is a valuable resource for business analy
tics. Therefore, it is necessary to have systems to analyze custo
mer reviews. This article is about categorizing and predicting
customer sentiments. In this article, a new framework for cate
gorizing and predicting customer sentiments was proposed.
The customer reviews were collected from an international
hotel. In the next step, the customer reviews processed, and
then entered into various machine learning algorithms. The
algorithms used in this paper were support vector machine
(SVM), artificial neural network (ANN), naive bayes (NB), decision
tree (DT), C4.5 and k-nearest neighbor (K-NN). Among these
algorithms, the DT provided better results. In addition, the
most important factors influencing the great customer experi
ence were extracted with the help of the DT. Finally, very inter
esting results were observed in terms of the effect of the
number of features on the performance of machine learning
algorithms.
Introduction
Customer reviews have been commonly recognized as valuable sources for
marketing intelligence and sentiment analysis (Dickinger and Mazanec
2015). Sentiment analysis seeks to build a system for analyzing and
evaluating customer reviews reflected on websites, blogs, Twitter, or
Instagram. In recent years, with the expansion of online systems, custo
mer reviews have a powerful impact on business development and attract
ing potential customers. Therefore, review categorization becomes the key
technology to organize textual data. Review categorization is defined as
assigning new documents to a set of pre-defined categories based on the
classification patterns (Uğuz 2011). In fact, customer sentiment is very
important for the hospitality industry and plays an important role in
providing better quality services (e.g., more adaptation to customer
requirements and customization of services), better customer relationship
Related Works
Sentiment Analytics
On the other word, due to the increasing growth of social networks and
websites, the obtained reviews are useful resources for analyzing and improv
ing business and developing products and services. Often these reviews are not
structured. With the help of machine learning methods, managers will be
provided with information for future use.
The most important task of sentiment analytics is analyzing the polarity of
sentiments and classifying texts into positive and negative emotional cate
gories. To successfully discover, interpret, and communicate extracted opi
nions and detected polarity, sentiment analytics relies on multidisciplinary
efforts including natural language processing and machine learning (Fu et al.
2018).
Existing approaches to sentiment analytics can be classified into two broad
categories: semantic orientation approaches and machine learning approaches
(Fu et al. 2018). Semantic orientation approaches hold that text is classified
into affect categories on the basis of the presence of fairly unambiguous affect
words, such as “happy,” “sad,” “afraid,” and “bored.” Semantic orientation
approaches are popular thanks to their accessibility and economy. However,
the weaknesses of these approaches include poor affect recognition given
complex linguistic rules, and heavy dependence on the depth and breadth of
the employed lexicon resources (Fu et al. 2018). For a domain lacking of such
resources, machine learning approaches can mitigate the above limitations. By
feeding a machine learning algorithm a training corpus of affectively anno
tated texts, machine learning approaches can not only learn the affective
polarity of affect keywords but can also consider the polarity of other arbitrary
keywords and word co-occurrence frequencies (Cambria 2017). However,
machine learning approaches rely on statistical models that are meaningful
when given a sufficiently large text input; therefore, the approaches can
achieve better performance on the document or paragraph level compared
to smaller text units, such as sentences or clauses (Fu et al. 2018).
With the proliferation of big data, there is an increasing trend for hotel
industry researchers to adopt computational methods in their studies (Fu et al.
2018). In particular, sentiment analytics, which works as an effective method
to automatically extract public opinions and analyze sentiment polarity from
massive textual data (Cambria 2016), has thus piqued researchers’ interest (Fu
et al. 2018).
A large amount of studies by different authors were conducted where
machine learning procedures were undertaken on hotel industry and hospi
tality data (Moro, Rita, and Coelho 2017). Ye, Zhang, and Law (2009) studied
sentiment classification methods in online reviews from travel blogs, compar
ing them with three machine learning techniques (Ye, Zhang, and Law 2009).
Cao, Duan, and Gan (2011) studied the impact of online review features
hidden in the reviews on the number of helpful votes by applying text mining
for extracting the review’s characteristics. In order to meet the requirement of
570 B. NOORI
Classification
Classification is one of the most commonly used methods in machine learning.
It is a process of finding a set of models that allows data classes to be identified
and distinguished. The aim of classification is to determine the class of future
data objects by using past information. In classification, a training set is usually
used to learn the model, and the learned information is then tested on the test
set. Many classification algorithms have been developed in the literature so far
since there is no perfect algorithm for all data sets (Gulsoy and Kulluk 2019).
(1) What are the key steps of sentiment analytics for hotel industry?
(2) What are the key design factors of feature engineering?
(3) How do these design factors influence the predictive accuracy of senti
ment analytics for hotel industry?
(4) How can machine learning methods be systematically incorporated to
improve the predictive accuracy of sentiment analytics?
The parts of proposed system structure are shown in Figure 1. These parts are
explained in the following subsections:
Data Collection
Data Cleaning
- duplicate removal, delete digits, punctuation marks, and numbers.
- delete stop-words and non-useful words like: pronouns, articles and
propositions. (Bahassine et al. 2018)
Removing of Stop-words
Words such as conjunctions and pronouns that are not related to the concept
of the text are called stop-words. This process involves removing certain
common words such as ‘a’, ‘an’, the’, etc., that occur commonly in all docu
ments. It is important to removing these high-frequency words because they
may misclassify the documents (Uğuz 2011).
Stemming
Stemming is a process of reducing inflected words into one form (stem or
root) by removing prefixes, suffixes and infixes (Bahassine et al. 2018). The
stemming process leaves out the root forms of the words. Thereby, terms
sharing the same root that seem like different words due to their affixes can be
determined. For example, “computer”, “computing”, “computation”, and
“computes” all have the same comput root (Uğuz 2011).
Term Weighting
After the words are transformed into terms, the presentation form of the
document, which means the expression thereof, terms have to be determined.
This process is called term weighting. Thereby, each document could be
written in a vector form depending on the terms they contained. To obtain
the weight vector, term frequency–inverse document frequency (TF-IDF)
feature weighting algorithm is used as its weight scheme. N-gram refers to
a sequence of n tokens based on words.
seen that a number of elements in the term vector is “0”. Therefore, we prune
the words that appear less than two times in the documents. This process
decreases the term vector dimension further (Uğuz 2011).
Feature Selection
Feature selection is a process that selects a subset from the original feature set
according to some criteria of feature importance (Uğuz 2011). A major pro
blem of sentiment categorization is the high dimensionality of the feature
space due to a large number of terms. This problem may cause the computa
tional complexity of machine learning methods used for sentiment categor
ization to be increased and may bring about inefficiency and results of low
accuracy due to irrelevant terms in the feature space. For a solution to this
problem, two techniques are used in this study: feature ranking and feature
selection (Uğuz 2011).
Sentiment mining has become a heated research in recent years. One of the
important means of sentiment mining is sentiment categorization. For many
problems of sentiment categorization, a good feature selection method can not
only reduce the computational complexity but also increase the categorization
performance. Feature selection is a process that selects a set of new features
from the original features and forms a distinct feature space. Apart from this,
feature selection is also perceived as a prerequisite for text categorization, so its
significance and importance can be imagined (Wang et al. 2015).
On the other hand, feature extraction produces a large feature set and
creates a high-dimensional vector space, which will ultimately lower the
efficiency and the effectiveness of sentiment classification. As a result, it is
critical to select features with significant sentiment distinguishing ability and
reduce the dimension of vector space (Wang et al. 2013). Features selection is
effective in the reduction of large data in text classification. It can enhance the
classification process. Feature selection deletes irrelevant and noisy data and
chooses a representative subset of all data to minimize the complexity of the
classification process (Dadgar, Araghi, and Farahani 2016).
Numerous techniques of features selection can be detected in the literature
such as: Chi-square (Bahassine et al. 2018) and Gini index (Manek et al. 2017).
The present research tried to introduce a modified version of Gini feature
selection method which will be presented hereafter (Bahassine et al. 2018).
A Gini Index-based feature selection method solves the problem mentioned
above. The experiments showed that the weight by Gini Index method has
better classification performance (Manek et al. 2017). At the end of the feature
selection step, terms of high importance in documents are acquired through
the Gini method.
In the current study, feature selection and feature sentiment extraction are
used to manage the high dimensionality of a feature space composed of a large
574 B. NOORI
number of terms, remove redundant and irrelevant features from the feature
space and thereby decrease the computational complexity of the machine
learning algorithms used in the text categorization and increase performances
thereof (Uğuz 2011).
In the first stage, each term in the text is ranked depending on their
importance for the classification in decreasing order using the Gini method.
Therefore, terms of high importance are assigned to the first ranks and terms
of less importance are assigned to the following ranks. In the second stage, the
PCA method selected for feature sentiment extraction.
Classification Methods
In this study, six separate machine learning methods are used in text categor
ization; the SVM, ANN, NB, DT, C4.5 and kNN methods are used due to their
usability and accuracy in text categorization. The reason for using a classifier is
to compare the performances of the six methods in the text categorization.
Brief descriptions of these methods are given, as follows.
SVM
SVM is a supervised learning method used for categorization. It is a useful
methodology that finds the best possible surface to separate the positive
samples from the negative samples. The basic goal of SVM, behind the training
process, is to find a maximum margin hyperplane to solve the feature review’s
classification task. There are unlimited possible boundaries to separate the two
classes. To select the best class, it is important to choose a decision boundary
that has a maximum margin between any points from both classes. The
decision boundary with a maximum margin would be less likely to make
prediction errors, which is close to the boundaries of one of the classes (Ali,
Kwak, and Kim 2016). The dot kernel type was selected because of its better
APPLIED ARTIFICIAL INTELLIGENCE 575
Naive Bayes
NB Tree is a supervised classifier that combines Bayesian rule and decision
tree. This algorithm uses the Bayes rule to calculate the likelihood of each
given class of instances, assuming that the properties are the conditional
576 B. NOORI
independent of the label given (Gulsoy and Kulluk 2019). A Naive Bayes
classifier is a simple probabilistic classifier based on applying Bayes’ theorem
with strong (naive) independence assumptions. In simple terms, a Naive Bayes
classifier assumes that the presence (or absence) of a particular feature of
a class (i.e. attribute) is unrelated to the presence (or absence) of any other
feature. For example, a fruit may be considered to be an apple if it is red,
round, and about 4 inches in diameter. Even if these features depend on each
other or upon the existence of the other features, a Naive Bayes classifier
considers all of these properties to independently contribute to the probability
that this fruit is an apple.
The advantage of the Naive Bayes classifier is that it only requires a small
amount of training data to estimate the means and variances of the variables
necessary for classification. Because independent variables are assumed, only
the variances of the variables for each label need to be determined and not the
entire covariance matrix.
Decision Trees
A decision tree is a tree-like graph or model. It is more like an inverted tree
because it has its root at the top and it grows downwards. This representation
of the data has the advantage compared with other approaches of being
meaningful and easy to interpret. Each interior node of tree corresponds to
one of the input attributes. The number of edges of a nominal interior node is
equal to the number of possible values of the corresponding input attribute.
Outgoing edges of numerical attributes are labeled with disjoint ranges. Each
leaf node represents a value of the label attribute given the values of the input
attributes represented by the path from the root to the leaf. Decision trees in
the Simple Cart algorithm are constructed by dividing each decision node into
two different branches using various separation criteria (Gulsoy and Kulluk
2019).
KNN
The KNN algorithm is a well-known instance-based approach that has been
widely applied to text categorization due to its simplicity and accuracy. To
categorize an unknown document, the KNN classifier ranks the document’s
neighbors among the training documents and uses the class labels of the
k most similar neighbors. Similarity between two reviews may be measured
by the Euclidean distance, cosine measure, etc. The similarity score of each
nearest neighbor document to the test document is used as the weight of the
classes of the neighbor document. If a specific category is shared by more than
one of the k-nearest neighbors, then the sum of the similarity scores of those
neighbors is obtained from the weight of that particular shared category (Uğuz
2011).
At the phase when classification is done by means of the KNN, the most
important parameter affecting classification is k-nearest neighbor number.
Usually, the optimal value of k is empirically determined. In our study,
k value is determined so that it would give the better classification accuracy
(k = 1 is determined). In addition, in the phase of finding the k-nearest
neighborhood, mix Euclidean distance is used as the distance metric (Uğuz
2011).
Results
Pre-processing and Feature Extraction
Performance Comparison
The accuracy performance metric values obtained by machine learning meth
ods on the sentiment data set are shown in Table 1. When Table 1 is examined,
it can be seen that the NB is given the maximum accuracy (65.5%) in first
dataset. Therefore, NB is very suitable for small feature set. It can be seen that
the SVM is given the maximum accuracy (68.6%) in second dataset. Moreover,
DT and C4.5 are given the maximum accuracy (98.9%) in third dataset. The
obtained accuracy is pretty high for real world problems. In the third dataset,
SVM algorithm is the second algorithm with 75.1% accuracy, and it is followed
by K-NN, and NB, respectively. ANN algorithm is the worst algorithm with
55.1% accuracy on large feature set. When the obtained maximum number of
features is taken into consideration, DT and C4.5 algorithms are the best
algorithms. Therefore, the DT and C4.5 produced significant results in terms
of accuracy and third feature set. The overall accuracies of the algorithms are
indicated in Figure 2. The accuracies of the selected algorithms are indicated in
Figure 3.
When the performances of the algorithms are evaluated in the first data set
in terms of recall performance metric, ANN classifier can be seen to give the
best result of 0.633 (Table 2). NB, K-NN, SVM, DT and C4.5 algorithms are
following this algorithm, respectively for first feature set. Furthermore, SVM
can be seen to give the best result of 0.675. ANN, K-NN, NB, C4.5 and DT
algorithms are following this algorithm, respectively for second feature set.
Recall values support the results obtained with accuracy values. Figure 4.
shows recall behavior. The DT classifier algorithm gave the best value of
0.98 for the recall performance metric in third feature set.
The DT classifier algorithm gave the best value of 0.991 for the precision
performance metric. The C4.5 is the second algorithm with 0.985 precision
value. SVM, K-NN, NB and ANN algorithms followed, respectively, the C4.5
algorithm according to precision (Table 3, Figure 5).
Additionally, the DT classifier algorithm gave the best value of 0.991 for the
F-measure performance metric, which is the harmonic mean of the precision
and recall metrics (Table 4). The C4.5 is the second algorithm with
0.985 F-measure value. SVM, K-NN, NB and ANN algorithms followed,
Meanwhile in the third feature set, the C4.5 algorithm gave the best result
with a value of 0.979 for the ROC area performance measure, which expresses
the residual under the ROC curve (Table 5). DT, SVM, NB, K-NN, and ANN
algorithms, respectively, followed the C4.5 algorithm.
Accordingly, it is seen that DT algorithm gives more favorable results than
the other 5 algorithms when the accuracy, precision, recall, and F-measure
metrics are taken into consideration. The C4.5 algorithm yielded more favor
able results for the area under the ROC curve only. In summary, both DT and
C4.5 algorithms can be used for classification purposes on the large feature set.
The DT and C4.5 classifiers may be to maximize the accuracy, to maximize the
APPLIED ARTIFICIAL INTELLIGENCE 583
area under the ROC-curve, or to minimize the square root of the mean square
error. As a result, with the DT, 98.9% accuracy value of the process of
584 B. NOORI
sentiment classification in the hotel sector was obtained. This value is quite
high for real-life data.
Yet, the SVM and NB algorithms are preferred in small and medium sized
feature sets in this study because it has high accuracy and F-measure as well as
being interpretable by the user. The results show that data mining can be used
to generate an objective measurement system that can be used for the senti
ment classification process of customers in the hotel industry.
Figure 9 shows results as a function of the number of selected terms that
resulted in the best accuracy for DT and C4.5, respectively. Considering our
results, we observed the following:
(7) All machine learning methods increase their accuracy as the number of
features increases, but ANN behaved differently, and its accuracy
decrease as the number of features increases.
(8) DT has a better performance of sentiment classification compared to the
other popular classifiers in large feature set and NB and SVM have
better performance of sentiment classification compared to the other
popular classifiers in smaller feature set.
Conclusion
Exploring and analyzing the invaluable hidden data of customer reviews has
become a major prerequisite for effective and successful marketing analytics.
The large volume of data and Big Data has become a fundamental feature of
today’s society, yet its ability to analyze, solidify and learn from them, has
become a useful element for competition as well as supporting the growth of
productivity and innovation. Although big data can certainly be considered as
good for decision, large data does not lead to better marketing because they are
related to some of the key challenges and issues (i.e. the lack of feature
engineering, need for useful analysis, and so on). Machine learning techniques
play an important role in the field of social media analysis (Ducange, Pecori,
and Mezzina 2018).
The sentiment classification, which make up a large part of the voice of
customer in the hotel sector, has become one of the most important problems
of the hospitality industry. This is because the estimation of whether the
customers will be none satisfied is a rather complex problem for the hotels.
Studies in the literature are predominantly related to the process for sentiment
classification. However, there are very few studies on the process of feature
identification, feature engineering and performance comparison of machine
learning algorithm. For this reason, in order to fill the gap in the literature,
feature identification and engineering for this sector has been studied. Then
real-life data was collected from an international five star Iranian hotel chain.
Moreover, the data were classified by using six different machine learning
algorithms. According to the results, sentiment predictions can be done with
98.9% accuracy.
The experimental results indicated the following main conclusions:
(3) Comparing the six machine learning methods, NB performed the best
with fewer features.
(4) The performance of almost all methods increased with increasing fea
tures except ANN. In other words, increasing the features will certainly
increase the accuracy of the model, except in the case of ANN.
Further research will be conducted in the following aspects: (1) reviews with
a neutral sentiment (neither positive nor negative), clustered SVM, multiclass
SVM and its contribution to sentiment classification will be discussed; (2) The
effect that punctuation (e.g. ‘!’, ‘?’, etc.) has on sentiment classification will be
analyzed; and (3) The ensemble learning (e.g., bagging and voting) in senti
ment classification.
References
Ali, F., K. Kwak, and Y. Kim. 2016. Opinion mining based on fuzzy domain ontology and
support vector machine: A proposal to automate online review classification. Applied Soft
Computing 47:235–50. doi:10.1016/j.asoc.2016.06.003.
Al-Smadi, M., O. Qawasmeh, M. Al-Ayyoub, Y. Jararweh, and B. Gupta. 2018. Deep Recurrent
neural network vs. support vector machine for aspect-based sentiment analysis of Arabic
hotels’ reviews. Journal of Computational Science 27:386–93. doi:10.1016/j.jocs.2017.11.006.
Bahassine, S., A. Madani, M. Al-Sarem, and M. Kissi. 2018. Feature selection using an
improved Chi-square for Arabic text classification. Journal of King Saud University-
Computer and Information Sciences 32(2):225-231. https://doi.org/10.1016/j.jksuci.2018.05.
010
Bi, J., Y. Liu, Z. Fan, and E. Cambria. 2019. Modelling customer satisfaction from online
reviews using ensemble neural network and effect-based Kano model. International Journal
of Production Research 57(22):1–21. https://doi.org/10.1080/00207543.2019.1574989
Cambria E., D. Das, S. Bandyopadhyay, and A. Feraco. 2017. Affective computing and senti
ment analysis. In A Practical Guide to Sentiment Analysis. Socio-Affective Computing vol 5.,
ed. Cambria E., Das D., Bandyopadhyay S., Feraco A., vol 5. Cham: Springer. https://doi.org/
10.1007/978-3-319-55394-8_1
Cao, Q., W. Duan, and Q. Gan. 2011. Exploring determinants of voting for the “helpfulness” of
online user reviews: A text mining approach. Decision Support Systems 50:511–21.
doi:10.1016/j.dss.2010.11.009.
Cardot, H., and D. Degras. 2018. Online principal component analysis in high dimension:
Which algorithm to choose? International Statistical Review 86:29–50. doi:10.1111/
insr.12220.
Dadgar, S., M. Araghi, and M. Farahani 2016. A novel text mining approach based on TF-IDF
and support vector machine for news classification. In A novel text mining approach based
on TF-IDF and support vector machine for news classification. 2016 IEEE International
Conference on Engineering and Technology (ICETECH), IEEE, Coimbatore, India. 112–16.
Dickinger, A., and J. Mazanec. 2015. Significant word items in hotel guest reviews: A feature
extraction approach. Tourism Recreation Research 40:353–63. doi:10.1080/
02508281.2015.1079964.
APPLIED ARTIFICIAL INTELLIGENCE 587
Duan, W., Y. Yu, Q. Cao, and S. Levy. 2016. Exploring the impact of social media on hotel
service performance: A sentimental analysis approach. Cornell Hospitality Quarterly
57:282–96. doi:10.1177/1938965515620483.
Ducange, P., R. Pecori, and P. Mezzina. 2018. A glimpse on big data analytics in the framework
of marketing strategies. Soft Computing 22:325–42. doi:10.1007/s00500-017-2536-4.
Fu, Y., J. Hao, X. Li, C. Hsu, X. Li, C. Hsu, and C. Hsu. 2018. Predictive accuracy of sentiment
analytics for tourism: A metalearning perspective on Chinese travel news. Journal of Travel
Research 57:0047287518772361. doi:10.1177/0047287517700317.
Gulsoy, N., and S. Kulluk. 2019. A data mining application in credit scoring processes of small
and medium enterprises commercial corporate customers. Wiley Interdisciplinary Reviews:
Data Mining and Knowledge Discovery 9:e1299.
Hu, Y., and K. Chen. 2016. Predicting hotel review helpfulness: The impact of review visibility,
and interaction between hotel stars and review ratings. International Journal of Information
Management 36:929–44. doi:10.1016/j.ijinfomgt.2016.06.003.
Ma, E., M. Cheng, and A. Hsiao. 2018. Sentiment analysis–a review and agenda for future
research in hospitality contexts. International Journal of Contemporary Hospitality
Management 30:3287–308. doi:10.1108/IJCHM-10-2017-0704.
Manek, A., P. Shenoy, M. Mohan, and K. Venugopal. 2017. Aspect term extraction for
sentiment analysis in large movie reviews using Gini Index feature selection method and
SVM classifier. World Wide Web 20:135–54. doi:10.1007/s11280-015-0381-x.
Moro, S., P. Rita, and J. Coelho. 2017. Stripping customers’ feedback on hotels through data
mining: The case of Las Vegas Strip. Tourism Management Perspectives 23:41–52.
doi:10.1016/j.tmp.2017.04.003.
Sánchez-Franco, M., A. Navarro-García, and F. Rondán-Cataluña. 2019. A naive Bayes strategy
for classifying customer satisfaction: A study based on online reviews of hospitality services.
Journal of Business Research 101:499-506.
Schuckert, M., X. Liu, and R. Law. 2015. Hospitality and tourism online reviews: Recent trends
and future directions. Journal of Travel & Tourism Marketing 32(5): 608–621. https://doi.
org/10.1080/10548408.2014
Talón-Ballestero, P., L. González-Serrano, C. Soguero-Ruiz, S. Muñoz-Romero, and J. Rojo-
Álvarez. 2018. Using big data from customer relationship management information systems
to determine the client profile in the hotel sector. Tourism Management 68:187–97.
doi:10.1016/j.tourman.2018.03.017.
Tran, T., H. Ba, and V. Huynh 2019. Measuring hotel review sentiment: An aspect-based
sentiment analysis approach. In: Seki H., Nguyen C., Huynh V. N., Inuiguchi M. (eds)
Integrated uncertainty in knowledge modelling and decision making. IUKM 2019. Lecture
notes in computer science, vol 11471. Springer, Cham. https://doi.org/10.1007/978-3-030-
14815-7_33
Tripathy, A., A. Agrawal, and S. Rath. 2016. Classification of sentiment reviews using n-gram
machine learning approach. Expert Systems with Applications 57:117–26. doi:10.1016/j.
eswa.2016.03.028.
Uğuz, H. 2011. A two-stage feature selection method for text categorization by using informa
tion gain, principal component analysis and genetic algorithm. Knowledge-Based Systems
24:1024–32. doi:10.1016/j.knosys.2011.04.014.
Uysal, A., and S. Gunal. 2014. The impact of preprocessing on text classification. Information
Processing & Management 50:104–12. doi:10.1016/j.ipm.2013.08.006.
Wang, F., C. Li, J. Wang, J. Xu, and L. Li. 2015. A two-stage feature selection method for text
categorization by using category correlation degree and latent semantic indexing. Journal of
Shanghai Jiaotong University (Science) 20:44–50. doi:10.1007/s12204-015-1586-y.
588 B. NOORI
Wang, H., P. Yin, J. Yao, and J. Liu. 2013. Text feature selection for sentiment classification of
Chinese online reviews. Journal of Experimental & Theoretical Artificial Intelligence
25:425–39. doi:10.1080/0952813X.2012.721139.
Ye, Q., Z. Zhang, and R. Law. 2009. Sentiment classification of online reviews to travel
destinations by supervised machine learning approaches. Expert Systems with Applications
36:6527–35. doi:10.1016/j.eswa.2008.07.035.
Zainuddin, N., A. Selamat, and R. Ibrahim. 2018. Hybrid sentiment classification on twitter
aspect-based sentiment analysis. Applied Intelligence 48:1218–1232.