Information 15 00277
Information 15 00277
Article
Enhancing E-Learning Adaptability with Automated Learning
Style Identification and Sentiment Analysis: A Hybrid Deep
Learning Approach for Smart Education
Tahir Hussain 1 , Lasheng Yu 1, * , Muhammad Asim 2,3, * , Afaq Ahmed 1 and Mudasir Ahmad Wani 2
1 Department of Computer Science and Technology, Central South University, 932 South Lushan Rd,
Changsha 410083, China; [email protected] (T.H.); [email protected] (A.A.)
2 EIAS Data Science Lab, College of Computer and Information Sciences, Prince Sultan University,
Riyadh 11586, Saudi Arabia; [email protected]
3 School of Computer Science and Technology, Guangdong University of Technology,
Guangzhou 510006, China
* Correspondence: [email protected] (L.Y.); [email protected] (M.A.)
Abstract: In smart education, adaptive e-learning systems personalize the educational process by
tailoring it to individual learning styles. Traditionally, identifying these styles relies on learners
completing surveys and questionnaires, which can be tedious and may not reflect their true pref-
erences. Additionally, this approach assumes that learning styles are fixed, leading to a cold-start
problem when automatically identifying styles based on e-learning platform behaviors. To address
these challenges, we propose a novel approach that annotates unlabeled student feedback using
multi-layer topic modeling and implements the Felder–Silverman Learning Style Model (FSLSM) to
identify learning styles automatically. Our method involves learners answering four FSLSM-based
questions upon logging into the e-learning platform and providing personal information like age,
gender, and cognitive characteristics, which are weighted using fuzzy logic. We then analyze learners’
behaviors and activities using web usage mining techniques, classifying their learning sequences into
specific styles with an advanced deep learning model. Additionally, we analyze textual feedback
Citation: Hussain, T.; Yu, L.; Asim, M.;
using latent Dirichlet allocation (LDA) for sentiment analysis to enhance the learning experience
Ahmed, A.; Wani, M.A. Enhancing
further. The experimental results demonstrate that our approach outperforms existing models in
E-Learning Adaptability with
Automated Learning Style
accurately detecting learning styles and improves the overall quality of personalized content delivery.
Identification and Sentiment Analysis:
A Hybrid Deep Learning Approach Keywords: e-learning system; smart education; sentiment analysis; fuzzy weights; FSLSM model;
for Smart Education. Information 2024, deep learning; LSTM; LDA
15, 277. https://doi.org/10.3390/
info15050277
have unique ways of engaging with learning materials that shape their individual learning
styles [3]. These styles reflect their preferred methods for processing and understanding
information. Recognizing these styles is crucial for e-learning systems to tailor content that
enhances learning effectively. Traditionally, identifying learning styles involved having
students complete questionnaires. However, this method has several drawbacks: it is
often a boring and time-consuming task, students may not fully understand or be aware
of their learning styles, leading to arbitrary responses, and the results are static, failing to
account for changes in learning styles over time, such a type of learning style identification
approach is known as an explicit approach [4]. These issues highlight the need for a more
dynamic and engaging approach to accurately ascertain students’ learning preferences.
Various automated methods have been developed to identify students’ learning styles
by analyzing their interactions with e-learning systems to address these limitations. These
automatic identification techniques offer several advantages over traditional methods.
Firstly, they eliminate the need for time-consuming questionnaires by gathering data
directly from students’ activities within the system [5]. Additionally, unlike static results
from questionnaires, the learning styles identified through these automated approaches are
dynamic; they can adapt and change in response to shifts in students’ behaviors, ensuring
a more accurate and personalized learning experience; such a type of approach is known as
an implicit approach [6]. The implicit approach is much better than the explicit approach,
but it may face a cold-start problem when new users log into the system. To overcome
these issues, this paper combined both approaches as we collected a minimal number of
input characteristics such as learners’ previous score (categorized by fuzzy weight logic),
age, gender, and lastly, filling our four questions which represented the FSLSM dimension.
Implementing automatic identification of learning styles requires using a learning style
(LS) model, which classifies students based on their preferred learning methods [7]. Various
learning style models have been discussed in this paper. However, the Felder–Silverman
Learning Style Model (FSLSM) is identified as particularly effective for adaptive e-learning
systems according to recent studies [8]. In this paper, the FSLSM model was chosen for the
reasons described in the next section.
This paper proposes a hybrid approach to improve the e-learning system by integrating
automatic learning style identification and simultaneously conducting sentiment analysis
to improve the learning object quality further. To identify automatic learning styles, using
web usage mining techniques, we analyze student behavior and obtain minimal attributes
from learners when they log into the e-learning platform. The collected data on student
activities from the e-learning platform’s log files organizes this information into sequences.
Each sequence consists of the learning objects that a student accessed during a session.
These learning objects are then aligned with the learning style combinations outlined in the
FSLSM. These sequences of student activities, their input attributes, and their corresponding
learning objects are used as input for long short-term memory (LSTM), a recurrent neural
network (RNN) architecture capable of learning long-term dependencies in data sequences.
This algorithm maps the sequences to various learning style categories.
Additionally, this paper analyzes textual feedback to improve the learning objects
further by using LDA for sentiment analysis to examine textual feedback from learners.
This paper proposes a novel approach to annotating unlabeled student feedback using
multi-layer topic modeling and also introduces and implements a novel algorithm for
sentiment extraction and mapping. Our proposed approach combines both implicit and
explicit approaches of learning style identification methods, which removes the limitations
of both approaches that exist individually. The experimental results of this paper not only
outperform existing models in accurately detecting learning styles but also enhance the
quality of learning objects, improving the overall adaptive e-learning systems.
This paper contributes the following points:
1. By using minimal learner inputs and learning sequences of learners, this paper de-
signs LSTM positional encoding to classify learners; our novel approach improves
the accuracy of learning style identification compared to other traditional machine
Information 2024, 15, 277 3 of 30
2. Related Works
In academic research, a range of classification methods have been applied to automati-
cally determine learning styles and sentiment across different models. Researchers have
utilized these techniques to better understand and cater to individual learning preferences,
enhancing the personalization of educational content.
According to Graf [9], a data-driven approach utilizing artificial intelligence algorithms
has been developed to automatically detect learning styles from learner interactions within
a system. This method uses real behavioral data as input, with the algorithm outputting
the learner’s style preferences, thereby enhancing accuracy. To ensure meaningful data
for classification, web mining techniques are employed to extract detailed behavioral
information, as detailed by Mahmood [10], making the system both effective and efficient
in adapting to individual learning needs.
In [11], a novel method was introduced to identify each learner’s style using the
FSLSM by extracting behaviors from Moodle logs. Decision trees were utilized for dynamic
classification based on these styles, with the method’s accuracy evaluated by comparing
behaviors to quiz results provided at the end of a course. This approach, however, was
tested with a limited sample of 35 learners in a single Moodle-based online course.
In [12], the authors integrated fuzzy logic with neural networks to train an algorithm
capable of recognizing various learning styles. However, the algorithm’s effectiveness was
limited to classifying just three dimensions of the FSLSM model: perception, input, and
understanding.
In [13], the authors employed Bayesian networks to analyze learner data from logs
of chats, forums, and processing activities, detecting only three Felder–Silverman learn-
ing styles: perception, processing, and understanding. Their study illustrates the use
of predictive modeling to discern specific educational traits based on interactive online
behaviors.
Fuzzy logic has been utilized to automatically determine learners’ styles, as demon-
strated by Troussas [14]. Expanding on this, Crockett [15] developed a fuzzy classification
tree within a predictive model that employs independent variables captured through
natural language dialogue, enhancing the precision of style assessments.
In [16], the authors applied the Fuzzy C-Means (FCM) algorithm to categorize learning
behavioral data into FSLSM categories. This clustering approach enabled a structured
analysis of behavioral patterns, aligning them with defined learning styles to enhance
personalized education strategies.
In [17], a model named adaptive e-learning recommender model using learning style
and knowledge-level modeling (AERM-KLLS) was developed to enhance student engage-
ment and performance through personalized materials using questionnaires and adaptive
Information 2024, 15, 277 4 of 30
Figure 1. Overview of the FSLSM: its four key dimensions, and their associated values [39].
Information 2024, 15, 277 6 of 30
3.2. Methodology
This paper’s aim is to improve adaptability in e-learning systems through learning
style identification and sentiment analysis. It aims to address critical issues such as the
inefficient identification of learning styles through traditional questionnaire-based meth-
ods, which are time-consuming and lack dynamic responsiveness to individual learner
preferences [44]. Additionally, it seeks to address the lack of incorporation of learners’
sentiments and feedback, which are vital for refining e-learning content and improving
learner experiences. To tackle these challenges, this paper introduces a novel hybrid
approach that integrates automatic learning style identification and sentiment analysis.
This approach aims to dynamically identify learners’ preferences and sentiments towards
educational content, thus facilitating a more personalized and engaging learning experi-
ence. By overcoming the limitations of traditional methods and incorporating sentiment
analysis, the proposed system seeks to enhance the effectiveness of e-learning platforms,
ultimately aiming to improve learner engagement, satisfaction, and educational outcomes.
The workflow starts when learners access the e-learning platform; initially, they have to
answer the four questions which are based on the FSLSM dimension [45], and additionally
provide their personal information such as age, gender, and prior academic performance.
Considering the uncertainty in grade classification, fuzzy weights are used to evaluate
prior academic performance. For example, a score of 75% might be seen as both somewhat
good and somewhat excellent. When the learners interact with learning objects in the
e-learning platform, the learners’ behaviors are extracted using web usage mining and
stored in the database, and their learning sequences are captured to identify their learning
style dynamically. To use a supervised learning algorithm, it is essential to convert the
sequences of learner interactions extracted from log files into a suitable input format for
the algorithm. A learner’s interaction sequence is characterized by the different learning
objects they engage with during a session. Each of these sequences is detailed with iden-
tifiers for the sequence, session, and learner, along with a list of the learning objects the
learner interacted with within that session. Following the extraction of these interaction
Information 2024, 15, 277 7 of 30
sequences, they are categorized based on the FSLSM, where each sequence is associated
with a distinct combination of learning styles. This paper used advanced deep learning
techniques, such as a Transformer model with LSTM positional encoding, to automatically
classify learners based on their learning styles because of its ability to handle sequential
data, leveraging its ability to remember long-term dependencies for tasks as compared to
other algorithms. This classifier takes into account various learner attributes, including
age, previous academic performance, gender, and learning sequences. This method enables
an automated process to determine the learning styles of students, aligning them with
specific dimensions without the necessity of completing the entire 44-question FSLSM
questionnaire, is a process known for being long and tedious. Additionally, this approach
utilizes the powerful LDA algorithm to extract key topics from student feedback, enabling
it to unearth valuable insights. As students engage with the e-learning platform, feedback
is collected through surveys and questionnaires. This feedback undergoes thorough pre-
processing to ensure accuracy and meaningful analysis. Using LDA, we identify prominent
aspects within the feedback, shedding light on student preferences, concerns, and senti-
ments. Furthermore, the extracted aspects and sentiments are subjected to aspect-based
sentiment analysis (ABSA), which assesses the sentiment associated with each aspect and
develops a novel algorithm for analyzing aspect-oriented feedback to determine each
piece of feedback’s sentiment orientation (positive, negative, or neutral). Its purpose is to
automatically categorize user feedback based on sentiment, helping to understand overall
user sentiment towards different aspects of feedback. This multifaceted analysis provides
critical insights that guide improvements in the overall e-learning experience. By employ-
ing a sentiment classification algorithm such as the convolutional neural network (CNN),
sentiment labels are assigned to learner feedback, categorizing it as positive, negative, or
neutral based on sentiment scores calculated using a sentiment lexicon. In the final stages of
the workflow, this approach summarizes the results of the aspect-based sentiment analysis
(ABSA) and sends them to educational institutions and instructors. This summary drives
the process of updating course materials and resources in alignment with the feedback
and sentiments expressed by learners. Ultimately, this comprehensive approach paves the
way for more engaging, adaptable, and satisfying e-learning experiences for students. The
overall framework is shown in Figure 2.
Figure 3. Fuzzy weighting for a student’s prior academic performance. (a) Mathematical Definitions
for Membership Functions (b) Graphical Representation of Membership Functions.
The membership functions described above assign values to the three fuzzy sets
(µb , µ I , µ A ) representing each student’s prior academic performance (Figure 3). These
numbers go from 0 to 1, where 1 denotes a total understanding of the subject matter being
taught. As a result, the equation µ B ( x ) + µ I ( x ) + µ A ( x ) = 1 states that the total of the
partition values for these fuzzy sets for the domain notion is always 1. These fuzzy sets
and their membership function thresholds were determined by a panel of 16 experts from
various domains. Among them, 10 were computer science experts from public universities’
computer science departments, and 6 were pedagogical experts from public universities’
education departments. These experts, each with over 10 years of experience, provided
descriptive assessments of student progress and performance levels along with defining
success intervals for each knowledge level.
xij
Pij = (1)
xi
a hyperparameter responsible for governing the weighting process and holds significant
importance in the context of GloVe. GloVe’s primary optimization objective lies in the
minimization of the loss function illustrated in Equation (2). This optimization is conducted
under the constraint outlined in Equation (3).
v v
J= ∑ ∑ f ( pij ) · (wiT · w j + bi + bj − log (Xij ))2 (2)
i =1 j =1
qi K jT qi K jT
! !
Lk
1
M̂(qi , K ) = max √
j d
−
LK ∑ √
d
(4)
j =1
The proposed approach makes use of “Log-Sum-Exp” to ascertain the qi average and
peak levels. When a relatively high M(qi , k) value is chosen, it amplifies the importance
of the attention probability P(k j |qi ). This querying technique effectively filters out less
critical information, creating a novel probabilistic sparse matrix, denoted as Q, which
encapsulates the most relevant specifics from M(qi , K ). By employing the sample sampling
factor u = c ln Q in the dot product calculation, we manage to reduce the complexity to
O( LK ln LQ). When calculating the dot product, the sample sampling factor u = cInQ
decreases in complexity to O( LK InLQ ). The dot product is then used to generate the
confidence score P, which highlights the factor that is being given the most attention.
Higher confidence scores indicate stronger associations. Equation (5) displays the formula
for P.
Q̂ · K
P= √ (5)
d
This approach uses the top-k(n) method to extract relevant information from P based
on the confidence score P. For each row in P, we isolate the top n scores, forming a pool
of key indices, where n is adjusted to c. This strategy maintains the intended meaning
while emphasizing the most important information. The final values are set to zero. This
strategy streamlines the model by selecting the most important scores while preserving
important data. The conversion of P into a sparse matrix P is achieved by utilizing the
softmax function, and this process is an integral part of the regularization procedure.
Finally, Equation (6) shows the ultimate result, S.
(θ jT xi )
e
P ( yi = j | x i ; θ ) = T i
(7)
∑kn=1 e(θn x )
3.7. Dataset
The dataset used in this paper is publicly available, originates from a prominent
northern Indian university, and serves as the foundation for generating the institutional
report through the analysis of student feedback data [62]. It encompasses six essential
categories: teaching quality, course content, lab experiences, library facilities, and the
environment of the institute. Each category in the dataset is represented by two columns,
allowing for sentiment labels—0 (neutral), 1 (positive), or −1 (negative)—to be assigned.
This dataset provides a comprehensive view of student opinions, making it a valuable
resource for evaluating various facets of the university experience, and the sample dataset
is shown in Figure 7.
Information 2024, 15, 277 13 of 30
3.8. Preprocessing
The data were cleaned during the preprocessing procedure [63] to enable efficient
computation. In the initial stages of text data preprocessing, our approach incorporates a
series of essential steps to ensure the consistency and quality of textual data for subsequent
analysis. These steps include the following.
• Word Tokenization: Word tokenization involves the task of segmenting the text into
individual words or tokens. It divides the text into discrete units where each unit is
typically a word [64]. For example, the sentence “The quick brown fox” would be
transformed into tokens as follows: [“The” “quick” “brown” “fox”]. Tokenization is
essential because it transforms unstructured text into a structured format that com-
puters can process. It enables subsequent analysis, such as counting word frequency,
determining sentence structure, and identifying significant terms.
• Lowercase Conversion: Lowercase conversion involves changing all the tokens in
the text to lowercase [65]. For instance, “Word” becomes “word”. This step ensures
uniformity in text analysis by treating words with different letter cases as identical.
Lowercasing simplifies text analysis by reducing the influence of letter case variations,
allowing for more accurate word matching and counting.
• Removal of Numbers and Punctuation: This step focuses on removing numerical
characters and punctuation marks from the text [66]. For example, When we transform
“The price is $100” into “The price is”, we eliminate the numbers and punctuation. This
simplification process streamlines the text, removing non-textual elements, and thus,
making it more amenable to the extraction of meaningful information from the text.
• Exclusion of Symbols and Hyphens: Symbols like “@” or “#” and hyphens are excluded
from the text. These characters often do not carry significant linguistic meaning and can
interfere with text analysis [67]. Excluding symbols and hyphens maintains the focus on
the textual content, reducing noise and distractions from non-textual elements.
• Stop Word Removal: Stop words consist of common, uninformative words such as “of”
“they” “is” “on”, etc. In this step, these words are removed from the text. Removing
stop words is critical for improving feature extraction because it reduces the influence
of frequently occurring but less meaningful words [68]. This enhances the quality of
text analysis and gives more attention to content-rich terms.
• Stemming and Lemmatization: Stemming and lemmatization are techniques to reduce
words to their base forms. Stemming typically removes suffixes, resulting in a root
form (e.g., “running” → “run”) [69]. Lemmatization focuses on semantic relevance,
ensuring that words are reduced to their dictionary form (e.g., “better” → “good”).
Using the LDA technique extracts aspect terms from unlabeled data for students’
textual feedback. These hidden aspects in student feedback were discovered using an
unsupervised LDA model similar to the latent topics represented by documents. The
generative workflow of LDA involves documents as mixtures over latent topics with words
characterized by topic-specific word probabilities. This approach allowed us to uncover
hidden aspects of student feedback effectively [71].
The generative process of LDA is shown in Figure 9, presenting an overview of its
key elements. In this context, ‘D’ symbolizes the corpus containing many documents, ‘z’
denotes topics, and ‘w’ represents the words present within these documents. Additionally,
within this visual representation, ‘α’ and ‘β’ are utilized to signify Dirichlet distributions,
while ‘θ’ and ‘ϕ’ are indicative of multinomial distributions.
j =1 i =1
P(W, Z, θ, ϕ, α, β) = ∏ P(θ j ; α) ∏ P(ϕi ; β)
M K
(8)
t =1
∏ P(Zj,t |θ j ) P(Wj,t |ϕzj,t )
N
Equation (8) depicts the LDA probability model, which consists of two sides. On the
left-hand side, the document’s probability is a mixture of latent topics, while the right-hand
side involves four distinct factors. The first two factors serve as the foundation of LDA,
while the latter two act as its driving mechanisms. Each factor calculates probabilities
and multiplies them to determine the final document probability. To be more specific,
the first term on the right-hand side corresponds to topics and is characterized by a
Dirichlet distribution, while the second factor represents words, also following a Dirichlet
distribution. In the same way, the third element of the equation is linked with topics, while
the fourth is associated with words. Both elements are defined by multinomial distributions.
Information 2024, 15, 277 15 of 30
Hyperparameter Value
Number of Topics 3
Number of Passes 10
Document–Topic Prior Alpha 0.01
Beta (Previous Topic–Word) 0.1
Random State 42
Maximum Iterations 500
Minimum Token Frequency 10
Maximum Features 5000
Chunk Size 2000
Learning Decay 0.7
Learning Offset 10
Perplexity Score Tolerance 0.1
Mean Change Tolerance 0.001
Learning Method ‘batch’
Evaluate Every 1
with zeros. Here, ‘Xi ’ represents an input instance while ‘xi ’ denotes the ith word within
that instance.
Layer 0: This initial layer incorporates word embeddings using Word2Vec representing
words as low-dimensional vectors ‘vi ∈ Rk’ where ‘k’ denotes the dimension of the word
vectors. These word vectors are concatenated to form the input instance ‘X ∈ Rl × k’.
Layer 1: In this layer, a one-dimensional convolution operation is applied with filter
weights ‘m’ of width ‘n’. For each word vector ‘vi ’, feature vectors are generated by
capturing contextual information within a window of size ‘n’. The output is calculated
using the ReLU activation function:
f j = m T v j:j+n−1
(9)
O = σ( f j + b)
Accuracy (A): Accuracy reflects the proportion of total predictions that a model has
classified correctly.
TP + TN
A=
TP + TN + FP + FN
F1 score: Harmonic mean of precision and recall, balancing both metrics.
P×R
F1 = 2 ×
P+R
LSTM Algorithm 75 82 72 75 91 98 87 80 74 76 71 74 90 95 86 72
MCQ Method 73 84 69 77 93 97 85 83 78 72 69 78 89 93 88 70
K-Means Algorithm 65 70 72 67 94 88 76 84 65 77 74 61 87 95 86 74
Information 2024, 15, 277 19 of 30
The 16 FSLSM categories are used to label the sequences. Since some sequences, based
on feature values, belong to more than one cluster, the total number of sequences clustered
using the LSTM algorithm is 1298, which is greater than the 1235 input sequences. In our
paper, i = 60, 100, and 200, and c = 4, n = 1235, d = 16, c = 4 correspond to the number of
data points, k to the number of clusters, c to the number of dimensions, and i to the number
of iterations. The results of our computation and a comparison of the temporal complexity
of both techniques are shown in Table 5.
Table 5. The time complexity of the proposed approach and FCM (Fuzzy C-Means) based on the
number of iterations.
The cross-validation outcomes for every dimension and algorithm are shown in
Table 6. Even though the proposed method has the highest precision, recall, and accuracy
scores in terms of size, as can be shown, the cross-validation score is just 78%. Other
algorithms with high cross-validation ratings are logistic regression random forest and
linear discriminant analysis. How will this help researchers and practitioners? Discussing
two significant aspects of this research will help to address this query. One is the distinction
and resemblance between hand marking and machine prediction in terms of consistency,
as can be seen in Table 7.
Algorithms Dimensions
Input Processing Understanding Perception
Proposed Approach 78% 83% 83% 89%
Random Forest 79% 77% 77% 91%
Classifier Using Decision Tree 73% 74% 74% 90%
Naive Bayes 77% 76% 76% 86%
Logistic Regression 79% 80% 80% 87%
K-Nearest Neighbor 76% 78% 78% 90%
Analyzing Linear Discriminant 79% 79% 79% 86%
Cross-Validation Scores
Table 7. The degree of consistency between machine prediction and manual marking about Gardner’s
multiple intelligence theory.
Figure 10. Performance evaluation based on accuracy and loss metrics: (a) accuracy and validation
accuracy; (b) training and validation loss.
Additionally, we evaluated the proposed model’s performance for each learning style,
as shown in Table 9. This detailed analysis offers precise insights into the model’s accuracy
within specific learning style categories, affirming its effectiveness in personalized learning
assessments.
Extracted Sentiment
Feedback
Aspect Orientation
1 Teacher’s behavior isn’t great for everyone. Teacher Negative
2 University should have more fun stuff, not just studying. University Neutral
The environment of the university is so bad, the environment should be clean,
3 University Negative
I don’t want to stay here because of poor environment.
4 I like the programming part of the course. Course Positive
Course,
5 The materials are not up to date, the teacher must have to provide latest materials. Negative
teacher
6 The materials we learned are not up to date; they are too old. Course Negative
7 The course selection procedure is so bad. Course Negative
8 University has less teaching staff, if increase teaching staff it will be good. University General
9 This course is interesting. Course Positive
10 University’s new library resources are very poor. University Negative
11 University providing more facilities. University Positive
12 Teacher well prepared lecture. Teacher Positive
Table 12. Extracted topics and their corresponding terms obtained through LDA.
Topic_Perc
Topic_Num Keywords Representative Text
_Contrib
Course, learn, study, teach, program,
[learn, research, based, course, dip, machine
0.0 0.9251 interesting, achieve, skill, understand,
learning, vision]
help
Teacher, empower, overlook, often, bad, [teacher, bad habit, tell, teacher, thing, tell,
1.0 0.9390
due, love, give, job, time parent, exact, opposite]
University, provide, experience, big, op-
[educational, institution, play, key, role, city,
2.0 0.9209 portunity, develop, game, offer, staff,
region, operate]
library
Information 2024, 15, 277 23 of 30
Figure 12. Performance evaluation of the model by analyzing accuracy and loss graphs based on our
specific dataset.
Figure 13. Illustrating the performance of our trained models using a confusion matrix.
Figure 14. Comparative evaluation of the proposed model with other existing models. In this context,
(a–c) signify the performance metrics concerning accuracy, precision, and recall, respectively.
Figure 15. Proposed model vs. other models for learning style identification.
In the bar chart for sentiment analysis of learner feedback shown in Figure 16, the
proposed model’s top performance is consistent, registering an accuracy of 93.81%. This
suggests that the model is not only good at identifying patterns in data for categorization
but is also adept at interpreting the nuances of language—a key feature for analyzing
feedback. The “Weakly Supervised Long Short-Term Memory (WS-LSTM)” [82], which
is only slightly less accurate, also demonstrates strong sentiment analysis capabilities.
However, as we move down the list to models like the “Recommendation System Based on
Sentiment Analysis (RSBSA)” [83], “Enhanced e-Learning Hybrid Recommender System
(ELHRS)” [84], and “Bidirectional Long Short-Term Memory (BiLSTM)” [85], the drop in
accuracy becomes more pronounced, indicating potential challenges in fully capturing
learner sentiments, especially when dealing with subtle or complex expressions.
The time complexity graph shown in Figure 17 compares the proposed algorithm
against the FCM (Fuzzy C-Means) algorithm over different iteration counts. Although the
proposed algorithm takes longer to execute than the FCM algorithm, the time increase is
not as pronounced as the iteration count grows. For instance, at 200 iterations, the proposed
algorithm takes 1389 s compared to 2529 s for the FCM algorithm, demonstrating better
scalability. This demonstrates that the proposed algorithm is more efficient at handling
large-scale data, a desirable trait for e-learning systems that need to process information
from many users simultaneously.
Figure 16. Proposed model vs. other models for sentiment classification performance.
Information 2024, 15, 277 26 of 30
The time complexity graph shown in Figure 17 compares the proposed algorithm
against the FCM (Fuzzy C-Means) algorithm over different iteration counts.
Figure 17. Proposed model and FCM algorithm time complexity comparison.
Although the proposed algorithm takes longer to execute than the FCM algorithm,
the time increase is not as pronounced as the iteration count grows. For instance, at
200 iterations, the proposed algorithm takes 1389 s compared to 2529 s for the FCM
algorithm, demonstrating better scalability. This demonstrates that the proposed algorithm
is more efficient at handling large-scale data, a desirable trait for e-learning systems that
need to process information from many users simultaneously.
of 95.54%, recall of 95.55%, and F1 score of 95.54%. The sentiment analysis further explores
student feedback using LDA for topic modeling. This approach effectively uncovers
key aspects and sentiments within the feedback, providing a nuanced understanding of
student opinions on different facets of the e-learning system. The performance of the
sentiment analysis was evaluated using a two-layer convolutional neural network (CNN),
which achieved a peak accuracy of 93.81%. In conclusion, the proposed hybrid deep
learning method revolutionizes e-learning by precisely tailoring content to individual
learning styles, enhancing student engagement and success. By analyzing sentiments, it
also offers valuable insights into student feedback, enabling continuous improvement. This
approach significantly boosts the effectiveness and accessibility of online education, directly
impacting students’ learning outcomes. In practice, it empowers educators to deliver more
personalized, responsive, and impactful e-learning experiences.
In the future, we aim to apply the proposed learning style identification across various
domains and among diverse age groups, including tertiary education students. Addition-
ally, we plan to develop an integrated LS model, combining the FSLSM model with other
cognitive frameworks. This will allow for a more comprehensive understanding of factors
influencing e-learning. The identification of these hybrid learning styles will be automated
through the use of advanced intelligent techniques. The proposed aspect sentiment anal-
ysis approach deals only with English language comments. However, in Pakistan, most
students use Roman Urdu in their feedback, so there is a significant opportunity to enhance
the system’s inclusivity and effectiveness. By incorporating the ability to process Roman
Urdu, future iterations of the proposed approach could delve deeper into understanding
the nuances of student feedback. Furthermore, students often employ a range of symbols
and emoticons to express their opinions in online feedback systems. Therefore, a critical
area of future research will involve examining how these visual elements correlate with the
sentiments being expressed. By systematically analyzing the weight and sentiment of each
attribute, future research aims to refine sentiment analysis techniques, thereby achieving
greater precision in classifying sentiments.
Author Contributions: Conceptualization, T.H., L.Y., M.A. and A.A.; methodology, T.H., L.Y., M.A.
and A.A.; software, T.H., L.Y. and M.A.; validation, L.Y., M.A. and M.A.W.; formal analysis, L.Y., M.A.
and M.A.W.; investigation, M.A. and M.A.W.; resources, L.Y. and M.A.W.; writing—original draft,
T.H.; writing—review and editing, L.Y., M.A., A.A. and M.A.W.; supervision, L.Y.; funding acquisition,
M.A. and M.A.W. All authors have read and agreed to the published version of the manuscript.
Funding: The authors would like to thank Prince Sultan University for paying the APC of this article.
Institutional Review Board Statement: Not applicable.
Informed Consent Statement: Not applicable.
Data Availability Statement: The data presented in this study are available on request from the
corresponding author.
Acknowledgments: This work was supported by EIAS Data Science Lab, College of Computer and
Information Sciences, Prince Sultan University. The authors would like to thanks Prince Sultan University
for their support.
Conflicts of Interest: The authors declare no conflicts of interest.
References
1. Imran, M.; Almusharraf, N.; Abdellatif, M.S.; Ghaffar, A. Teachers’ perspectives on effective English language teaching practices
at the elementary level: A phenomenological study. Heliyon 2024, 10, e29175. [CrossRef] [PubMed]
2. Farooq, U.; Naseem, S.; Mahmood, T.; Li, J.; Rehman, A.; Saba, T.; Mustafa, L. Transforming educational insights: Strategic
integration of federated learning for enhanced prediction of student learning outcomes. J. Supercomput. 2024, 1–34. [CrossRef]
3. Sivarajah, R.T.; Curci, N.E.; Johnson, E.M.; Lam, D.L.; Lee, J.T.; Richardson, M.L. A review of innovative teaching methods. Acad.
Radiol. 2019, 26, 101–113. [CrossRef] [PubMed]
4. Karagiannis, I.; Satratzemi, M. An adaptive mechanism for Moodle based on automatic detection of learning styles. Educ. Inf.
Technol. 2018, 23, 1331–1357. [CrossRef]
Information 2024, 15, 277 28 of 30
5. Granić, A. Educational technology adoption: A systematic review. Educ. Inf. Technol. 2022, 27, 9725–9744. [CrossRef] [PubMed]
6. Shoeibi, A.; Khodatars, M.; Jafari, M.; Ghassemi, N.; Sadeghi, D.; Moridian, P.; Khadem, A.; Alizadehsani, R.; Hussain, S.; Zare,
A.; et al. Automated detection and forecasting of COVID-19 using deep learning techniques: A review. Neurocomputing 2024,
577, 127317. [CrossRef]
7. Zhang, H.; Huang, T.; Liu, S.; Yin, H.; Li, J.; Yang, H.; Xia, Y. A learning style classification approach based on deep belief network
for large-scale online education. J. Cloud Comput. 2020, 9, 1–17. [CrossRef]
8. Muhammad, B.A.; Qi, C.; Wu, Z.; Ahmad, H.K. An evolving learning style detection approach for online education using
bipartite graph embedding. Appl. Soft Comput. 2024, 152, 111230. [CrossRef]
9. Graf, S. Adaptivity in Learning Management Systems Focussing on Learning Styles. Ph.D. Thesis, Technische Universität Wien,
Vienna , Austria, 2007. Available online: http://hdl.handle.net/20.500.12708/10843 (accessed on 27 March 2024).
10. Jalal, A.; Mahmood, M. Students’ behavior mining in e-learning environment using cognitive processes with information
technologies. Educ. Inf. Technol. 2019, 24, 2797–2821. [CrossRef]
11. Abdullah, M.A. Learning style classification based on student’s behavior in moodle learning management system. Trans. Mach.
Learn. Artif. Intell. 2015, 3, 28. [CrossRef]
12. Zatarain-Cabada, R.; Barrón-Estrada, M.L.; Angulo, V.P.; García, A.J.; García, C.A.R. A learning social network with recognition of
learning styles using neural networks. In Proceedings of the Advances in Pattern Recognition: Second Mexican Conference on Pattern
Recognition, MCPR 2010, Puebla, Mexico, 27–29 September 2010; Springer: Berlin/Heidelberg, Germany, 2010; pp. 199–209. [CrossRef]
13. García, P.; Amandi, A.; Schiaffino, S.; Campo, M. Evaluating Bayesian networks’ precision for detecting students’ learning styles.
Comput. Educ. 2007, 49, 794–808. [CrossRef]
14. Troussas, C.; Chrysafiadi, K.; Virvou, M. An intelligent adaptive fuzzy-based inference system for computer-assisted language
learning. Expert Syst. Appl. 2019, 127, 85–96. [CrossRef]
15. Crockett, K.; Latham, A.; Mclean, D.; O’Shea, J. A fuzzy model for predicting learning styles using behavioral cues in an
conversational intelligent tutoring system. In Proceedings of the 2013 IEEE International Conference on Fuzzy Systems
(FUZZ-IEEE), Hyderabad, India, 7–10 July 2013; IEEE: New York, NY, USA, 2013; pp. 1–8. [CrossRef]
16. Kolekar, S.V.; Pai, R.M.; MM, M.P. Prediction of Learner’s Profile Based on Learning Styles in Adaptive E-learning System. Int. J.
Emerg. Technol. Learn. 2017, 12, 31–51. [CrossRef]
17. Aziz, A.S.; El-Khoribi, R.A.; Taie, S.A. Adaptive E-learning recommendation model based on the knowledge level and learning
style. J. Theor. Appl. Inf. Technol. 2021, 99, 5241–5256.
18. Kaouni, M.; Lakrami, F.; Labouidya, O. The design of an adaptive E-learning model based on Artificial Intelligence for enhancing
online teaching. Int. J. Emerg. Technol. Learn. 2023, 18, 202. [CrossRef]
19. Madhavi, A.; Nagesh, A.; Govardhan, A. A framework for automatic detection of learning styles in e-learning. AIP Conf. Proc.
2024, 2802, 120012. [CrossRef]
20. Rashid, A.B.; Ikram, R.R.R.; Thamilarasan, Y.; Salahuddin, L.; Abd Yusof, N.F.; Rashid, Z.B. A Student Learning Style Auto-
Detection Model in a Learning Management System. Eng. Technol. Appl. Sci. Res. 2023, 13, 11000–11005. [CrossRef]
21. Essa, S.G.; Celik, T.; Human-Hendricks, N. Personalised adaptive learning technologies based on machine learning techniques to
identify learning styles: A systematic literature review. IEEE Access 2023, 11, 48392–48409. [CrossRef]
22. Alshmrany, S. Adaptive learning style prediction in e-learning environment using levy flight distribution based CNN model.
Clust. Comput. 2022, 25, 523–536. [CrossRef]
23. Raleiras, M.; Nabizadeh, A.H.; Costa, F.A. Automatic learning styles prediction: A survey of the State-of-the-Art (2006–2021).
J. Comput. Educ. 2022, 9, 587–679. [CrossRef]
24. Gomede, E.; Miranda de Barros, R.; de Souza Mendes, L. Use of deep multi-target prediction to identify learning styles. Appl. Sci.
2020, 10, 1756. [CrossRef]
25. Khan, F.A.; Akbar, A.; Altaf, M.; Tanoli, S.A.K.; Ahmad, A. Automatic student modelling for detection of learning styles and
affective states in web based learning management systems. IEEE Access 2019, 7, 128242–128262. [CrossRef]
26. Pashler, H.; McDaniel, M.; Rohrer, D.; Bjork, R. Learning styles: Concepts and evidence. Psychol. Sci. Public Interest 2008,
9, 105–119. [CrossRef] [PubMed]
27. Hauptman, H.; Cohen, A. The synergetic effect of learning styles on the interaction between virtual environments and the
enhancement of spatial thinking. Comput. Educ. 2011, 57, 2106–2117. [CrossRef]
28. Jegatha Deborah, L.; Baskaran, R.; Kannan, A. Learning styles assessment and theoretical origin in an E-learning scenario: A
survey. Artif. Intell. Rev. 2014, 42, 801–819. [CrossRef]
29. Choudhary, L.; Swami, S. Exploring the Landscape of Web Data Mining: An In-depth Research Analysis. Curr. J. Appl. Sci.
Technol. 2023, 42, 32–42. [CrossRef]
30. Roy, R.; Giduturi, A. Survey on pre-processing web log files in web usage mining. Int. J. Adv. Sci. Technol. 2019, 29, 682–691.
31. Fawzia Omer, A.; Mohammed, H.A.; Awadallah, M.A.; Khan, Z.; Abrar, S.U.; Shah, M.D. Big Data Mining Using K-Means and
DBSCAN Clustering Techniques. In Big Data Analytics and Computational Intelligence for Cybersecurity; Springer: Berlin/Heidelberg,
Germany, 2022; pp. 231–246. [CrossRef]
32. Nafea, S.M.; Siewe, F.; He, Y. On recommendation of learning objects using felder-silverman learning style model. IEEE Access
2019, 7, 163034–163048. [CrossRef]
Information 2024, 15, 277 29 of 30
33. Deng, Y.; Lu, D.; Chung, C.J.; Huang, D.; Zeng, Z. Personalized learning in a virtual hands-on lab platform for computer science
education. In Proceedings of the 2018 IEEE Frontiers in Education Conference (FIE), San Jose, CA, USA, 3–6 October 2018; IEEE:
New York, NY, USA, 2018; pp. 1–8. [CrossRef]
34. Hu, J.; Peng, Y.; Chen, X.; Yu, H. Differentiating the learning styles of college students in different disciplines in a college English
blended learning setting. PLoS ONE 2021, 16, e0251545. [CrossRef]
35. Hmedna, B.; El Mezouary, A.; Baz, O. How does learners’ prefer to process information in MOOCs? A data-driven study. Procedia
Comput. Sci. 2019, 148, 371–379. [CrossRef]
36. Reardon, M.; Derner, S. Strategies for Great Teaching: Maximize Learning Moments; Taylor & Francis: Abingdon, UK, 2023. [CrossRef]
37. Seghroucheni, Y.Z.; Chekour, M. How Learning Styles Can Withstand the Demands of Mobile Learning Environments? Int. J.
Interact. Mob. Technol. 2023, 17, 84–99. [CrossRef]
38. Othmane, Z.; Derouich, A.; Talbi, A. A comparative study of the Most influential learning styles used in adaptive educational
environments. Int. J. Adv. Comput. Sci. Appl. 2019, 10, 520–528.
39. Sihombing, J.H.; Laksitowening, K.A.; Darwiyanto, E. Personalized e-learning content based on felder-silverman learning style
model. In Proceedings of the 2020 8th International Conference on Information and Communication Technology (ICoICT),
Yogyakarta, Indonesia, 24–26 June 2020; IEEE: New York, NY, USA, 2020; pp. 1–6. [CrossRef]
40. Hidayat, N.; Wardoyo, R.; Sn, A.; Surjono, H.D. Enhanced performance of the automatic learning style detection model using a
combination of modified K-means algorithm and Naive Bayesian. Int. J. Adv. Comput. Sci. Appl. 2020, 11, 638–648. [CrossRef]
41. Staudemeyer, R.C. Applying long short-term memory recurrent neural networks to intrusion detection. S. Afr. Comput. J. 2015,
56, 136–154. [CrossRef]
42. Kumaravel, G.; Sankaranarayanan, S. PQPS: Prior-Art Query-Based Patent Summarizer Using RBM and Bi-LSTM. Mob. Inf. Syst.
2021, 2021, 1–19. [CrossRef]
43. Pamir; Javaid, N.; Javaid, S.; Asif, M.; Javed, M.U.; Yahaya, A.S.; Aslam, S. Synthetic theft attacks and long short term
memory-based preprocessing for electricity theft detection using gated recurrent unit. Energies 2022, 15, 2778. [CrossRef]
44. Saberi, N.; Montazer, G.A. A new approach for learners’ modeling in e-learning environment using LMS logs analysis. In
Proceedings of the 6th National and 3rd International Conference of E-Learning and E-Teaching, Tehran, Iran, 3–4 March 2012;
IEEE: New York, NY, USA, 2012; pp. 25–33. [CrossRef]
45. Graf, S.; Viola, S.R.; Leo, T.; Kinshuk. In-depth analysis of the Felder-Silverman learning style dimensions. J. Res. Technol. Educ.
2007, 40, 79–93. [CrossRef]
46. Prottasha, N.J.; Sami, A.A.; Kowsher, M.; Murad, S.A.; Bairagi, A.K.; Masud, M.; Baz, M. Transfer learning for sentiment analysis
using BERT based supervised fine-tuning. Sensors 2022, 22, 4157. [CrossRef] [PubMed]
47. Stevens-Smith, D.A. Brain-based teaching: Differentiation in teaching, learning, and motor skills. J. Phys. Educ. Recreat. Danc.
2020, 91, 34–42. [CrossRef]
48. Hall, A.H.; Gao, Q.; Guo, Y.; Xie, Y. Examining the effects of kindergarten writing instruction on emergent literacy skills: A
systematic review of the literature. Early Child Dev. Care 2023, 193, 334–346. [CrossRef]
49. Al-Maroof, R.S.; Alahbabi, N.M.N.; Akour, I.; Alhumaid, K.; Ayoubi, K.; Alnnaimi, M.; Thabit, S.; Alfaisal, R.; Aburayya, A.;
Salloum, S. Students’ perception towards behavioral intention of audio and video teaching styles: An acceptance study. Int. J.
Data Netw. Sci. 2022, 6, 603. [CrossRef]
50. Hakami, Z. Comparison between virtual and traditional learning methods for orthodontic knowledge and skills in dental
students: A Quasi-Experimental Study. Healthcare 2021, 9, 1092. [CrossRef] [PubMed]
51. Ariastuti, M.D.; Wahyudin, A.Y. Exploring academic performance and learning style of undergraduate students in English
Education program. J. Engl. Lang. Teach. Learn. 2022, 3, 67–73. [CrossRef]
52. Chen, J.; Zhang, Y.; Wei, Y.; Hu, J. Discrimination of the contextual features of top performers in scientific literacy using a machine
learning approach. Res. Sci. Educ. 2021, 51, 129–158. [CrossRef]
53. Saadat, M.; Bayat, M. Prediction of the unconfined compressive strength of stabilised soil by Adaptive Neuro Fuzzy Inference
System (ANFIS) and Non-Linear Regression (NLR). Geomech. Geoengin. 2022, 17, 80–91. [CrossRef]
54. Venkatesh, V.; Zhang, A.; Deakins, E.; Luthra, S.; Mangla, S. A fuzzy AHP-TOPSIS approach to supply partner selection in
continuous aid humanitarian supply chains. Ann. Oper. Res. 2019, 283, 1517–1550. [CrossRef]
55. Selva Birunda, S.; Kanniga Devi, R. A review on word embedding techniques for text classification. In Innovative Data
Communication Technologies and Application: Proceedings of ICIDCA 2020; Springer: Singapore, 2021; pp. 267–281. [CrossRef]
56. Yang, L.; Zhou, G. Dissecting The Analects: An NLP-based exploration of semantic similarities and differences across English
translations. Humanit. Soc. Sci. Commun. 2024, 11, 50. [CrossRef]
57. Pan, M.; Huang, J.X.; He, T.; Mao, Z.; Ying, Z.; Tu, X. A simple kernel co-occurrence-based enhancement for pseudo-relevance
feedback. J. Assoc. Inf. Sci. Technol. 2020, 71, 264–281. [CrossRef]
58. Vilar, E. Word embedding, neural networks and text classification: What is the state-of-the-art? J. Manag. Sci. 2019, 4, 35–62.
[CrossRef]
59. Ma, K.; Tan, Y.; Tian, M.; Xie, X.; Qiu, Q.; Li, S.; Wang, X. Extraction of temporal information from social media messages using
the BERT model. Earth Sci. Informatics 2022, 15, 573–584. [CrossRef]
60. Le, T.H.; Chen, H.; Babar, M.A. Deep learning for source code modeling and generation: Models, applications, and challenges.
ACM Comput. Surv. (CSUR) 2020, 53, 1–38. [CrossRef]
Information 2024, 15, 277 30 of 30
61. Garg, S.; Roy, D. A Birds Eye View on Knowledge Graph Embeddings, Software Libraries, Applications and Challenges. arXiv
2022, arXiv:2205.09088. [CrossRef].
62. Khaiser, F.K.; Saad, A.; Mason, C. Sentiment analysis of students’ feedback on institutional facilities using text-based classification
and natural language processing (NLP). J. Lang. Commun. 2023, 10, 101–111. [CrossRef]
63. Tabassum, A.; Patil, R.R. A survey on text pre-processing & feature extraction techniques in natural language processing. Int. Res.
J. Eng. Technol. 2020, 7, 4864–4867.
64. Dogra, V.; Verma, S.; Chatterjee, P.; Shafi, J.; Choi, J.; Ijaz, M.F. A complete process of text classification system using state-of-the-art
NLP models. Comput. Intell. Neurosci. 2022, 2022, 1883698. [CrossRef] [PubMed]
65. Bhattacharjee, A.; Karami, M.; Liu, H. Text transformations in contrastive self-supervised learning: A review. arXiv 2022,
arXiv:2203.12000. [CrossRef].
66. Egger, R.; Gokce, E. Natural Language Processing (NLP): An Introduction: Making Sense of Textual Data. In Applied Data Science in
Tourism: Interdisciplinary Approaches, Methodologies, and Applications; Springer: Berlin/Heidelberg, Germany, 2022; pp. 307–334. [CrossRef]
67. Altrabsheh, N.; Cocea, M.; Fallahkhair, S. Sentiment analysis: Towards a tool for analysing real-time students feedback. In Proceedings
of the 2014 IEEE 26th International Conference on Tools with Artificial Intelligence, Limassol, Cyprus, 10–12 November 2014; IEEE:
New York, NY, USA, 2014; pp. 419–423. [CrossRef]
68. Kalaivani, E.R.; Marivendan, E.R. The effect of stop word removal and stemming in datapreprocessing. Ann. Rom. Soc. Cell Biol.
2021, 25, 739–746.
69. Pramana, R.; Subroto, J.J.; Gunawan, A.A.S.; Anderies. Systematic literature review of stemming and lemmatization performance
for sentence similarity. In Proceedings of the 2022 IEEE 7th International Conference on Information Technology and Digital
Applications (ICITDA), Yogyakarta, Indonesia, 4–5 November 2022; IEEE: New York, NY, USA , 2022; pp. 1–6. [CrossRef]
70. Zhou, F.; Ayoub, J.; Xu, Q.; Jessie Yang, X. A machine learning approach to customer needs analysis for product ecosystems.
J. Mech. Des. 2020, 142, 011101. [CrossRef]
71. Xuan, T.Y.; Yahya, N.; Khan, Z.; Badruddin, N.; Yusoff, M.Z. EEG Motor Classification Using Multi-band Signal and Common Spa-
tial Filter. In Proceedings of the International Conference on Intelligent Human Computer Interaction, Daegu, Republic of Korea,
24–26 November 2020; Springer: Berlin/Heidelberg, Germany, 2020; pp. 120–131. [CrossRef]
72. Shams, H.; Jan, T.; Khalil, A.A.; Ahmad, N.; Munir, A.; Khalil, R.A. Fingerprint image enhancement using multiple filters. PeerJ
Comput. Sci. 2023, 9, e1183. [CrossRef]
73. Kumar, A.; Malik, A.; Batra, I.; Ahmad, N.; Johar, S. Digital society social interactions and trust analysis model. PeerJ Comput. Sci.
2022, 8, e1129. [CrossRef]
74. Novera, C.N.; Ahmed, Z.; Kushol, R.; Wanke, P.; Azad, M.A.K. Internet of Things (IoT) in smart tourism: A literature review.
Span. J. Mark. ESIC 2022, 26, 325–344. [CrossRef]
75. Khoo, C.S.; Johnkhan, S.B. Lexicon-based sentiment analysis: Comparative evaluation of six sentiment lexicons. J. Inf. Sci. 2018,
44, 491–511. [CrossRef]
76. Syed, S.; Spruit, M. Exploring symmetrical and asymmetrical Dirichlet priors for latent Dirichlet allocation. Int. J. Semant. Comput.
2018, 12, 399–423. [CrossRef]
77. Azzi, I.; Jeghal, A.; Radouane, A.; Yahyaouy, A.; Tairi, H. A robust classification to predict learning styles in adaptive E-learning
systems. Educ. Inf. Technol. 2020, 25, 437–448. [CrossRef]
78. Alruwais, N.; Zakariah, M. Student-Engagement Detection in Classroom Using Machine Learning Algorithm. Electronics 2023,
12, 731. [CrossRef]
79. El Aissaoui, O.; El Madani, Y.E.A.; Oughdir, L.; El Allioui, Y. Combining supervised and unsupervised machine learning
algorithms to predict the learners’ learning styles. Procedia Comput. Sci. 2019, 148, 87–96. [CrossRef]
80. Aziz, A.S.; El-Khoribi, R.A.; Taie, S.A. AFCM model to predict the learner style based on questionnaire and fuzzy C mean
algorithm. J. Theor. Appl. Inf. Technol. 2021, 99, 343–356.
81. Ayyoub, H.Y.; Al-Kadi, O.S. Learning Style Identification Using Semi-Supervised Self-Taught Labeling. IEEE Trans. Learn. Technol.
2024, 17, 1093–1106. [CrossRef]
82. Kastrati, Z.; Imran, A.S.; Kurti, A. Weakly supervised framework for aspect-based sentiment analysis on students’ reviews of
MOOCs. IEEE Access 2020, 8, 106799–106810. [CrossRef]
83. Alatrash, R.; Priyadarshini, R.; Ezaldeen, H.; Alhinnawi, A. Augmented language model with deep learning adaptation on
sentiment analysis for E-learning recommendation. Cogn. Syst. Res. 2022, 75, 53–69. [CrossRef]
84. Ezaldeen, H.; Misra, R.; Bisoy, S.K.; Alatrash, R.; Priyadarshini, R. A hybrid E-learning recommendation integrating adaptive
profiling and sentiment analysis. J. Web Semant. 2022, 72, 100700. [CrossRef]
85. Alzaid, M.; Fkih, F. Sentiment Analysis of Students’ Feedback on E-Learning Using a Hybrid Fuzzy Model. Appl. Sci. 2023,
13, 12956. [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual
author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to
people or property resulting from any ideas, methods, instructions or products referred to in the content.