1 s2.0 S2666307424000184 Main
1 s2.0 S2666307424000184 Main
Keywords: In the rapidly advancing era of educational technology, customized learning materials have the potential
Blending ensemble technique to enhance individuals’ learning capacities. This research endeavors to devise an effective method for
Content modification detecting a learner’s preferred learning style and subsequently adapting the learning content to align with
Learning style detection
that style, utilizing artificial intelligence AI techniques. Our investigation finds that analyzing learners’ web
Named entity recognition
tracking logs for activity classification and categorizing individual responses for feedback classification are
Text-to-text transfer transformer (T5) model
highly effective methods for identifying a learner’s learning styles, such as visual, auditory, and kinesthetic.
A custom dataset has been constructed in this research comprising approximately 506 samples and 22
features utilizing the Moodle learning management system (LMS), successfully categorizing students into
their respective learning styles. Furthermore, decision tree, random forest, support vector machine (SVM),
logistic regression, XGBoost, blending ensemble, and convolutional neural network (CNN) algorithms with
corresponding optimized hyperparameters and synthetic minority oversampling technique (SMOTE) have been
applied for learning behavior classification. The blending ensemble technique with the XGBoost meta-learning
model accomplished the best performance for learning style detection with an accuracy of 97.56%. Next, the
text content of the electronic documents is modified by employing different natural language processing (NLP)
techniques, including named entity recognition of spaCy, knowledge graph, generative pre-trained transformer
3 (GPT-3), and text-to-text transfer transformer (T5) model, to accommodate diverse learning styles. Various
approaches, such as color coding, audio scripts, mind maps, flashcards, etc., are implemented to adapt the
content effectively for the detected categories of learners. The spaCy NLP-based named entity recognition
(NER) model demonstrates a 94.16% F1 score and 0.92 exact match ratio for color coding text generation of
ten electronic documents comprising 790 distinct individual words. These modifications aim to cater to the
unique preferences of learners, fostering a more personalized and engaging educational experience. To the best
of our knowledge, this is the first time an integrated learning style detection and content modification system
has been developed in this work utilizing efficient AI techniques and a private dataset.
1. Introduction videos. Auditory learners derive benefits from listening to lectures and
engaging in discussions. Kinesthetic learners, on the other hand, thrive
Learning is a complex and multifaceted process, and comprehending in hands-on activities and movements, such as experiments and role-
individual differences in learning styles is imperative for effective edu- playing. Various studies indicate a distribution ratio among visual,
cation. The term ‘‘learning style’’ pertains to an individual’s preferred auditory, and kinesthetic learners of 65:30:5 (Rogowsky, Calhoun, &
method of processing information and acquiring knowledge (Whitman, Tallal, 2020).
2023). It involves discerning how a person learns best through visual, Our motivation for undertaking this research stems from a keen
auditory, read/write, or kinesthetic (hands-on) approaches. Learning interest in understanding individual students’ diverse adaptation capa-
style theory posits that individuals exhibit distinct learning prefer- bilities and learning outcomes in response to various content and study
ences, and understanding these preferences can significantly enhance materials. Consequently, we aim to identify each student’s learning
the efficacy of teaching and learning (Newton & Salvi, 2020). Visual style and tailor the learning materials accordingly. This customiza-
tion is envisioned to render the learning process more enjoyable,
learners favor learning through visual aids such as images, charts, and
∗ Corresponding author.
E-mail address: [email protected] (R. Khan).
https://doi.org/10.1016/j.ijcce.2024.06.002
Received 24 February 2024; Received in revised form 10 June 2024; Accepted 11 June 2024
Available online 5 July 2024
2666-3074/© 2024 The Authors. Publishing services by Elsevier B.V. on behalf of KeAi Communications Co. Ltd. This is an open access article under the CC
BY-NC-ND license (http://creativecommons.org/licenses/by-nc-nd/4.0/).
M.K.H. Kanchon et al. International Journal of Cognitive Computing in Engineering 5 (2024) 269–278
enhance engagement with educational content, and inspire motivation El Alami, Oughdir, and El Allioui (2018) proposed an approach for
to acquire knowledge in alignment with their preferred learning style. detecting learning styles using machine learning algorithms to imple-
Consequently, our objective is to detect an individual’s learning style ment adaptive e-learning systems. The authors used a private dataset
and facilitate the optimization of their learning experiences, ultimately and an unsupervised method (K-means clustering) based on FSLSM. The
contributing to improved academic performance. This work aims to naive Bayes tree technique estimated a new learner’s preferred learning
develop a learning style detection and content modification system, method with 89.06% accuracy.
offering the following contributions: Kuttattu, Gokul, Prasad, Murali, and Nair (2019) applied the visual-
auditory kinesthetic (VAK) model, focusing on the simultaneous iden-
• The work introduces a novel learning style detection system tification of multiple learning styles. The authors conducted a sur-
through a dataset comprising 506 samples with 22 attributes. vey with fifteen multiple-choice questions, gathering data from 1,123
The dataset is meticulously gathered from diverse undergraduate students. Implementing the VAK model, they utilized the k-means
and graduate students at North South University, Bangladesh. algorithm to cluster the dataset, resulting in 51 clusters. Achieving
Participants engaged in a Moodle-based course, allowing com- an accuracy rate of 92.87%, the SVM algorithm was employed to
prehensive tracking and analysis of various factors related to predict learning styles. Subsequently, they integrated the predictions
learner interaction, duration of course completion, assessment using the decision tree algorithm, incorporating inputs from cluster
of diverse content types, surveys, and feedback, facilitating the values, individual learning styles, and combinations via the Rand in-
detection of their learning styles. Employing synthetic minority dex approach. Wibirama, Sidhawara, Pritalia, and Adji (2020) dis-
oversampling technique (SMOTE) data balancing and six machine cussed the strengths and limitations of three main approaches to distin-
learning (ML) algorithms — decision tree, random forest, support guishing learning styles: the conventional, artificial, and sensor-based
vector machine (SVM), logistic regression, XGBoost and blending approaches. They used web-based learning programs, object-oriented
ensemble, and convolutional neural network (CNN) deep learn- modeling lectures for Bayesian network and NBTree algorithms, an
ing. The curated dataset is classified into three categories: visual,
accelerometer, and an eye-tracker to measure different approaches to
auditory, and kinesthetic.
learning style detection. Furthermore, they introduced a theoretical
• For the content modification, we have worked on multiple modi- way to detect learning styles utilizing SVM classification algorithms.
fication processes for the learning materials, which include color Vaishnav and Chirayu (2013) investigated three prominent learning
code, audio content, mind maps, flashcards, rhymes, etc. The
styles, visual, auditory, and kinesthetic and their influence on students’
methods we used for content modification are classification
academic performance. A sample of 200 students was employed for
method, text summarization, knowledge graph, and question–
the study. Analysis of the sample reveals that the percentage of vi-
answer techniques with the help of multiple natural language
sual, auditory, and kinesthetic learners is 33.50%, 28.50%, and 38%,
processing (NLP) models and libraries such as named entity
respectively. The study also proposes distinct learning methods for
recognition (NER) with spaCy, text-to-text transfer transformer
VAK learners. Visual learners are recommended to watch videos, draw
(T5), and generative pre-trained transformer 3 (GPT-3).
maps, take notes, and use highlights. Auditory learners are advised to
• This work aims to identify learning styles and modify contents
participate in group discussions, record lectures, and listen to videos.
based on particular learning styles. Through this, we aim to
Kinesthetic learners are encouraged to study in brief sessions and attend
enhance personalized learning experiences by accurately detect-
lab classes. Thushara, Mownika, and Mangamuru (2019) conducted a
ing learning styles and modifying content to cater to individual
comparative study on keyword and keyphrase extraction algorithms
learners’ needs.
without utilizing a corpus. The authors implemented TextRank, Rapid
While learner-type detection and content modification exist sepa- Automatic Keyword Extraction (RAKE), and PositionRank algorithms
rately, the novelty of this work is to develop an integrated system for the study. While the RANK algorithm encompasses candidate and
with specialized features like mind maps, audio content, flashcards, keyword selection, TextRank and PositionRank involve graph-building,
etc., enhancing personalized learning experiences based on advanced sorting, and forming keyphrases. The authors conducted experiments
machine learning and NLP techniques. with the algorithms on research documents. The applied PositionRank
We present the related works concerning learning style classification model provided the most relevant keyphrases in less time than Tex-
and didactic style in Section 2 and discuss the proposed system in tRank and RAKE. TextRank requires more time for large documents,
Section 3, where we have presented the datasets and models used and the keyphrases extracted from RAKE tend to be lengthy and less
for this research. Section 4 is dedicated to the significant results of pertinent.
the proposed learning style detection and content modification system. Obeng (2023) implemented machine learning techniques for auto-
Finally, Section 5 concludes the article with some directions for the matic learning style classification of computer science undergraduate
future improvement of this work. students. The applied OneR technique attained approximately 80%
accuracy for classifying 198 instances and five categories of VARK
2. Related works didactic styles. Ahmad Muhammad, Qi, Wu, and Kabir Ahmad (2022)
assigned a graph representation-based embedded learning approach for
Learning style detection has long captivated the interest of re- learning style detection in an online class. The model demonstrated
searchers in the field of education, as educators endeavor to craft 88.25% accuracy and a 0.87 F1-score for the KDDCup dataset. Nguyen
personalized and engaging learning experiences for their students. The et al. (2023) demonstrated the XGBoost ensemble approach produced
traditional available approaches of detecting learning styles and con- 96% accuracy for an automatic learning style identification custom
verting content to individual preferences are implemented by various dataset of 550 volunteers.
independent tools. This section demonstrates a comprehensive review Naseer et al. (2021) discussed different NER techniques and the
of relevant studies to explore different methodologies employed by advantages and disadvantages of various libraries and compared them
recent artificial intelligence (AI)-based articles on automatic learning based on their performances. The NER techniques they discussed in-
style classification and content modification. clude rule-based, learning-based, and hybrid approaches. Among the
Rasheed and Wahid (2021) applied the Felder–Silverman learning libraries and NER tools, spaCy, StanfordNLP, TensorFlow, and Apache
style model (FSLSM) and a diverse questionnaire of 498 samples to OpenNLP are the subject of the discussion. The NER tools show that
classify learning styles. The authors used various ML algorithms, where spaCy has the highest F1-score, followed by TensorFlow, OpenNLP, and
the SVM algorithm attained the highest accuracy of 75.55% and preci- StanfordNLP. spaCy provides the best result with 100% training accu-
sion and recall of 73% and 76%, respectively. El Aissaoui, El Madani racy, F1 score, and prediction probability and with the least training
270
M.K.H. Kanchon et al. International Journal of Cognitive Computing in Engineering 5 (2024) 269–278
loss compared to all the tools. Ranganathan and Abuka (2022) proposed 3.2. Dataset preprocessing
a text summarization technique using the T5 model. To perform the
theoretical text summarization, the authors fine-tuned the T5 model. The curated dataset consists of four sections, each represented
The authors used the University of California, Irvine’s drug reviews and by an Excel file containing grades, participation, activity logs, event
BBC news datasets. The efficacy of the model was measured using the completion, and event types (ground truth). We mapped each section
ROUGE metrics with ROUGE-1 values of 45.62 and 69.05 for the UCI using the ‘‘User full name’’ column from the activity logs and concate-
and BBC News datasets, respectively. nated all sections into a single data frame. The resulting data frame
Popescu-Belis et al. (2023) implemented a novel solution to gener- comprises 506 samples and 22 columns. We have carefully examined
ate synthetic data based on a rhyming algorithm that helps learn to
the data for null values and duplicates, subsequently removing any
rhyme. The authors used the GPT-2 model and successfully obtained
rows exhibiting either. The ‘‘Time’’ column contains timestamps for
60% of rhyming lines of (AABB) consecutive rhymes. However, it
each event completion or initiation by each student. Next, this column
failed to work with alternative versus (ABAB) as it has longer-range
is converted to the ‘‘Datetime’’ datatype and chronologically sorted the
dependencies. Chen, Jia, and Xiang (2020) represented the knowledge
DataFrame based on both ‘‘User full name’’ and ‘‘Time’’. Subsequently,
graph, which works according to knowledge reasoning. Pirnau et al.
we generated a new column, ‘‘Time difference’’, by calculating the
(2024) initiated the development of NLP-based automatic techniques
time difference between consecutive events for each user using the
for content analysis in scientific research articles. The authors used over
1,000 COVID-19-related articles to classify its contents into three cate- diff() method with user grouping. The initial time difference is set
gories. The Bidirectional Encoder Representations from Transformers to 0. The label encoding approach (Talib, Albahri, & Edoh, 2023) is
(BERT) model demonstrated the most efficient results with a cosine applied to convert categorical columns into numerical representations.
similarity of 94%. In a similar article, Tunca, Sezen, and Balcioglu We assessed the data balance and identified the presence of outliers. We
(2023) applied various NLP approaches to investigate the content of applied the SMOTE technique to achieve balanced data representation.
newspaper articles. Among the applied models, the BERT technique Finally, the preprocessed dataset has been split into training and testing
obtained the maximum F1 score of 73%. sets using an 8:2 ratio. Hyperparameters optimization (Tasin, Nabil,
Significant work on learning style detection and didactic content Islam, & Khan, 2023) of the applied models is performed on the training
modification has been completed individually in various articles. How- samples. This comprehensive data preprocessing pipeline ensures our
ever, existing solutions relying on features such as user surveys and data is well-suited for subsequent machine learning analysis.
eye movements to identify learning styles have proven ineffective and
are often constrained in their implementation settings. In contrast, 3.2.1. Preprocessing techniques for content modification
the proposed solution presented herein is designed to seamlessly inte- To prepare the raw text data of the electronic document for further
grate with standard learning management systems (LMSs), eliminating analysis, we employed a comprehensive preprocessing pipeline, i.e., to-
the need for additional devices or dedicated time for learning be- kenization, segmenting the text into individual units, lemmatization
havior detection. While some content converters exist independently, to normalize words to their base forms, and grouping words with
centralized systems are limited to specific models or conversions. Con-
similar meanings. Additionally, the texts are converted to lowercase
sequently, this work introduces a comprehensive and generalized learn-
for consistency, and stop words, which hold minimal semantic value,
ing style classification and content modification system to address these
are removed. Finally, we eliminated extraneous symbols and condensed
limitations.
multiple spaces into single spaces using regular expressions within a
Python library, effectively cleaning and standardizing the text data for
3. Proposed system
downstream processing.
The proposed AI-based learning style detection and content modifier
system is divided into two major parts, as shown in Fig. 1. The proposed 3.3. Learning style detection
system identifies a user’s learning style from their LMS activity and
modifies content accordingly. A custom dataset has been created to To identify a student’s learning style, we input the student’s data
train various AI models for learning style detection. Different NLP from the LMS and utilize the corresponding parameters to the identify-
techniques were used to modify the content based on the detected ing methods. In this method, we identify the student’s style by his/her
learning style. The system can convert text content into audio and activities in the Moodle Learning Management System, such as average
visual formats to cater to different learning preferences. The descrip- video-watching time, audio listing time, quiz scores, lesson completion
tions of individual components of the proposed system are presented time, etc. Given the absence of a suitable existing dataset for activity
in subsequent paragraphs. classification, a customized dataset has been constructed in this work.
The actual label (ground truth) of a student’s learning style has been
3.1. Dataset
determined from the corresponding student’s comments and feedback
in text form as input.
In this work, a customized dataset comprising 506 samples and 22
features has been created by introducing a course to bachelor’s and
master’s students of North South University, Bangladesh, utilizing the 3.3.1. Machine learning models
Moodle LMS platform. This course included videos, audio, and text, and Various machine learning models, decision tree, random forest,
students engaged with the content by taking quizzes, participating in SVM, logistic regression, and XGBoost techniques have been used for
a survey, and providing feedback. Moodle allowed us to track time, learning approach classification of the customized dataset.
assess marks, and give feedback. Treating this information as data, we
discern the learner types. Students aged 19 to 27 years participated • Decision Tree (DT): The DT uses a flowchart-like tree structure to
in this study, with the gender distribution being 57.91% male and learn simple decision rules and predict outcomes. It begins with a
42.09% female. Fig. 2 summarizes the demographic distribution of root node, splits into branches based on the decision rules, travels
the custom dataset. The dataset was collected from July to October through child nodes, and ends with the leaf nodes that denote
2023. We considered students’ survey responses the ground truth for the result of the algorithm (Lee, Lee, Mun, & Kim, 2022). The
identifying their learning styles. From the dataset, we found that 65% uncertainty in the dataset decides whether a feature should be the
of the students who took the course were visual learners, 30% were root node, child node, or leaf node. Although DT is a supervised
auditory learners, and 5% were kinesthetic learners, as shown in Fig. 3. algorithm, it can also solve regression problems.
271
M.K.H. Kanchon et al. International Journal of Cognitive Computing in Engineering 5 (2024) 269–278
Fig. 1. Working sequences of the proposed learning style detection and content modifier system.
272
M.K.H. Kanchon et al. International Journal of Cognitive Computing in Engineering 5 (2024) 269–278
Fig. 4. Working sequences of the proposed blending ensemble model for learning style detection.
Fig. 5. Architecture of the proposed CNN technique for learning style detection.
meta-model were considered as the final predictions. Working 3.4. Content modifier
sequences of the proposed blending ensemble model for learning
style detection are illustrated in Fig. 4 and Algorithm 1. The text of an electronic document (PDF, text, Word document,
etc.) is considered learning content and converted into various learning
Algorithm 1 Algorithm of the proposed blending ensemble model for contents for three different learning styles.
learning style detection
• Visual: Three contents can be created for visual learners.
Start
Step 1: Split the custom dataset into training, validation, and test. – Color Code: Combining the Named Entity Recognition and
Step 2: Train the random forest, SVM, and decision tree base models Key Phase Extraction techniques, we are identifying the
with the training samples. words, which are persons’ names, locations, year, time, etc.
Step 3: Generate the new dataset to train the XGBoost meta-model. These words will be colored according to the type so that a
Step 4: Train the meta-model with the validation samples. virtual learner can easily notice the necessary information.
Step 5: Inference using the test samples with the trained meta-model. – Mind Map: Mind maps are helpful to visual learners to
End understand a topic better. We use Key Phase Extraction and
Knowledge Graph techniques to extract the key points and
• Convolutional Neural Network (CNN): CNN is an integral com- the relations between the elements.
ponent of artificial neural networks and a widely utilized deep – Flashcards: Using Knowledge Graph, Text-to-image, and
Question-answering techniques, we enlist the most essential
learning algorithm. It consists of layers that use convolving filters
information and represent that. It helps visual learners to
to extract local features. This study employs two one-dimensional
memorize better.
convolutional layers with a kernel size of three and 32 filters.
Furthermore, a dropout value of 0.2 was set, and a MaxPooling1D • Auditory: Three contents can be created for auditory learners.
layer was added. The activation function utilized in the convo-
– Audio Script: Auditory learners grasp the study matters
lutional part was ‘‘ReLU’’, while the final dense layer employed
best by listening, so we are creating a script for the audio
the softmax function. KerasTuner framework has been used to
lesson. The lesson is designed with some questions and men-
select the values of various hyperparameters of the applied CNN tions what to expect in this lesson at the beginning. After
model. The proposed CNN technique’s architecture employed in that, the lesson will continue following the primary learn-
this work is demonstrated in Fig. 5. ing content. Text summarization and question-generation
• Synthetic Minority Oversampling Technique (SMOTE): SMOTE is techniques have been applied to construct the audio scripts.
a popular machine learning preprocessing approach for unbal- – Summary: After an audio lesson, an overview of the lesson
anced data (Solayman, Aumi, Mery, Mubassir, & Khan, 2023). is best for this learner to take out the key points. We use the
To adjust imbalances in datasets, it creates synthetic samples text summarization technique to get an abstract summary of
focusing on the minority class. the learning content.
273
M.K.H. Kanchon et al. International Journal of Cognitive Computing in Engineering 5 (2024) 269–278
The text contents can be converted into audio content using the
text-to-speech technique, as audio is the best way for auditory
learners to learn.
• Kinesthetic: For auditory learners, efficient content can be devel-
oped.
dent according to the detected learning style. Random SVM Decision XGBoost None False
Forest Tree (Blending) (0)
3.4.1. Applied NLP techniques
We used various techniques for different types of modification of
the contents. A brief discussion of these methods is stated below.
4.1. Learning style detection
• SpaCy: SpaCy is a library for NLP built to extract information (Hu,
Gong, & He, 2022). The text is initially tokenized using spaCy, The learning styles within the custom dataset have been classified
which then undergoes a processing pipeline that includes an using various machine learning techniques. Due to the kinesthetic
entity recognizer, lemmatizer, tagger, and parser. category containing fewer instances compared to visual and auditory
• Named Entity Recognition: Named Entity Recognition (NER), an learners, the SMOTE data balancing approach has been employed.
element of natural language processing (NLP), also known as Optimization of different hyperparameters of the applied models has
entity chunking or entity extraction, recognizes preset categories resulted in improved performance.
of objects inside a text (Jehangir, Radhakrishnan, & Agarwal, Initially, the visual (label: 2), auditory (label: 0) and kinesthetic
2023). In this work, the NER model of the spaCy library has been (label: 1) categories constitute 122, 20, and 263 training instances,
used. respectively. Following the application of the SMOTE method, each
• Key Phase Extraction: In natural language processing (NLP), category comprises 300 samples. Fig. 6 shows the imbalanced and
keyphrase or keyword extraction is a text analysis approach that balanced number of samples concerning various classes of the training
pulls keywords and phrases out of the input text. Numerous
dataset.
activities can benefit from using these key terms, including con-
Table 1 presents the hyperparameters used in the proposed blending
tent categorization, document summarization, and information
ensemble technique, where Random Forest, SVM, and Decision Tree
retrieval.
• Knowledge Graph: A Knowledge Graph is a knowledge base that are employed as base estimators, and XGBoost is used as the meta
employs a graph-structured data model (Kejriwal, 2022). It is estimator. The blending process does not utilize cross-validation, and
a type of network graph showing qualitative connections be- the pass-through parameter is set to false, indicating that the original
tween facts, concepts, and events of real-world entities. The mind features are not passed to the meta-estimator.
map for the visual learner has been generated employing this Table 2 describes various metrics of the applied models for learning
approach. style classification. According to this table, the blending ensemble
• Text-to-Text Transfer Transformer (T5): T5 restructures different model with its optimized hyperparameters achieves the maximum ac-
NLP actions into a single text-to-text format, with text strings as curacy of 97.56%, with the highest number of correct predictions out of
both the input and the output (Raffel et al., 2023). This tech- the total predictions. The blending model’s F1 score of 96.76% indicates
nique considers the model text as input and trains it to generate that it maintains an outstanding trade-off between correctly identifying
target text. Translation, question-answering, summarization, and positive instances and overall correct predictions across various classes
categorization are comprehensively produced using a text-to-text within the dataset.
methodology. In this study, this transformer has been employed The total training time in seconds of the applied models is illustrated
for question–answer generation and text summarization. in Table 3. According to this table, the CNN model comprises the
• Generative Pre-trained Transformer 3 (GPT-3): To carry out a longest training time of 600.37 s, followed by the blending ensemble
range of natural language activities, GPT-3 processes text input. technique. The decision tree and SVM techniques have significantly
It combines natural language processing with natural language shorter training times of 70.99 s and 79.41 s, respectively, indicating
generation to comprehend and produce text written in real human faster training processes than more complex models like CNN and
language (Brown et al., 2020). For generating automated text of
blending ensemble.
rhyme for auditory learners, we used a GPT-3 transformer model
Fig. 7 shows that both training and test accuracy rapidly increase
in this study.
and stabilize around 96.8% after a few epochs for the applied CNN
4. Results and discussion deep learning model. The training and test loss sharply decrease and
converge, indicating effective learning and minimal overfitting. It is
The performance of the proposed AI-based automatic learning ap- worth mentioning that, different hyperparameters of the proposed CNN
proach detection and content modifier system has been measured uti- model are optimized utilizing the KerasTuner framework, leading to
lizing various metrics, which have been discussed in this section. improved performance.
274
M.K.H. Kanchon et al. International Journal of Cognitive Computing in Engineering 5 (2024) 269–278
Table 2
Performance metrics of the applied models for learning style classification.
Model Accuracy F1 score Precision Recall/Sensitivity Specificity AUC-ROC
Decision Tree 92.63% 91.14% 91.26% 91.02% 91.84% 0.91
Random Forest 96.34% 94.78% 94.92% 94.65% 94.95% 0.94
SVM 95.57% 94.07% 94.13% 94.01% 94.79% 0.94
Logistic Regression 91.48% 89.71% 89.84% 89.58% 90.74% 0.90
CNN 96.82% 96.13% 96.24% 96.03% 96.45% 0.96
XGBoost 97.03% 96.39% 96.63% 96.17% 96.66% 0.96
Blending 97.56% 96.76% 96.94% 96.59% 96.41% 0.96
275
M.K.H. Kanchon et al. International Journal of Cognitive Computing in Engineering 5 (2024) 269–278
Fig. 9. Converted content for visual learners employing colored texts, flashcards, and mindmap.
Table 4
Hyperparameters used in tokenizer.
Hyperparameters/Tasks Max length Pad to max length Truncation Return tensor
Summarization 512 False True ‘pt’
Short Question Generation 512 False True ‘pt’
Short Answer Generation X False False ‘pt’
Broad Question Generation 512 True True ‘pt’
Rhyme Generation X False False ‘pt’
Table 5
Hyperparameters used in output generation.
Hyperparameters/ Num beams Early Num return No repeat Min Max Output Return dict
Task stopping sequences NGram size length length scores in generate
Summarization 3 False 3 2 75 200 True True
Short Question Generation 3 False 3 3 20 50 True True
Short Answer Generation 1 False 1 0 10 X False False
Broad Question Generation 10 False 3 2 10 40 True True
Rhyme Generation 5 True 1 2 10 3000 False False
276
M.K.H. Kanchon et al. International Journal of Cognitive Computing in Engineering 5 (2024) 269–278
Table 7
Comparison of the proposed learning style detection and content modification system with other works.
Ref. Dataset Learning style detection Content Accuracy F1 score
modification
Rasheed and Wahid (2021) Custom Decision tree N/A 92% 92%
Samples: 498
Kuttattu et al. (2019) Custom SVM N/A 92.87% N/A
Samples: 1,123
Lokare and Jadhav (2024) Custom EEG Random forest N/A 87.5% 86.8%
Samples: 52
Obeng (2023) Custom OneR N/A 80% 87.5%
Samples: 198
Ahmad Muhammad et al. (2022) Custom Graph representation N/A 88.25% 78.50%
Samples: 198 learning
Nguyen et al. (2023) KDDCup XGBoost N/A 96% 96%
Public
Proposed Custom Blending ensemble with GPT-3, T5, 97.56% 96.76%
Samples: 506 XGBoost meta learner spaCy, etc.
Fig. 10. Converted content for (a) auditory and (b) kinesthetic learners.
Fig. 12. User feedback ratings of the proposed content modification system.
277
M.K.H. Kanchon et al. International Journal of Cognitive Computing in Engineering 5 (2024) 269–278
278