Final Draft
Final Draft
ABSTRACT- With the goal of completely changing the sessions. In traditional classroom teaching, teachers
paradigm of the teaching and learning process from evaluate their students’ learning effect, the level of
traditional classroom instruction to online learning understanding and comprehension, by mainly observing
platforms, e-learning is rapidly moving towards students’ behavior. The behavior aspects may include body
personalised learning. This change is predicated on the language, eye gaze, facial expressions, and emotions
notion that accurate emotion recognition in e-learning exhibited through vocal feedback. Analysing all these
platforms enables students' learning experiences to be attributes a tutor in a physical classroom could give
tailored to their preferences. Three modes are used by e- personalized feedback to a particular student. But in online
learners to express their sentiment: text, image, and
learning analyzing these attributes is a challenge in the
audio. Online students engage in a range of learning
absence of a physical tutor. Detection of a student’s
activities, such as writing, reading, watching tutorial
videos, taking online tests, and attending online emotional state is crucial to personalize their learning in an
conferences. They exhibit a range of engagement levels automated learning platform. Multiple researchers have
while taking part in these educational activities, proposed the use of natural language processing, hand
including neutral, confusion, delight, boredom, and gesture recognition, eye gaze estimation, facial emotion
frustration. It is critical for online educators to recognition, and body language detection to estimate
accurately and effectively assess the engagement level of learners' learning effects and provide a measure that will
their online learners in order to offer individualised provide a more effective learning experience. It is
pedagogical support through interventions. Various becoming essential for e-learning platforms to be able to
methods from the fields of machine learning and educate their learners according to their personalized
computer vision are used to detect engagement. We features.
investigate the viability of utilising machine learning on Due to the enormous development in the Machine
data from eye trackers and camera sensors to measure Learning sphere, the LMSs can now classify the learners
learner engagement and categorise the degree of according to their different personalized features like
engagement. After watching videos of the students and learning style, attentiveness, cognitive ability, specialized
their screens, we categorise the gathered data as requirements, etc. The learners belonging to different
Engaged or Not Engaged. Perceptual user features (e.g., classes are recommended different learning objects,
body posture, facial points, and gaze) are extracted from learning contents, tutors, etc. after discovering the most
the collected data. Feature selection and classification appropriate one for them by applying different Machine
techniques are used to create classifiers that can
Learning algorithms. Recognizing a learner’s emotional
determine whether or not a student is engaged. After
state (fear, anger, depression, joy, confusion, confidence,
identifying the difficulties in detecting engagement, we
investigate the datasets and performance metrics that etc.) is also very important for personalizing their learning
are currently available and offer suggestions for how to experience. Recommending appropriate learning content
improve the technology in the future to detect according to their emotional state could enhance their
engagement in online learning. learning quality as a result improving the overall quality of
the learning platform.
1. Introduction Researchers have developed several frameworks and
models predominantly based on Artificial Intelligence to
The usage of computer-based technology is growing in
recognize the emotional state of a person from textual,
numerous directions due to its easy availability,
effectiveness, etc. Technological advancements like audio, image, and video content. In this work, we have
smartphones, laptops, and other intelligent devices help us reviewed different research works for recognizing the
use online learning facilities termed E-learning. E-learning emotion of an e-learner through analyzing their video or
platforms have become a significant tool for knowledge image.
sharing and understanding for almost every student, The purpose of this survey is to encourage further study into
especially after the pandemic. E-learning has numerous ways to improve the learning drive of an e-learner. Since a
advantages like eco-friendliness as it saves paper, reduces person’s motivation is closely related to their emotion we
the cost of traveling and time, etc. The students can attend have surveyed recognizing emotions of a person from their
the classes from their places even if they are not feeling video mostly in the context of e-learning. This will direct
well. During a pandemic situation like COVID-19, all new avenues to enhance personalized learning in Intelligent
parents and students were worried about the futures of their Tutoring Systems to attract and sustain a greater number of
children in lockdown situations, but E-learning has solved students in online learning. Using the above exposition, we
this problem, and every student is capable of learning
justify the novelty of this work.
through their mobile phones and laptops.
On the other hand, the physical education system has This paper is organized in the following sections. In
benefits like interaction between teacher and students, section 2 we have conducted a brief literature review of the
assessment of student’s understandability, and hands-on topic. Section 3 represents a comparative study of the
research works discussed in section 2. In Section 4 we have system is useful because it works well even if faces are at
discussed the future direction of this research topic in brief. different angles or partly covered; making it more versatile
for real-world situations.
2. Brief Survey
In [5] authors emphasized more on multi-modal emotion
We observed as we explored through the previous works recognition than the single modal ones. They used
that frameworks developed for the emotional recognition of
a person are majorly based on two categories. Either deep Affectnet as their FER model. The use of predicted
learning-based or Conventional Machine learning-based. So emotions extends beyond understanding student behavior to
as per our findings we have divided our review in two include visual summarization of classroom films and
categories. In category 2.1 we have reviewed those works classification of the group-level emotions on videos.
which are developed around deep learning-based algorithms
and in 2.2 we have reviewed those works in which deep Based on their behavior and biological data, several
learning algorithms have hardly intervened attempts have been made to ascertain the e-learners'
level of focus through a neuro-fuzzy inference system that
tracks the position of the eye’s iris to find out the
concentration level. In it, SVM is also used to determine
concentration level [6]. This proposed model has an
overhead of timing for the preprocessing of data which in
the future could be eliminated by fully automating the
process so that it could be implemented in real time also.
1. Non-verbal behavioural pattern classification feed-forward MLP Question type, gender, and demographic NVB modelling and behaviour labelling
of e-learners & MSE variables are not considered for classification can be improved using deep learning
techniques.
2. Detecting Learning Affect in E-Learning OpenCV, HAAR, fuzzy rough theoretical methods for rule
Platform Using Facial Emotion Expression CNN2 generation and feature selection would be
developed.
3. Estimating Student Learning Affect Using CNN, FER2013 Using two predictions facial emotion for include the use of multi-model pattern
Facial Emotions dataset learning affect detection is more ideal than analysis such as body expression, eye
using one or three predictions. gaze, and head movements to achieve a
more accurate result
4. EAC-Net: Deep Nets with Enhancing and CNN, BP4D, and Automatic generation of attention maps Finding more responsive areas for
Cropping for Facial Action Unit Detection DISFA AU datasets enhancing and cropping nets rather than
manually locating the positions at present.
5. Classifying Emotions and Engagement in CNN, Affectnet Improve the quality of the engagement 1) Predict arousal and valence in addition
Online Learning Based on a Single Facial prediction engine to facial expressions 2) Face clustering
Expression Recognition Neural Network
6. A methodology to predict e-learners’ Recurrent neural 1)applied only to well-structured process 1) Automation of the process.
concentration networks, long models. 2) applied in a controlled environment. 2) Should be tested in a real environment.
short-term memory, 3) Expanding the model to CNN and
TCNs.
7. Implementation of a deep-learning-based temporal relational To learn the long-term temporal dependencies Combine the TRN technique with other
facial image analysis model to estimate the network, MLP, FER between frames deep neural networks suffer modalities such as the audio modality with
learning effect and to reflect on the level of other datasets.
student engagement.
8. Proposes a method for accurate facial Stacked sparse auto- Limited to training on CPU-based machines, Use a framework that could support GPU,
expression recognition using a lightweight encoder (SSAE) which is why it took a longer time for training. which will improve the training time.
deep learning model.
9.21. A comprehensive overview of the research on SVM, CNN, KNN, 1) High memory and sophisticated computing explore approaches for long-term
FER in online learning. DNN & LSTM requirements. 2)It has low illumination and a monitoring of facial emotions, such as
lack of frontal pose. wearable sensors or continuous video
recording.
10.2 An algorithm that can accurately detect Adaptive Weighted Many one-screen learning pages in which To deal with a short learning content
2. students' engagement in online learning Local Gray Code students need not scroll to update content. page..
environments Patterns (LGCP),
11.23 develop approaches for the automatic CNNs Limitation of short-term laboratory studies. Focusing on long-term learning situations
. recognition of student engagement from their
facial expression
12.2 To investigate student engagement levels LSTM Limited data set and computational power. They want to merge the information
4. context of online learning through the analysis currently provided by our system with the
of facial behaviour. information
13.2 To develop deep learning models for real-time CNNs This only limits the selection in the aspect of Collect more data for learning the models,
5 facial expression recognition (FER) in the the age range. and re-conduct the classroom experiment
context of active teaching with better camera positioning
14.9. Analyze online lecture videos, detect students' CNNS Camera Orientation, Obstructions and Lighting They aim to conduct a user study in the
engagement levels and emotions Conditions, Engagement Detection from future and solve the limitations.
Profile Pictures
15.26 The development of a robust and accurate ResNet18, Triplet Overfitting, network complexity, linear Multimodal approaches that combine
system for classifying human facial emotions Loss Function bottleneck structure, complex facial structures. image- and video-based methodologies,.
16.2 A systematic literature review on the use of SVM Low Illumination, Lack of Frontal Pose, 1) Long-term monitoring of facial
7 Facial Expression Recognition (FER) systems sample number of Dataset emotions.2) privacy, consent, and potential
in the classification of academic emotions. biases
17.1 Detecting engagement levels and emotions of CNNS Emotion Recognition Accuracy, Ethical and group engagement detections and evaluate
0 online learners using facial expression Privacy Concerns the valence and arousal of the group
recognition.
18.11 An autonomous monitoring system using UML, ANN, CNN 1. Developing an autonomous Facial expression recognition accuracy.
. facial expression recognition and gamification monitoring system. Generalization to different contexts
methods to support the learning process with a 2. Implementation using cloud
blended learning model. computing.
3. Testing facial expression
recognition and gamification
References extraction and visualization for engagement
detection in online learning”
1. Mike Holmes, Annabel Latham, Keeley Crockett, 15. Irfan Haider, Hyung-Jeong Yang, Guee-Sang Lee,
and James D. O’Shea “Near real-time Soo-Hyung Kim, Wataru Sato “Robust Human
comprehension classification with artificial Face Emotion Classification Using Triplet-Loss-
neural networks: decoding e-Learner non-verbal Based Deep CNN Features and SVM”
behaviour” 16. Jeniffer Xin-Ying Lek and Jason Teo, Sarah
2. Sukrit Bhattacharya, Vaibhav Shaw, Pawan [Link] “Academic Emotion Classification
Kumar Singh, Ram Sarkar, and Debotosh Using FER: A Systematic Review”
Bhattacharjee “SV-NET: A Deep Learning 17. Michael Moses Thiruthuvanathan, Balachandran
Approach to Video Based Human Activity Krishnan, Madhavi Rangaswamy “Engagement
Recognition” Detection Through Facial Emotional Recognition
3. Benisemeni Esther Zakka and Hina Vadapalli Using a Shallow Residual Convolutional Neural
“Detecting Learning Affect in E-Learning Networks”
Platform Using Facial Emotion Expression” 18. Indra Kurniawan, Yeffry Handoko Putra
4. Wei Li, Farnaz Abtahi, Zhigang Zhu, and Lijun “Autonomous Monitoring with Facial Expression
Yin “EAC-Net: Deep Nets with Enhancing and Recognition and Gamification to Support
Cropping for Facial Action Unit Detection” Blended Learning Model”
5. Andrey V. Savchenko, Lyudmila V. Savchenko, 19. Islam Alkabany, Asem Ali, Amal Farag, Ian
Ilya Makarov “Classifying Emotions and Bennett, Mohamad Ghanoum, Aly Farag
Engagement in Online Learning Based on a “Measuring Student Engagement Level Using
Single Facial Expression Recognition Neural Facial Information”
Network” 20. Guary Tonguc, Betul Ozaydin Ozkara
6. Young-Sang Jeong, Nam-Wook Cho “Evaluation “Automatic Recognition of Student Emotions
of e-learners’ concentration using recurrent neural from Facial Expressions during a Lecture”
networks” 21. Bing Gong and Jing Wei “Quantitative Analysis
7. Anil Pise, Hima Vadapalli, Ian Sanders “Facial of Facial Expression Recognition in Classroom
emotion recognition using temporal relational Teaching Based on FACS and KNN
network: an application to E-Learning” Classification Algorithm”
8. Mubashir Ahmad, Saira, Omar Alfandi, Asad 22. Unqua Laraib, Arslan Shaukat, Rizwan Ahmed
Masood Khattak, Syed Furqan Qadri, Iftikhar Khan, Zartasha Mustansar, Muhammad Usman
Ahmed Saeed, Salabat Khan, Bashir Hayat and Akram and Umer Asgher “Recognition of
Arshad Ahmad “Facial expression recognition Children’s Facial Expressions Using Deep
using lightweight deep learning modeling” Learned Features”
9. Jennifer Xin-Ying Lek and Jason Teo “Academic 23. K. Keerthana, D. Pradeep Dr. B. Vanathi
Emotion Classification Using FER: A Systematic “Learner’s Engagement Analysis for E-Learning
Review” Platform”
10. Zhaoli Zhang, Zhenhua Li, Hai Liu, Taihe Cao, 24. Ati Jain, Dr. Hare Ram Sah and Harsha Atre
and sannyuya Liu “Data-driven Online Learning “Student’s Emotion Recognition through Facial
Engagement and Mouse Behavior Recognition Expressions during E-Learning using Fuzzy
Technology” Logic and CNN classification”
11. Jacob Whitehill, Zewelanji Serpell, Yi-Ching Lin, 25. M. Ali Akber Dewan, Mahbub Murshed and
Aysha Foster, and Javier R. Movellan “The Faces Fuhua Lin “Engagement detection in online
of Engagement: Automatic Recognition of learning: a review”
Student Engagement from Facial Expressions” 26. Sana Ikram, Haseeb Ahmad, Nasir Mahmood,
12. Unqua Laraib, Arslan Shaukat, Rizwan Ahmed C.M. Nadeem Faisal, Qaisar Abbas, Imran
Khan, Zartasha Mustansar, Muhammad Usman Qureshi and Ayyaz Hussain “Recognition of
Akram, and Umer Asgher “Recognition of Student Engagement State in a Classroom
Children’s Facial Expressions Using Deep Environment Using Deep and Efficient Transfer
Learned Features” Learning Algorithm”
13. David Dukic, Ana sovic Krzic “Real-Time Facial 27. Fezile Ozdamli, Aayat Alijarrah, Damla
Expression Recognition Usinig Deep Learnin Karagozlu, and Mustafa Ababneh “Facial
with Application in the Active Classroom Recognition System to Detect Student Emotions
Environment ” and Cheating in Distance Learning”
14. Mohammad Nehal Hasnine, Huyen [Link], 28. Yi Chen, Jin Zhou, Qiating Gao, Jing Gao, and
Thuy Thi Thu Tran, Ho Tran Nguyen, Gokhan Wei Zhang “MDNN: Predicting Student
Akcapinar, Hiroshi Ueda “Students emotion Enagegement via Gaze Direction and Facial
Expression in Collaborative Learning”
29. Nha Tran, Hung Nguyen, Hien Luong, Minh
Nguyen, Khiet Luong, Huy Tran “Recognition of
Student Behaviour through Actions in the
Classroom”
30. Benyoussef Abdellaoui, Aniss MOUMEN,
Younes ELBOUZEKRI EL IDRISSI, Ahmed
Remaida “Face Detection to Recognize Students
Emotion and Their Engagement: A Systematic
Review”