Open Access Open Access  Restricted Access Subscription Access

Emotion Recognition Through Deep Learning in Various Modes


Affiliations
1 Ministry of Electronics and Information Technology, Government of India, India
2 Department of Engineering Physics, Indian Institute of Technology Banaras Hindu University, Varanasi, India

   Subscribe/Renew Journal


The man-machine interface encompasses a crucial area—emotion recognition through facial expressions. Despite its significance, emotion recognition faces challenges such as facial accessories, nonuniform illuminations, pose variations, audio speeches, text conversations, and hand and facial gestures. Understanding emotions like happiness, anger, anxiety, joy, and shock, along with their varying degrees and overlaps, is essential for accurate recognition. These nuances, inherent to humans, pose difficulties and costs in achieving standard results through facial recognition. Recognizing someone’s mood through facial expression, conversation, voice modulation, and gestures is a skill humans excel at. However, replicating this ability through facial recognition has proven challenging and costly. This paper addresses these challenges by proposing diverse approaches to emotion detection. By exploring various modes, including facial expressions, conversation analysis, voice modulation, and gestures, the paper tackles current research problems and holds practical applications in public experiments and exhaustive sentiment analysis. The paper presents a good combo of various modes of emotion recognition on multiple datasets (tried and tested widely before amalgamating all to produce an excellent optimal result as an output of the model).

Keywords

FER, ASR, MFCC, Multimodal Deep Learning
Subscription Login to verify subscription
User
Notifications
Font Size

Abstract Views: 101




  • Emotion Recognition Through Deep Learning in Various Modes

Abstract Views: 101  | 

Authors

Bharat Gupta
Ministry of Electronics and Information Technology, Government of India, India
Manas Gupta
Department of Engineering Physics, Indian Institute of Technology Banaras Hindu University, Varanasi, India

Abstract


The man-machine interface encompasses a crucial area—emotion recognition through facial expressions. Despite its significance, emotion recognition faces challenges such as facial accessories, nonuniform illuminations, pose variations, audio speeches, text conversations, and hand and facial gestures. Understanding emotions like happiness, anger, anxiety, joy, and shock, along with their varying degrees and overlaps, is essential for accurate recognition. These nuances, inherent to humans, pose difficulties and costs in achieving standard results through facial recognition. Recognizing someone’s mood through facial expression, conversation, voice modulation, and gestures is a skill humans excel at. However, replicating this ability through facial recognition has proven challenging and costly. This paper addresses these challenges by proposing diverse approaches to emotion detection. By exploring various modes, including facial expressions, conversation analysis, voice modulation, and gestures, the paper tackles current research problems and holds practical applications in public experiments and exhaustive sentiment analysis. The paper presents a good combo of various modes of emotion recognition on multiple datasets (tried and tested widely before amalgamating all to produce an excellent optimal result as an output of the model).

Keywords


FER, ASR, MFCC, Multimodal Deep Learning