Open Access
Subscription Access
Emotion Recognition Through Deep Learning in Various Modes
Subscribe/Renew Journal
The man-machine interface encompasses a crucial area—emotion recognition through facial expressions. Despite its significance, emotion recognition faces challenges such as facial accessories, nonuniform illuminations, pose variations, audio speeches, text conversations, and hand and facial gestures. Understanding emotions like happiness, anger, anxiety, joy, and shock, along with their varying degrees and overlaps, is essential for accurate recognition. These nuances, inherent to humans, pose difficulties and costs in achieving standard results through facial recognition. Recognizing someone’s mood through facial expression, conversation, voice modulation, and gestures is a skill humans excel at. However, replicating this ability through facial recognition has proven challenging and costly. This paper addresses these challenges by proposing diverse approaches to emotion detection. By exploring various modes, including facial expressions, conversation analysis, voice modulation, and gestures, the paper tackles current research problems and holds practical applications in public experiments and exhaustive sentiment analysis. The paper presents a good combo of various modes of emotion recognition on multiple datasets (tried and tested widely before amalgamating all to produce an excellent optimal result as an output of the model).
Keywords
FER, ASR, MFCC, Multimodal Deep Learning
Subscription
Login to verify subscription
User
Font Size
Information
Abstract Views: 101