UGC Approved Journal no 63975(19)
New UGC Peer-Reviewed Rules

ISSN: 2349-5162 | ESTD Year : 2014
Volume 13 | Issue 1 | January 2026

JETIREXPLORE- Search Thousands of research papers



WhatsApp Contact
Click Here

Published in:

Volume 13 Issue 1
January-2026
eISSN: 2349-5162

UGC and ISSN approved 7.95 impact factor UGC Approved Journal no 63975

7.95 impact factor calculated by Google scholar

Unique Identifier

Published Paper ID:
JETIRHG06002


Registration ID:
573656

Page Number

20-27

Share This Article


Jetir RMS

Title

Emotion Recognition Using Multimodal AI

Abstract

Emotion recognition plays a significant role in human–computer interaction, mental health analysis, social robotics, and surveillance systems. Traditional emotion recognition methods primarily rely on single modalities such as facial expressions or speech. However, single-modality systems suffer from noise, occlusion, accent variations, and environmental disturbances. To overcome these challenges, this research proposes a Multimodal AI-based Emotion Recognition System that integrates facial expression analysis, speech features, and text sentiment to enhance accuracy and reliability. The model uses a combination of Convolutional Neural Networks (CNN) for visual features, Mel-Spectrogram + LSTM for audio analysis, and BERT for text-based sentiment extraction. The modalities are fused using a CMAF (Cross-Modal Attention Fusion) mechanism. The system is evaluated on multimodal datasets and shows improved accuracy compared to unimodal methods.

Key Words

Emotion recognition plays a significant role in human–computer interaction, mental health analysis, social robotics, and surveillance systems. Traditional emotion recognition methods primarily rely on single modalities such as facial expressions or speech. However, single-modality systems suffer from noise, occlusion, accent variations, and environmental disturbances. To overcome these challenges, this research proposes a Multimodal AI-based Emotion Recognition System that integrates facial expression analysis, speech features, and text sentiment to enhance accuracy and reliability. The model uses a combination of Convolutional Neural Networks (CNN) for visual features, Mel-Spectrogram + LSTM for audio analysis, and BERT for text-based sentiment extraction. The modalities are fused using a CMAF (Cross-Modal Attention Fusion) mechanism. The system is evaluated on multimodal datasets and shows improved accuracy compared to unimodal methods.

Cite This Article

"Emotion Recognition Using Multimodal AI", International Journal of Emerging Technologies and Innovative Research (www.jetir.org), ISSN:2349-5162, Vol.13, Issue 1, page no.20-27, January-2026, Available :http://www.jetir.org/papers/JETIRHG06002.pdf

ISSN


2349-5162 | Impact Factor 7.95 Calculate by Google Scholar

An International Scholarly Open Access Journal, Peer-Reviewed, Refereed Journal Impact Factor 7.95 Calculate by Google Scholar and Semantic Scholar | AI-Powered Research Tool, Multidisciplinary, Monthly, Multilanguage Journal Indexing in All Major Database & Metadata, Citation Generator

Cite This Article

"Emotion Recognition Using Multimodal AI", International Journal of Emerging Technologies and Innovative Research (www.jetir.org | UGC and issn Approved), ISSN:2349-5162, Vol.13, Issue 1, page no. pp20-27, January-2026, Available at : http://www.jetir.org/papers/JETIRHG06002.pdf

Publication Details

Published Paper ID: JETIRHG06002
Registration ID: 573656
Published In: Volume 13 | Issue 1 | Year January-2026
DOI (Digital Object Identifier):
Page No: 20-27
Country: -, -, India .
Area: Engineering
ISSN Number: 2349-5162
Publisher: IJ Publication


Preview This Article


Downlaod

Click here for Article Preview

Download PDF

Downloads

0004

Print This Page

Current Call For Paper

Jetir RMS