Work place: Department of Computer Science and Information Technology, K L University, Vaddeswaram, Guntur District 522302, India
E-mail: nazeerakhamar@gmail.com
Website: https://orcid.org/0009-0000-0242-5262
Research Interests:
Biography
Shaik Nazeera Khamar is an information technology professional currently working as a Fullstack web Developer, with prior experience at Codetree. Nazeera is an alumna of KL University, which equipped her with both theoretical knowledge and practical skills relevant to the fast-evolving tech landscape. Through her career, she has demonstrated expertise with various platforms and digital transformation initiatives. Nazeera is recognized among her peers for collaborative teamwork, adaptiveness, and a rapid learning curve that enables her to solve complex business and technical challenges efficiently.
By Srinivas P. V. V. S. Shaik Nazeera Khamar Nohith Borusu Mohan Guru Raghavendra Kota Harika Vuyyuru Sampath Patchigolla
DOI: https://doi.org/10.5815/ijigsp.2026.01.07, Pub. Date: 8 Feb. 2026
In the field of affective computing research, multi-modal emotion detection has gained popularity as a way to boost recognition robustness and get around the constraints of processing a multiple type of data. Human emotions are utilized for defining a variety of methodologies, including physiological indicators, facial expressions, as well as neuroimaging tactics. Here, a novel deep attention mechanism is used for detecting multi-modal emotions. Initially, the data are collected from audio and video features. For dimensionality reduction, the audio features are extracted using Constant-Q chromagram and Mel-Frequency Cepstral Coefficients (MM-FC2). After extraction, the audio generation is carried out by a Convolutional Dense Capsule Network (Conv_DCN) is used. Next is video data; the key frame extraction is carried out using Enhanced spatial-temporal and Second-Order Gaussian kernels. Here, Second-Order Gaussian kernels are a powerful tool for extracting features from video data and converting it into a format suitable for image-based analysis. Next, for video generation, DenseNet-169 is used. At last, all the extracted features are fused, and emotions are detected using a Weighted Late Fusion Deep Attention Neural Network (WLF_DAttNN). Python tool is used for implementation, and the performance measure achieved an accuracy of 97% for RAVDESS and 96% for CREMA-D dataset.
[...] Read more.Subscribe to receive issue release notifications and newsletters from MECS Press journals