Vineela Krishna. Suri

Work place: Department of CSE, Jawaharlal Nehru Technological University, Kakinada, Andhra Pradesh, India

E-mail: vineela.suri@gmail.com

Website: https://orcid.org/0000-0002-4563-1979

Research Interests:

Biography

Vineela Krishna. Suri is a Research Scholar in the Department of Computer Science and Engineering at Jawaharlal Nehru Technological University, Kakinada. She holds an M.Tech in Computer Science and Engineering from JNTUK, Kakinada. Her areas of interest include Machine Learning, Deep Learning, and Web Programming. She has contributed to various research publications in reputed journals and conferences. Her research focuses on developing innovative solutions using artificial intelligence and data-driven technologies. Additionally, she is passionate about exploring emerging trends in web development and AI-driven applications.

Author Articles
A Hybrid CNN-Transformer Model for Multimodal Fake News Detection Using Feature Fusion

By Vineela Krishna. Suri Prasad. GVSNRV

DOI: https://doi.org/10.5815/ijmecs.2026.02.08, Pub. Date: 8 Apr. 2026

The widespread distribution of fake news poses a critical societal challenge by influencing public opinion and shaping political discourse. Addressing this problem requires models that can capture multimodal cues beyond text alone. This work proposes a lightweight Multimodal Cross-attention Fusion–based Fake News Detection (MCAF-FND) model which combines textual and visual features through cross-attention strategy. The study evaluates MCAF-FND on the Fakeddit benchmark, a large-scale dataset comprising 682,996 multimodal samples collected from social media. Textual features are extracted using DistilBERT, while spatially aware image representations are derived from VGG-19 convolutional layers. The cross-attention module enables semantic alignment between text tokens and image patches, modeling inter-modal dependencies more effectively than conventional fusion strategies. The fused representation is classified using a Multilayer Perceptron(MLP) with softmax, ensuring contributions from both modalities. Experimental results demonstrate that MCAF-FND consistently outperforms unimodal baselines and traditional fusion methods, achieving 93.2% accuracy with strong precision, recall, and F1-score. Cross-attention based visualizations illustrate how the model aligns textual cues with salient visual regions, enhancing interpretability. By combining computational efficiency with robust multimodal reasoning, the proposed approach provides a reliable and extensible solution for automated fake news detection.

[...] Read more.
Other Articles