Kamal Kant Verma

Work place: School of Computer Science and Engineering, IILM University Greater Noida

E-mail: dr.kamalverma83@gmail.com

Website:

Research Interests:

Biography

Kamal Kant Verma is working as Professor in School of Computer Science and Engineering at IILM University Greater Noida India. He did PhD in Computer Science and Engineering from Uttarakhand Technical University Dehradun India in 2021. He did B. Tech in Information Technology in 2006, MTech in CSE in 2012. He has 18+ years of teaching and research experience. His research area is Human Activity Recognition, Human Computer Interface, Pattern Recognition and Signal Processing. He has published more than 40 research papers in reputed national/ international journal and conferences such as Springer, Elsevier, IJIMAI, etc.

Author Articles
Lightweight 3DCNN-BiLSTM Model for Human Activity Recognition using Fusion of RGBD Video Sequences

By Vijay Singh Rana Ankush Joshi Kamal Kant Verma

DOI: https://doi.org/10.5815/ijitcs.2025.06.10, Pub. Date: 8 Dec. 2025

Over the past two decades, the automatic recognition of human activities has been a prominent research field. This task becomes more challenging when dealing with multiple modalities, different activities, and various scenarios. Therefore, this paper addresses activity recognition task by fusion of two modalities such as RGB and depth maps. To achieve this, two distinct lightweight 3D Convolutional Neural Network (3DCNN) are employed to extract space time features from both RGB and depth sequences separately. Subsequently, a Bidirectional LSTM (Bi-LSTM) network is trained using the extracted spatial temporal features, generating activity score corresponding to each sequence in both RGB and depth maps. Then, a decision level fusion is applied to combine the score obtained in the previous step. The novelty of our proposed work is to introduce a lightweight 3DCNN feature extractor, designed to capture both spatial and temporal features form the RGBD video sequences. This improves overall efficiency while simultaneously reducing the computational complexity. Finally, the activities are recognized based the fusion scores. To assess the overall efficiency of our proposed lightweight-3DCNN and BiLSTM method, it is validated on the 3D benchmark dataset UTKinectAction3D, achieving an accuracy of 96.72%. The experimental findings confirm the effectiveness of the proposed representation over existing methods.

[...] Read more.
Other Articles