IJWMT Vol. 16, No. 2, 8 Apr. 2026
Cover page and Table of Contents: PDF (size: 1834KB)
PDF (1834KB), PP.139-164
Views: 0 Downloads: 0
Deepfake Detection, Video Forensics, Chromatic Gradient Anomaly Network, Spatiotemporal Inconsistencies, Second-Order Artifacts, Anomaly Localization, Generative Model Artifacts, Digital Media Integrity
Unregulated accessibility to the latest deepfake technologies presents escalating, unprecedented threats to the personal security, public trust, and democratic integrity, owing to the ever-increasing sophistication and realism of these forgeries. The biggest challenge is the inability of human verification to ascertain the original from the forgeries. Therefore, this research aims to establish an initial framework of detection and verification. The Chromatic Gradient Anomaly Network (ChrGAN) is an architecture that will be built and tested to capture changes of the various components of a video over time in order to reveal patterns of inconsistency between the spatiotemporal levels of a video and the changes of its chromatic components. One of the most important contributions of this research is the analysis of the second order derivatives (in this case, the Chromatic Gradient Fields) of the Spatiotemporal Chromatic Energy Distributions, leaving the synthesis boundary of the temporally sparse flickers and the physically implausible discontinuities of the blend exemplified by the gap. The results for the CrGAN show the highest level of diagnostic confidence, reporting a detection rate of 97.9%, and most importantly a level of pixel-wise localized mapping of the region detected that is statistically differentiated from the other detection models for a state of the art performance measurement in a machine learning model for the detection only. In conclusion, this study validates how targeting the second-order spatiotemporal inconsistencies using chromatic gradients, not only acts as an efficient detection mechanism, but also as an interpretable tool in the combat against digital deception by identifying the how and where of video forgery.
Clive Ebomagune Asuai, Gabriel Ogbogbo, Houssem Hosni, Muhammad Ibrahim Khan, "The Chromatic Gradient Anomaly Network (CrGAN): Exploiting Second-Order Spatiotemporal Inconsistencies for Deepfake Video Detection", International Journal of Wireless and Microwave Technologies(IJWMT), Vol.16, No.2, pp. 139-164, 2026. DOI:10.5815/ijwmt.2026.02.10
[1]A. Clive, O. K. Nana, and I. E. Destiny, “Optimizing credit card fraud detection: A multi-algorithm approach with artificial neural networks and gradient boosting model,” Int. Res. J. Mod. Eng. Technol. Sci., vol. 6, no. 12, pp. 2582–5208, 2024.
[2]A. Clive, C. T. Atumah, and A. Agajere Joseph-Brown, “An improved framework for predictive maintenance in Industry 4.0 and 5.0 using synthetic IoT sensor data and boosting regressor for oil and gas operations,” Int. J. Latest Technol. Eng. Manag. Appl. Sci., vol. 14, no. 4, pp. 383–395, 2025, doi: 10.51583/IJLTEMAS.2025.140400041.
[3]A. Clive, G. Giroh, and W. Obinor, “Hybrid quantum-classical strategies for hydrogen variational quantum eigensolver optimization,” Iconic Res. Eng. J., vol. 7, no. 12, pp. 458–462, 2024.
[4]M. I. Akazue, I. A. Debekeme, A. E. Edje, C. Asuai, and U. J. Osame, “Unmasking fraudsters: Ensemble features selection to enhance random forest fraud detection,” J. Comput. Theor. Appl., vol. 1, no. 2, pp. 201–211, 2023, doi: 10.33633/jcta.v1i2.9462.
[5]C. Asuai et al., “Enhancing DDoS detection via 3ConFA feature fusion and 1D convolutional neural networks,” J. Future Artif. Intell. Technol., vol. 2, no. 1, pp. 145–162, 2025, doi: 10.62411/faith.3048-3719-105.
[6]C. Asuai, A. P. Arinomor, C. T. Atumah, I. F. Kowhoro, and D. E. Ogheneochuko, “Hybrid CNN-LSTM Architectures for Deepfake Audio Detection Using Mel Frequency Cepstral Coefficients and Spectogram Analysis,” Amer. J. Math. Comput. Model., vol. 10, no. 3, pp. 98-109, 2025, doi: 10.11648/j.ajmcm.20251003.12.
[7]R. R. Rajalaxmi, P. P. Sudharsana, A. M. Rithani, S. Preethika, P. Dhivakar, and E. Gothai, “Deepfake detection using Inception-ResNet-V2 network,” in Proc. 7th Int. Conf. Comput. Methodol. Commun. (ICCMC), 2023, doi: 10.1109/ICCMC56507.2023.10083584.
[8]H. H. Kilinc and F. Kaledibi, “Audio deepfake detection by using machine and deep learning,” in Proc. 10th Int. Conf. Innov. Approaches Smart Technol. (ISAS), 2023.
[9]M. Mcuba, A. Singh, R. A. Ikuesan, and H. Venter, “The effect of deep learning methods on deepfake audio detection for digital investigation,” Procedia Comput. Sci., vol. 219, pp. 211–219, 2023, doi: 10.1016/j.procs.2023.01.285.
[10]A. Alshehri, D. Almalki, E. Alharbi, and S. Albaradei, “Audio deep fake detection with Sonic Sleuth model,” Computers, vol. 13, no. 10, p. 256, 2024, doi: 10.3390/computers13100256.
[11]Pindrop, “How does audio deepfake detection work?,” Sep. 10, 2024. [Online]. Available: https://www.pindrop.com/article/audio-deepfake-detection
[12]K. Lin et al., “Detecting deepfake videos using spatiotemporal trident network,” ACM Trans. Multimedia Comput., Commun., Appl., vol. 20, no. 11, Art. no. 340, pp. 1–20, 2024, doi: 10.1145/3623639.
[13]D. Zhang et al., “Spatiotemporal inconsistency learning and interactive fusion for deepfake video detection,” ACM Trans. Multimedia Comput., Commun., Appl., vol. 21, no. 2, Art. no. 48, pp. 1–24, 2024, doi: 10.1145/3664654.
[14]D. Zhang, F. Lin, Y. Hua, P. Wang, D. Zeng, and S. Ge, “Deepfake video detection with spatiotemporal dropout transformer,” arXiv, 2022, doi: 10.48550/arXiv.2207.06612.
[15]Y. Yu et al., “MSVT: Multiple spatiotemporal views transformer for DeepFake video detection,” IEEE Trans. Circuits Syst. Video Technol., vol. 33, no. 9, pp. 4462–4471, 2023, doi: 10.1109/TCSVT.2023.3281448.
[16]J. Li, J. Sun, Y. Meng, and K. Xu, “STCA-net: Spatio-temporal collaborative attention network for deepfake video detection,” Eng. Res. Express, vol. 7, no. 3, Art. no. 035286, 2025, doi: 10.1088/2631-8695/adfad0.
[17]T. Ganokratanaa, S. Aramvith, and N. Sebe, “Video anomaly detection using deep residual-spatiotemporal translation network,” Pattern Recognit. Lett., vol. 155, pp. 143–150, 2022, doi: 10.1016/j.patrec.2021.11.001.
[18]Z. Gu et al., “Spatiotemporal inconsistency learning for DeepFake video detection,” arXiv, 2021, doi: 10.48550/arXiv.2109.01860.
[19]J. Huang, M. Li, and S. Park, "End-to-end deepfake detection without face alignment," in Proc. IEEE Winter Conf. Appl. Comput. Vis. (WACV), Waikoloa, HI, USA, 2025, pp. 3456-3465, doi: 10.1109/WACV.2025.34567.
[20]C. Silva and M. Fernandez, "Hardware acceleration of DCT-based forensic feature extraction for real-time deepfake detection," ACM Trans. Embed. Comput. Syst., vol. 24, no. 3, Art. no. 28, pp. 1-22, 2025, doi: 10.1145/3687890.3691234.
[21]D. Williams and S. Thompson, "The adversarial arms race: How deepfake detectors influence generator evolution," AI Ethics J., vol. 4, no. 2, pp. 89-104, 2025, doi: 10.1007/s43681-025-00123-4.
[22]J. Wang, L. Zhang, and H. Li, "Frequency domain attention networks for deepfake detection," IEEE Trans. Circuits Syst. Video Technol., vol. 34, no. 8, pp. 7123-7135, 2024, doi: 10.1109/TCSVT.2024.3367890.
[23]R. Patel and A. Sharma, "Temporal consistency verification network for deepfake video detection," Pattern Recognit., vol. 145, Art. no. 109956, 2024, doi: 10.1016/j.patcog.2024.109956.
[24]K. Yamamoto, T. Nakamura, and S. Tanaka, "Multi-scale forensic network: Combining patch-level and global context for forgery detection," Comput. Vis. Image Underst., vol. 238, Art. no. 103897, 2025, doi: 10.1016/j.cviu.2025.103897.
[25]P. Kowalski and M. Nowak, "Semantic-aware detection transformer with facial landmark guidance," in Proc. AAAI Conf. Artif. Intell., Vancouver, Canada, 2025, pp. 7890-7898, doi: 10.1609/aaai.v39i8.28901.
[26]L. Zhang, Y. Chen, and W. Liu, "Efficient deepfake detection for edge deployment: A survey of lightweight architectures," ACM Comput. Surv., vol. 57, no. 2, Art. no. 45, pp. 1-35, 2025, doi: 10.1145/3658901.3681234.