Work place: Department of Computer Science and Information Technology, Institute of Technical Education and Research (ITER), Siksha 'O' Anusandhan (SOA) Deemed to be University, Bhubaneswar, 751030, India
E-mail: deepakpatel@soa.ac.in
Website:
Research Interests: Cloud Computing
Biography
Deepak K. Patel is currently working as an Associate Professor at Siksha ’O’ Anusandhan Deemed to be University, Bhubaneswar, Odisha, India. He received a Ph.D degree in Computer Science & Engineering from Veer Surendra Sai University of Technology (VSSUT), Burla, Odisha, India, in 2019. He has received a B.Tech degree in computer science & engineering from Biju Pattanaik University of Technology, Odisha, India, and an M.Tech degree in computer science & engineering from Veer Surendra Sai University of Technology (Government of Odisha, India), in 2011 and 2014, respectively. His current research focuses on Grid Computing and Cloud Computing.
By Tapas K. Das Santosh K. Das Swarupananda Bissoyi Deepak K. Patel
DOI: https://doi.org/10.5815/ijisa.2026.01.08, Pub. Date: 8 Feb. 2026
The rapid growth of IoT ecosystems has intensified the complexity of fog–cloud infrastructures, necessitating adaptive and energy-efficient task offloading strategies. This paper proposes MADRL-MAML, a Multi-Agent Deep Reinforcement Learning framework enhanced with Model-Agnostic Meta-Learning for dynamic fog–cloud resource allocation. The approach integrates curriculum learning, centralized attention-based critics, and KL-divergence regularization to ensure stable convergence and rapid adaptation to unseen workloads. A unified cost-based reward formulation is used, where less negative values indicate better joint optimization of energy, latency, and utilization. MADRL-MAML is benchmarked against six baselines Greedy, Random, Round-Robin, PPO, Federated PPO, and Meta-RL using consistent energy, latency, utilization, and reward metrics. Across these baselines, performance remains similar: energy (3.64–3.71 J), latency (85.4–86.7 ms), and utilization (0.51–0.54). MADRL-MAML achieves substantially better results with a reward of $-21.92 \pm 3.88$, energy 1.16 J, latency 12.80 ms, and utilization 0.39, corresponding to 68\% lower energy and 85\% lower latency than Round-Robin. For unseen workloads characterized by new task sizes, arrival rates, and node heterogeneity, the meta-learned variant (MADRL-MAML-Unseen) achieves a reward of $-6.50 \pm 3.98$, energy 1.14 J, latency 12.76 ms, and utilization 0.73, demonstrating strong zero-shot generalization. Experiments were conducted in a realistic simulated environment with 10 fog and 2 cloud nodes, heterogeneous compute capacities, and Poisson task arrivals. Inference latency remains below 5 ms, confirming real-time applicability. Overall, MADRL-MAML provides a scalable and adaptive solution for energy-efficient and latency-aware orchestration in fog–cloud systems.
[...] Read more.Subscribe to receive issue release notifications and newsletters from MECS Press journals