Swarupananda Bissoyi

Work place: Department of Computer Application, Maharaja Sriram Chandra Bhanja Deo University, Baripada, 757003, India

E-mail: swarupananda.bissoyi@odisha.gov.in

Website:

Research Interests: Data Mining

Biography

Swarupananda Bissoyi received his Ph.D. degree in Computer Science from the Department of Computer Science, Berhampur University, Berhampur, India, in 2022. He has been working as a full-time Assistant Professor at Maharaja Sriram Chandra Bhanja Deo University, Baripada, India, since 2014. Before entering academia, he gained 11 years of experience in the software industry, working for companies such as Mahindra Comviva and Samsung Research India – Bangalore. His research interests include Data Mining, Recommender Systems, Digital Humanities, Natural Language Processing, and Computer Vision.

Author Articles
Self-adaptive Resource Allocation in Fog-Cloud Systems Using Multi-agent Deep Reinforcement Learning with Meta-learning

By Tapas K. Das Santosh K. Das Swarupananda Bissoyi Deepak K. Patel

DOI: https://doi.org/10.5815/ijisa.2026.01.08, Pub. Date: 8 Feb. 2026

The rapid growth of IoT ecosystems has intensified the complexity of fog–cloud infrastructures, necessitating adaptive and energy-efficient task offloading strategies. This paper proposes MADRL-MAML, a Multi-Agent Deep Reinforcement Learning framework enhanced with Model-Agnostic Meta-Learning for dynamic fog–cloud resource allocation. The approach integrates curriculum learning, centralized attention-based critics, and KL-divergence regularization to ensure stable convergence and rapid adaptation to unseen workloads. A unified cost-based reward formulation is used, where less negative values indicate better joint optimization of energy, latency, and utilization. MADRL-MAML is benchmarked against six baselines Greedy, Random, Round-Robin, PPO, Federated PPO, and Meta-RL using consistent energy, latency, utilization, and reward metrics. Across these baselines, performance remains similar: energy (3.64–3.71 J), latency (85.4–86.7 ms), and utilization (0.51–0.54). MADRL-MAML achieves substantially better results with a reward of $-21.92 \pm 3.88$, energy 1.16 J, latency 12.80 ms, and utilization 0.39, corresponding to 68\% lower energy and 85\% lower latency than Round-Robin. For unseen workloads characterized by new task sizes, arrival rates, and node heterogeneity, the meta-learned variant (MADRL-MAML-Unseen) achieves a reward of $-6.50 \pm 3.98$, energy 1.14 J, latency 12.76 ms, and utilization 0.73, demonstrating strong zero-shot generalization. Experiments were conducted in a realistic simulated environment with 10 fog and 2 cloud nodes, heterogeneous compute capacities, and Poisson task arrivals. Inference latency remains below 5 ms, confirming real-time applicability. Overall, MADRL-MAML provides a scalable and adaptive solution for energy-efficient and latency-aware orchestration in fog–cloud systems.

[...] Read more.
Other Articles