Asynchronous Multi-Agent Reinforcement Learning for Scheduling in Subnetworks
Loading...
Access rights
openAccess
acceptedVersion
URL
Journal Title
Journal ISSN
Volume Title
A4 Artikkeli konferenssijulkaisussa
This publication is imported from Aalto University research portal.
View publication in the Research portal (opens in new window)
View/Open full text file from the Research portal (opens in new window)
View publication in the Research portal (opens in new window)
View/Open full text file from the Research portal (opens in new window)
Unless otherwise stated, all rights belong to the author. You may download, display and print this publication for Your own personal use. Commercial use is prohibited.
Date
Major/Subject
Mcode
Degree programme
Language
en
Pages
6
Series
Proceedings of the IEEE 101st Vehicular Technology Conference, IEEE Vehicular Technology Conference
Abstract
We address radio resource scheduling in a network of multiple in-X subnetworks providing wireless Ultra-Reliable Low- Latency Communication (URLLC) service. Each subnetwork is controlled by an agent responsible for scheduling resources to its devices. Agents rely solely on interference measurements for information about other agents, with no explicit coordination. Subnetwork mobility and fast-fading effects create a non-stationary environment, adding to the complexity of the scheduling problem. This scenario is modeled as a multi-agent Markov Decision Process (MDP). To address the problem, we propose a Multi- Agent Deep Reinforcement Learning (MADRL) approach under URLLC constraints, which integrates Long Short-Term Memory (LSTM) with the Deep Deterministic Policy Gradient (DDPG) algorithm to manage non-stationarity and high-dimensional action spaces. We apply an asynchronous update strategy, where one agent is updating at a time. This reduces learning variability, resolves policy conflicts, and improves the interpretability of the MADRL approach. Simulation results demonstrate that the asynchronous update mechanism outperforms synchronous updates and baseline methods, achieving superior reliability, resource utilization, and explainability.Description
Other note
Citation
Srinivasan, A, Zhang, J & Tirkkonen, O 2025, Asynchronous Multi-Agent Reinforcement Learning for Scheduling in Subnetworks. in Proceedings of the IEEE 101st Vehicular Technology Conference . IEEE Vehicular Technology Conference, IEEE, IEEE Vehicular Technology Conference, Oslo, Norway, 17/06/2025. https://doi.org/10.1109/VTC2025-Spring65109.2025.11174624