Multi-Agent Reinforcement Learning for Coordinating Communication and Control

Federico Mason*, Federico Chiariotti, Andrea Zanella, Petar Popovski

*Corresponding author for this work

Research output: Contribution to journalJournal articleResearchpeer-review

2 Citations (Scopus)

Abstract

The automation of factories and manufacturing processes has been accelerating over the past few years, leading to an ever-increasing number of scenarios with networked agents whose coordination requires reliable wireless communication. In this context, goal-oriented communication adapts transmissions to the control task, prioritizing the more relevant information to decide which action to take. Instead, networked control models follow the opposite pathway, optimizing physical actions to address communication impairments. In this work, we propose a joint design that combines goal-oriented communication and networked control into a single optimization model, an extension of a multi-agent Partially Observable Markov Decision Process (POMDP), which we call Cyber-Physical POMDP. The proposed model is flexible enough to represent a large variety of scenarios and we illustrate its potential in two simple use cases with a single agent and a set of supporting sensors. Our results assess that the joint optimization of communication and control tasks radically improves the performance of networked control systems, particularly in the case of constrained resources, leading to implicit coordination of communication actions.

Original languageEnglish
JournalIEEE Transactions on Cognitive Communications and Networking
Volume10
Issue number4
Pages (from-to)1566-1581
Number of pages16
ISSN2332-7731
DOIs
Publication statusPublished - 2024

Bibliographical note

Publisher Copyright:
© 2015 IEEE.

Keywords

  • goal-oriented communications
  • Markov decision processes
  • multi-agent reinforcement learning
  • networked control systems

Fingerprint

Dive into the research topics of 'Multi-Agent Reinforcement Learning for Coordinating Communication and Control'. Together they form a unique fingerprint.

Cite this