A Method for Solving Reconfiguration Blueprints Based on Multi-Agent Reinforcement Learning
Computer Science and Information Systems, Tome 21 (2024) no. 4.

Voir la notice de l'article provenant de la source Computer Science and Information Systems website

Integrated modular avionics systems primarily achieve system fault tolerance by reconfiguring the system configuration blueprints. In the design of manual reconfiguration, the quality of reconfiguration blueprints is influenced by various unstable factors, leading to a certain degree of uncertainty. The effectiveness of reconfiguration blueprints depends on various factors, including load balancing, the impact of reconfiguration, and the time required for the process. Solving high-quality reconfiguration configuration blueprints can be regarded as a type of multi-objective optimization problem. Traditional algorithms have limitations in solving multi-objective optimization problems. Multi-Agent Reinforcement Learning (MARL) is an important branch in the field of machine learning. It enables the accomplishment of more complex tasks in dynamic real-world scenarios through interaction and decision-making. Combining Multi-Agent Reinforcement Learning algorithms with reconfiguration techniques and utilizing MARL methods to generate blueprints can optimize the quality of blueprints in multiple ways. In this paper, an Improved Value-Decomposition Networks (VDN) based on the average sequential cumulative reward is proposed. By refining the characteristics of the integrated modular avionics system, mathematical models are developed for both the system and the reconfiguration blueprint. The Improved VDN algorithm demonstrates superior convergence characteristics and optimization effects compared with traditional reinforcement learning algorithms such as Q-learning, Deep Q-learning Network (DQN), and VDN. This superiority has been confirmed through experiments involving single and continuous faults.
Keywords: Integrated modular avionics system, Multi-Agent Reinforcement Learning, reconfiguration blueprint, multi-objective optimization problem
@article{CSIS_2024_21_4_a9,
     author = {Jing Cheng and Wen Tan and Guangzhe Lv and Guodong Li and Wentao Zhang and Zihao Liu},
     title = {A {Method} for {Solving} {Reconfiguration} {Blueprints} {Based} on {Multi-Agent} {Reinforcement} {Learning}},
     journal = {Computer Science and Information Systems},
     publisher = {mathdoc},
     volume = {21},
     number = {4},
     year = {2024},
     url = {http://geodesic.mathdoc.fr/item/CSIS_2024_21_4_a9/}
}
TY  - JOUR
AU  - Jing Cheng
AU  - Wen Tan
AU  - Guangzhe Lv
AU  - Guodong Li
AU  - Wentao Zhang
AU  - Zihao Liu
TI  - A Method for Solving Reconfiguration Blueprints Based on Multi-Agent Reinforcement Learning
JO  - Computer Science and Information Systems
PY  - 2024
VL  - 21
IS  - 4
PB  - mathdoc
UR  - http://geodesic.mathdoc.fr/item/CSIS_2024_21_4_a9/
ID  - CSIS_2024_21_4_a9
ER  - 
%0 Journal Article
%A Jing Cheng
%A Wen Tan
%A Guangzhe Lv
%A Guodong Li
%A Wentao Zhang
%A Zihao Liu
%T A Method for Solving Reconfiguration Blueprints Based on Multi-Agent Reinforcement Learning
%J Computer Science and Information Systems
%D 2024
%V 21
%N 4
%I mathdoc
%U http://geodesic.mathdoc.fr/item/CSIS_2024_21_4_a9/
%F CSIS_2024_21_4_a9
Jing Cheng; Wen Tan; Guangzhe Lv; Guodong Li; Wentao Zhang; Zihao Liu. A Method for Solving Reconfiguration Blueprints Based on Multi-Agent Reinforcement Learning. Computer Science and Information Systems, Tome 21 (2024) no. 4. http://geodesic.mathdoc.fr/item/CSIS_2024_21_4_a9/