信息物理融合视角下的多机器人系统安全控制与学习综述

Survey on Safety and Security Control and Learning in Multi-robot Systems from a Cyber-physical System Perspective

  • 摘要: 多机器人系统在智能制造、灾难救援与无人作战等场景中展现出显著的协同优势,但随着应用环境从结构化空间走向开放、对抗与高动态场景,系统安全性成为制约其大规模部署的核心瓶颈。基于威胁来源与作用机理,可将多机器人系统的安全问题概括为3个层面:物理空间的避碰安全、系统功能的容错安全及信息空间的网络弹性安全。围绕上述维度,本文从信息物理融合的统一视角系统梳理了多机器人安全控制与学习的发展脉络,涵盖从启发式方法到形式化控制、从鲁棒容错到分布式弹性控制、从模型驱动到学习增强的技术演进。本文重点综述了安全避碰控制(以控制障碍函数为代表的形式化方法及其数据驱动增强)、容错控制(故障检测诊断、分布式协同容错、系统重构)、弹性控制(攻击检测辨识、安全状态估计、弹性恢复策略)及安全强化学习(安全屏蔽、约束优化、信任机制) 4大关键技术方向的研究现状与最新进展。通过对各方向的理论基础、代表性进展及其适用边界进行系统分析,本文揭示当前研究在安全-性能权衡、多源不确定性处理、跨层协同设计与可扩展性方面的核心瓶颈。最后,本文讨论了学习与形式化保证的深度融合、弹性架构及具备自解释性的安全决策等未来发展方向,为多机器人系统的可靠自治提供理论参考与技术指引。

     

    Abstract: Multi-robot systems demonstrate significant collaborative advantages in scenarios such as intelligent manufacturing, disaster rescue, and unmanned combat. However, as application environments shift from structured spaces to open, adversarial, and highly dynamic settings, system security has become a core bottleneck constraining their large-scale deployment. Based on threat sources and mechanisms, security challenges in multi-robot systems can be categorized into three dimensions: collision avoidance safety in physical space, fault-tolerant safety for system functionality, and cyber resilience in information space. From a unified cyber-physical integration perspective, we systematically review the evolution of safety control in multi-robot systems, covering technological progress from heuristic methods to formal control, from robust fault tolerance to distributed resilient control, and from model-driven to learning-enhanced approaches. The study focuses on key technical areas including safety-critical collision avoidance control, exemplified by formal methods such as control barrier functions and their data-driven enhancements; fault-tolerant control, covering fault detection and diagnosis, distributed cooperative fault tolerance, and system reconfiguration; resilient control, involving attack detection and identification, secure state estimation, and resilience recovery strategies; and safe reinforcement learning, which encompasses safety shielding, constrained optimization, and trust mechanisms. Through a systematic analysis of the theoretical foundations, representative advances, and applicability boundaries of these directions, the paper identifies core bottlenecks in current research, including safety–performance trade-offs, handling multi-source uncertainties, cross-layer cooperative design, and scalability. Finally, we discuss future research directions, such as the deep integration of learning with formal guarantees, cross-domain resilient architectures, and self-explainable safety decision-making, providing theoretical references and technical guidance for reliable autonomy in multi-robot systems.

     

/

返回文章
返回