STRFormer: A Spatial Topological Relationship-guided Multi-modal Variational Fusion Network for Intelligent Health State Diagnosis of the Manipulator

Bo Zhao, Qiqiang Wu, Yongbo Li, Weihua Li, Xianmin Zhang*, Zijun Zhang*

*Corresponding author for this work

Research output: Journal Publications and ReviewsRGC 21 - Publication in refereed journalpeer-review

1 Citation (Scopus)

Abstract

The precision and reliability of advanced precision machinery equipment hinge critically on the health state of their embedded precise manipulators. In real-world industrial scenarios, the manipulators’ intricate coupled topological structures and the complex operating environment invariably introduce diverse interference and uncertainty factors into the monitored performance data. This significantly complicates high-quality feature extraction and effective information fusion. To address this issue, a novel spatial topological relationship-guided multi-modal variational fusion network, termed STRFormer, is developed and employed for the intelligent health state diagnosis task of manipulators. Specifically, the intrinsic properties characterizing the health state of the manipulator embedded in each independent modal information are initially extracted and strengthened through the constructed integrated scale-wise self-attention (ISSA) module, which takes into account both intra- and inter-scale joint information. Subsequently, from the perspective of the manipulator’s kinematic chain branches, the modal information affiliated with the corresponding kinematic chain branches is further fused through the established hybrid cross-modal variational fusion (HCVF) module, in which the variational inference mechanism is employed to suppress the interference of uncertainty factors. Furthermore, the spatial topological relationships between the individual kinematic chain branches and the manipulator are abstracted into a graph structure, and the ultimate system-level information fusion and task matching are achieved based on this. Through a typical 3-PRR planar parallel manipulator in multi-scenario cases, the feasibility, superiority, robustness, and anti-interference capability of the proposed fusion method are comprehensively verified and discussed. © 2025 Elsevier Ltd.
Original languageEnglish
Article number103499
JournalAdvanced Engineering Informatics
Volume67
Online published9 Jun 2025
DOIs
Publication statusPublished - Sept 2025

Funding

This work was supported in part by the Joint Funds of the National Natural Science Foundation of China (Key Program) under grant U24A20108, in part by the National Natural Science Foundation of China under grant 52130508, in part by the Hong Kong RGC General Research Fund Project under grant 11213124, in part by the Hong Kong RGC Collaborative Research Fund Project under grant C1049-24GF, in part by the Shenzhen-Hong Kong-Macau Science and Technology Category C Project under grant SGDX20220530111205037, in part by the Guangdong Province Technological Project under grant 2023A0505030014, in part by the Hong Kong ITC Innovation and Technology Fund Project under grant ITS/034/22MS, and in part by the Guangdong Provincial Basic and Applied Basic Research-Offshore Wind Power Joint Fund Project under grant 2022A1515240066.

Research Keywords

  • Transformer
  • Multi-modal information fusion
  • Variational inference
  • Health state diagnosis
  • Manipulator

RGC Funding Information

  • RGC-funded

Fingerprint

Dive into the research topics of 'STRFormer: A Spatial Topological Relationship-guided Multi-modal Variational Fusion Network for Intelligent Health State Diagnosis of the Manipulator'. Together they form a unique fingerprint.

Cite this