Indexed by:
Abstract:
Fault diagnosis based on single-modal features struggles to capture the coupling relationship between multiple fault factors, resulting in inferior diagnosis accuracy. To address this problem, a transfer learning-based multimodal feature fusion model (TL-MMFF) is proposed for fault diagnosis. First, a continuous wavelet transform-based modal expression method is employed to transform raw vibration signals into time-frequency representations. Then, this high-resolution time-frequency modal can be utilized to capture transient vibration and energy changes in non-stationary signals. Second, a multi-modal feature fusion strategy is proposed, which designs learnable parameters to dynamically weight the time-domain features of torque and the time-frequency features of vibration signals. This adaptive weighting strategy optimizes the fusion process based on the correlation of different modal feature sets, thereby enhancing the ability to describe fault characteristics. Third, a maximum mean discrepancy-based transfer learning algorithm is designed to reduce the distribution differences between fused features under different operating conditions. Then, the model can identify fault characteristics across varying operating conditions. Finally, experiments on the Paderborn University dataset demonstrate that TL-MMFF achieves 99.1% accuracy and converges 30% faster than single-modal methods. These results validate the effectiveness of the model in integrating multimodal data and generalizing across domains. © 1963-2012 IEEE.
Keyword:
Reprint Author's Address:
Email:
Source :
IEEE Transactions on Instrumentation and Measurement
ISSN: 0018-9456
Year: 2025
5 . 6 0 0
JCR@2022
Cited Count:
SCOPUS Cited Count:
ESI Highly Cited Papers on the List: 0 Unfold All
WanFang Cited Count:
Chinese Cited Count:
30 Days PV: 9
Affiliated Colleges: