Abstract
Underwater acoustic target recognition (UATR) is typically challenging due to the complex underwater environment and poor prior knowledge. Deep learning (DL)-based UATR methods have demonstrated their effectiveness by extracting more discriminative features on time–frequency (T–F) spectrograms. However, the existing methods exhibit the lack of robustness and ability to capture the time–frequency correlation inherent in the T–F representation. To this end, we first introduce the Wavelet Scattering Transform (WST) to obtain the T–F scattering coefficients of underwater acoustic signals. Then, we treat the scattering coefficients as multivariate time-series data and design a new Two-Stream Time–Frequency (newTSTF) transformer. This model can simultaneously extract temporal and frequency-related features from the scattering coefficients, enhancing accuracy. Specifically, we introduce the Non-stationary encoder to recover the temporal features lost during normalization. Experimental results on real-world data demonstrate that our model achieves high accuracy in UATR.
Original language | English |
---|---|
Article number | 109891 |
Journal | Signal Processing |
Volume | 231 |
DOIs | |
State | Published - Jun 2025 |
Keywords
- Transformer
- Underwater acoustic target recognition (UATR)
- Wavelet scattering transform