Through-Ice Acoustic Source Tracking Using Vision Transformers with Ordinal Classification

Sensors (Basel). 2022 Jun 22;22(13):4703. doi: 10.3390/s22134703.

Abstract

Ice environments pose challenges for conventional underwater acoustic localization techniques due to their multipath and non-linear nature. In this paper, we compare different deep learning networks, such as Transformers, Convolutional Neural Networks (CNNs), Long Short-Term Memory (LSTM) networks, and Vision Transformers (ViTs), for passive localization and tracking of single moving, on-ice acoustic sources using two underwater acoustic vector sensors. We incorporate ordinal classification as a localization approach and compare the results with other standard methods. We conduct experiments passively recording the acoustic signature of an anthropogenic source on the ice and analyze these data. The results demonstrate that Vision Transformers are a strong contender for tracking moving acoustic sources on ice. Additionally, we show that classification as a localization technique can outperform regression for networks more suited for classification, such as the CNN and ViTs.

Keywords: Vision Transformers; ice acoustics; localization; ordinal classification.

MeSH terms

  • Acoustics*
  • Ice*
  • Neural Networks, Computer

Substances

  • Ice