Skip to main content

Flow Vision for Autonomous Underwater Vehicles via an Artificial Lateral Line

Abstract

Most fish have the capability of sensing flows and nearby movements even in dark or murky conditions by using the lateral line organs. This enables them to perform a variety of underwater activities, such as localizing prey, avoiding predators, navigating in narrow spaces, and schooling. To emulate this capability for Autonomous Underwater Vehicles, we developed an artificial lateral line using an array of Micro-Electro-Mechanical-Systems (MEMS) flow sensors. The signals collected via the artificial lateral line are then processed by an adaptive beamforming algorithm developed from Capon's method. The system produces 3D images of source locations for different hydrodynamic activities, including the vibration of a dipole source and the movement of a tail-flicking crayfish. A self-calibration algorithm provides the capability of self-adaptation to different environments. Lastly, we give a Cramer-Rao bound on the theoretical performance limit which is consistent with experimental results.

Publisher note

To access the full article, please see PDF.

Author information

Affiliations

Authors

Corresponding author

Correspondence to Nam Nguyen.

Rights and permissions

Open Access This article is distributed under the terms of the Creative Commons Attribution 2.0 International License (https://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

Reprints and Permissions

About this article

Cite this article

Nguyen, N., Jones, D.L., Yang, Y. et al. Flow Vision for Autonomous Underwater Vehicles via an Artificial Lateral Line. EURASIP J. Adv. Signal Process. 2011, 806406 (2011). https://doi.org/10.1155/2011/806406

Download citation

Keywords

  • Source Location
  • Lateral Line
  • Performance Limit
  • Full Article
  • Autonomous Underwater Vehicle