Loading [a11y]/accessibility-menu.js
Live demonstration: 3D sonar sensing using low-cost MEMS arrays | IEEE Conference Publication | IEEE Xplore

Live demonstration: 3D sonar sensing using low-cost MEMS arrays


Abstract:

In recent years, our research group has developed state of the art 3D sonar sensors [1]-[3] which use a low-cost MEMS microphone array for real-time acoustic imaging in a...Show More

Abstract:

In recent years, our research group has developed state of the art 3D sonar sensors [1]-[3] which use a low-cost MEMS microphone array for real-time acoustic imaging in air. Using this sensor, various robotic applications have been developed [2], including obstacle avoidance and corridor following and SLAM [1]. The developed sensor is capable of localizing an arbitrary number of reflectors, and generates 2D (range versus azimuth) or 3D (range versus azimuth and elevation) acoustic images of the environment by emitting a broadband, spatially omnidirectional acoustic emission. This emission is reflected back by the environment to the microphone array. Using array beamforming algorithms an acoustic image of the environment is created, which subsequently can be either visualized or used in various control algorithms. The overview of the developed sensor array can be seen in figure 1.
Published in: 2017 IEEE SENSORS
Date of Conference: 29 October 2017 - 01 November 2017
Date Added to IEEE Xplore: 25 December 2017
ISBN Information:
Conference Location: Glasgow, UK
Citations are not available for this document.

I. Introduction

In recent years, our research group has developed state of the art 3D sonar sensors [1]–[3] which use a low-cost MEMS microphone array for real-time acoustic imaging in air. Using this sensor, various robotic applications have been developed [2], including obstacle avoidance and corridor following and SLAM [1]. The developed sensor is capable of localizing an arbitrary number of reflectors, and generates 2D (range versus azimuth) or 3D (range versus azimuth and elevation) acoustic images of the environment by emitting a broadband, spatially omnidirectional acoustic emission. This emission is reflected back by the environment to the microphone array. Using array beamforming algorithms an acoustic image of the environment is created, which subsequently can be either visualized or used in various control algorithms. The overview of the developed sensor array can be seen in figure 1.

Cites in Papers - |

Cites in Papers - IEEE (2)

Select All
1.
Dennis Laurijssen, Anthony Schenck, Girmi Schouten, Robin Kerstens, Sebastiaan Aussems, Eric Paillet, Randy Gomez, Keisuke Nakamura, Walter Daems, Jan Steckel, "Bio-Inspired Gesture Recognition with Baffled Transducers Using Temporal and Spectral Features", 2023 IEEE SENSORS, pp.1-4, 2023.
2.
Dennis Laurijssen, Wouter Jansen, Jan Steckel, "Live Demonstration: Mixed Reality 3D In-Air Ultrasound Applications", 2022 IEEE Sensors, pp.1-1, 2022.

References

References is not available for this document.