Published January 1, 2021 | Version v1
Conference paper Open

FINO-Net: A Deep Multimodal Sensor Fusion Framework for Manipulation Failure Detection

  • 1. Istanbul Tech Univ, Fac Comp & Informat Engn, Artificial Intelligence & Robot Lab, Maslak, Turkey
  • 2. Halmstad Univ, Ctr Appl Intelligent Syst Res, Sch Informat Technol, Halmstad, Sweden

Description

We need robots more aware of the unintended outcomes of their actions for ensuring safety. This can be achieved by an onboard failure detection system to monitor and detect such cases. Onboard failure detection is challenging with a limited set of onboard sensor setup due to the limitations of sensing capabilities of each sensor. To alleviate these challenges, we propose FINO-Net, a novel multimodal sensor fusion based deep neural network to detect and identify manipulation failures. We also introduce FAILURE, a multimodal dataset, containing 229 real-world manipulation data recorded with a Baxter robot. Our network combines RGB, depth and audio readings to effectively detect failures. Results indicate that fusing RGB with depth and audio modalities significantly improves the performance. FINO-Net achieves %98.60 detection accuracy on our novel dataset. Code and data are publicly available at https://github.com/ardai/fino-net.

Files

bib-1bf2a1b4-42a4-4dc0-8374-bb50c787f293.txt

Files (220 Bytes)

Name Size Download all
md5:79efa3d7c44b6ba467a5b63ce2ef6af5
220 Bytes Preview Download