Determining Exception Context in Assembly Operations from Multimodal Data

Sensors (Basel). 2022 Oct 19;22(20):7962. doi: 10.3390/s22207962.

Abstract

Robot assembly tasks can fail due to unpredictable errors and can only continue with the manual intervention of a human operator. Recently, we proposed an exception strategy learning framework based on statistical learning and context determination, which can successfully resolve such situations. This paper deals with context determination from multimodal data, which is the key component of our framework. We propose a novel approach to generate unified low-dimensional context descriptions based on image and force-torque data. For this purpose, we combine a state-of-the-art neural network model for image segmentation and contact point estimation using force-torque measurements. An ensemble of decision trees is used to combine features from the two modalities. To validate the proposed approach, we have collected datasets of deliberately induced insertion failures both for the classic peg-in-hole insertion task and for an industrially relevant task of car starter assembly. We demonstrate that the proposed approach generates reliable low-dimensional descriptors, suitable as queries necessary in statistical learning.

Keywords: autonomous assembly; autonomous exception handling; peg-in-hole; predictive clustering trees; sensor fusion.

MeSH terms

  • Humans
  • Image Processing, Computer-Assisted* / methods
  • Neural Networks, Computer*