Learning Instance-Level N-Ary Semantic Knowledge At Scale For Robots Operating in Everyday Environments


Weiyu Liu (Georgia Tech),
Dhruva Bansal (Georgia Tech),
Angel Daruna (Georgia Tech),
Sonia Chernova (Georgia Tech)
Paper Website
Code
Paper #035
Interactive Poster Session III Interactive Poster Session VI

0d 00h 00m

0d 00h 00m


Abstract

Robots operating in everyday environments need to effectively perceive, model, and infer semantic properties of objects. Existing knowledge reasoning frameworks only model binary relations between an object’s class label and its semantic properties, unable to collectively reason about object properties detected by different perception algorithms and grounded in diverse sensory modalities. We bridge the gap between multimodal perception and knowledge reasoning by introducing an n-ary representation that models complex, inter-related object properties. To tackle the problem of collecting n-ary semantic knowledge at scale, we propose a transformer neural network that directly generalizes knowledge from observations of object instances. The learned model can reason at different levels of abstraction, effectively predicting unknown properties of objects in different environmental contexts given different amounts of observed information. We quantitatively validate our approach against five prior methods on LINK, a unique dataset we contribute that contains 1457 situated object instances with 15 multimodal properties types and 200 total properties. Compared to the prior state of the art Markov Logic Network, our model obtains a 10% improvement in predicting unknown properties of novel object instances while reducing training and inference time by 150 times. Additionally, we apply our work to a mobile manipulation robot, demonstrating its ability to leverage n-ary reasoning to retrieve objects and actively detect object properties.

Spotlight Presentation

Previous Paper Paper Website Next Paper