Timezone: »
Given a natural language instruction, and an input and an output scene, our goal is to train a neuro-symbolic model which can output a manipulation program that can be executed by the robot on the input scene resulting in the desired output scene. Prior approaches for this task possess one of the following limitations: (i) rely on hand-coded symbols for concepts limiting generalization beyond those seen during training (R. Paul et. al., 2016) (ii) infer action sequences from instructions but require dense sub-goal supervision (C. Paxton et. al., 2019) or (iii) lack semantics required for deeper object- centric reasoning inherent in interpreting complex instructions (M. Shridhar et. al., 2022). In contrast, our approach is neuro-symbolic and can handle linguistic as well as perceptual variations, is end-to-end differentiable requiring no intermediate supervision, and makes use of symbolic reasoning constructs which operate on a latent neural object- centric representation, allowing for deeper reasoning over the input scene. Our experiments on a simulated environment with a 7-DOF manipulator, consisting of instructions with varying number of steps, as well as scenes with different number of objects, and objects with unseen attribute combinations, demonstrate that our model is robust to such variations, and significantly outperforms existing baselines, particularly in generalization settings.
Author Information
Namasivayam Kalithasan (Indian Institute of Technology, Delhi)
Himanshu Singh (IIT Delhi)
Vishal Bindal (Indian Institute of Technology Delhi)
Arnav Tuli (Indian Institute of Technology, Delhi)
Vishwajeet Agrawal
Rahul Jain
Parag Singla (Indian Institute of Technology Delhi)
Rohan Paul (Indian Institute of Technology Delhi)
More from the Same Authors
-
2022 : Image Manipulation via Neuro-Symbolic Networks »
Harman Singh · Poorva Garg · Mohit Gupta · Kevin Shah · Arnab Kumar Mondal · Dinesh Khandelwal · Parag Singla · Dinesh Garg -
2022 : Few Shot Generative Domain Adaptation Via Inference-Stage Latent Learning in GANs »
Arnab Kumar Mondal · Piyush Tiwary · Parag Singla · Prathosh AP -
2022 Poster: A Solver-free Framework for Scalable Learning in Neural ILP Architectures »
Yatin Nandwani · Rishabh Ranjan · - Mausam · Parag Singla -
2021 : Panel Discussion 2 »
Susan L Epstein · Yoshua Bengio · Lucina Uddin · Rohan Paul · Steve Fleming -
2021 Workshop: Metacognition in the Age of AI: Challenges and Opportunities »
Ingmar Posner · Francesca Rossi · Lior Horesh · Steve Fleming · Oiwi Parker Jones · Rohan Paul · Biplav Srivastava · Andrea Loreggia · Marianna Ganapini -
2019 Poster: A Primal Dual Formulation For Deep Learning With Constraints »
Yatin Nandwani · Abhishek Pathak · Mausam · Parag Singla -
2018 : Spotlights 2 »
Mausam · Ankit Anand · Parag Singla · Tarik Koc · Tim Klinger · Habibeh Naderi · Sungwon Lyu · Saeed Amizadeh · Kshitij Dwivedi · Songpeng Zu · Wei Feng · Balaraman Ravindran · Edouard Pineau · Abdulkadir Celikkanat · Deepak Venugopal -
2015 Poster: Fast Lifted MAP Inference via Partitioning »
Somdeb Sarkhel · Parag Singla · Vibhav Gogate -
2015 Poster: Lifted Inference Rules With Constraints »
Happy Mittal · Anuj Mahajan · Vibhav Gogate · Parag Singla -
2015 Poster: Lifted Symmetry Detection and Breaking for MAP Inference »
Timothy Kopp · Parag Singla · Henry Kautz -
2014 Poster: An Integer Polynomial Programming Based Framework for Lifted MAP Inference »
Somdeb Sarkhel · Deepak Venugopal · Parag Singla · Vibhav Gogate -
2014 Poster: New Rules for Domain Independent Lifted MAP Inference »
Happy Mittal · Prasoon Goyal · Vibhav Gogate · Parag Singla