Timezone: »

Fast Abductive Learning by Similarity-based Consistency Optimization
Yu-Xuan Huang · Wang-Zhou Dai · Le-Wen Cai · Stephen H Muggleton · Yuan Jiang

Wed Dec 08 12:30 AM -- 02:00 AM (PST) @ Virtual

To utilize the raw inputs and symbolic knowledge simultaneously, some recent neuro-symbolic learning methods use abduction, i.e., abductive reasoning, to integrate sub-symbolic perception and logical inference. While the perception model, e.g., a neural network, outputs some facts that are inconsistent with the symbolic background knowledge base, abduction can help revise the incorrect perceived facts by minimizing the inconsistency between them and the background knowledge. However, to enable effective abduction, previous approaches need an initialized perception model that discriminates the input raw instances. This limits the application of these methods, as the discrimination ability is usually acquired from a thorough pre-training when the raw inputs are difficult to classify. In this paper, we propose a novel abduction strategy, which leverages the similarity between samples, rather than the output information by the perceptual neural network, to guide the search in abduction. Based on this principle, we further present ABductive Learning with Similarity (ABLSim) and apply it to some difficult neuro-symbolic learning tasks. Experiments show that the efficiency of ABLSim is significantly higher than the state-of-the-art neuro-symbolic methods, allowing it to achieve better performance with less labeled data and weaker domain knowledge.

Author Information

Yu-Xuan Huang (Nanjing University)
Wang-Zhou Dai (Imperial College London)
Le-Wen Cai (Nanjing University)
Stephen H Muggleton (Imperial College London)
Yuan Jiang (National Key lab for Novel Software Technology)

More from the Same Authors