Effective Search of Logical Forms for Weakly Supervised Knowledge-Based Question Answering

Effective Search of Logical Forms for Weakly Supervised Knowledge-Based Question Answering

Tao Shen, Xiubo Geng, Guodong Long, Jing Jiang, Chengqi Zhang, Daxin Jiang

Proceedings of the Twenty-Ninth International Joint Conference on Artificial Intelligence
Main track. Pages 2227-2233. https://doi.org/10.24963/ijcai.2020/308

Many algorithms for Knowledge-Based Question Answering (KBQA) depend on semantic parsing, which translates a question to its logical form. When only weak supervision is provided, it is usually necessary to search valid logical forms for model training. However, a complex question typically involves a huge search space, which creates two main problems: 1) the solutions limited by computation time and memory usually reduce the success rate of the search, and 2) spurious logical forms in the search results degrade the quality of training data. These two problems lead to a poorly-trained semantic parsing model. In this work, we propose an effective search method for weakly supervised KBQA based on operator prediction for questions. With search space constrained by predicted operators, sufficient search paths can be explored, more valid logical forms can be derived, and operators possibly causing spurious logical forms can be avoided. As a result, a larger proportion of questions in a weakly supervised training set are equipped with logical forms, and fewer spurious logical forms are generated. Such high-quality training data directly contributes to a better semantic parsing model. Experimental results on one of the largest KBQA datasets (i.e., CSQA) verify the effectiveness of our approach and deliver a new state-of-the-art performance.
Keywords:
Machine Learning: Knowledge-based Learning
Data Mining: Mining Graphs, Semi Structured Data, Complex Data
Natural Language Processing: Natural Language Semantics