Mapping Natural-language Problems to Formal-language Solutions Using Structured Neural Representations

作者: Paul Smolensky , Hamid Palangi , Kenneth D. Forbus , Jianfeng Gao , Qiuyuan Huang

DOI:

关键词:

摘要: Generating formal-language programs represented by relational tuples, such as Lisp or mathematical operations, to solve problems stated in natural language is a challenging task because it requires explicitly capturing discrete symbolic structural information implicit the input. However, most general neural sequence models do not capture information, limiting their performance on these tasks. In this paper, we propose new encoder-decoder model based structured representation, Tensor Product Representations (TPRs), for mapping Natural-language Formal-language solutions, called TP-N2F. The encoder of TP-N2F employs TPR `binding' encode natural-language structure vector space and decoder uses `unbinding' generate, space, sequential program each consisting relation (or operation) number arguments. considerably outperforms LSTM-based seq2seq two benchmarks creates state-of-the-art results. Ablation studies show that improvements can be attributed use TPRs both decoder. Analysis learned structures shows how enhance interpretability

参考文章(21)
Susan Goldin-Meadow, Dedre Gentner, Language in Mind: Advances in the Study of Language and Thought MIT Press. ,(2003)
Diederik P. Kingma, Jimmy Ba, Adam: A Method for Stochastic Optimization arXiv: Learning. ,(2014)
Pat Langley, Crafting Papers on Machine Learning international conference on machine learning. pp. 1207- 1216 ,(2000)
Thang Luong, Hieu Pham, Christopher D. Manning, Effective Approaches to Attention-based Neural Machine Translation empirical methods in natural language processing. pp. 1412- 1421 ,(2015) , 10.18653/V1/D15-1166
Paul Smolensky, Li Deng, Jianfeng Gao, Wen-tau Yih, Xiaodong He, Moontae Lee, Basic Reasoning with Tensor Product Representations. arXiv: Artificial Intelligence. ,(2016)
Kenneth D. Forbus, Chen Liang, Irina Rabkina, Representation and Computation in Cognitive Models Topics in Cognitive Science. ,vol. 9, pp. 694- 718 ,(2017) , 10.1111/TOPS.12277
Kenneth D Forbus, Kezhen Chen, Action recognition from skeleton data via analogical generalization over qualitative representations national conference on artificial intelligence. pp. 638- 645 ,(2018)
Jürgen Schmidhuber, Imanol Schlag, Learning to Reason with Third Order Tensor Products neural information processing systems. ,vol. 31, pp. 9981- 9993 ,(2018)
Krzysztof Krawiec, Karol Piaskowski, Jakub Bednarek, Ain't Nobody Got Time for Coding: Structure-Aware Program Synthesis from Natural Language arXiv: Learning. ,(2018)
Qiuyuan Huang, Li Deng, Dapeng Wu, Chang Liu, Xiaodong He, Attentive Tensor Product Learning national conference on artificial intelligence. ,vol. 33, pp. 1344- 1351 ,(2019) , 10.1609/AAAI.V33I01.33011344