A copy of this work was available on the public web and has been preserved in the Wayback Machine. The capture dates from 2020; you can also visit the original URL.
The file type is application/pdf
.
Exploiting Rich Syntactic Information for Semantic Parsing with Graph-to-Sequence Model
2018
Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing
Existing neural semantic parsers mainly utilize a sequence encoder, i.e., a sequential LSTM, to extract word order features while neglecting other valuable syntactic information such as dependency or constituent trees. In this paper, we first propose to use the syntactic graph to represent three types of syntactic information, i.e., word order, dependency and constituency features; then employ a graph-tosequence model to encode the syntactic graph and decode a logical form. Experimental results
doi:10.18653/v1/d18-1110
dblp:conf/emnlp/XuWWYCS18
fatcat:ztx75adlwvdwrfzib6p67yclru