@inproceedings{liu-zhang-2017-encoder,
title = "Encoder-Decoder Shift-Reduce Syntactic Parsing",
author = "Liu, Jiangming and
Zhang, Yue",
editor = "Miyao, Yusuke and
Sagae, Kenji",
booktitle = "Proceedings of the 15th International Conference on Parsing Technologies",
month = sep,
year = "2017",
address = "Pisa, Italy",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/W17-6315",
pages = "105--114",
abstract = "Encoder-decoder neural networks have been used for many NLP tasks, such as neural machine translation. They have also been applied to constituent parsing by using bracketed tree structures as a target language, translating input sentences into syntactic trees. A more commonly used method to linearize syntactic trees is the shift-reduce system, which uses a sequence of transition-actions to build trees. We empirically investigate the effectiveness of applying the encoder-decoder network to transition-based parsing. On standard benchmarks, our system gives comparable results to the stack LSTM parser for dependency parsing, and significantly better results compared to the aforementioned parser for constituent parsing, which uses bracketed tree formats.",
}
<?xml version="1.0" encoding="UTF-8"?>
<modsCollection xmlns="http://www.loc.gov/mods/v3">
<mods ID="liu-zhang-2017-encoder">
<titleInfo>
<title>Encoder-Decoder Shift-Reduce Syntactic Parsing</title>
</titleInfo>
<name type="personal">
<namePart type="given">Jiangming</namePart>
<namePart type="family">Liu</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Yue</namePart>
<namePart type="family">Zhang</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<originInfo>
<dateIssued>2017-09</dateIssued>
</originInfo>
<typeOfResource>text</typeOfResource>
<relatedItem type="host">
<titleInfo>
<title>Proceedings of the 15th International Conference on Parsing Technologies</title>
</titleInfo>
<name type="personal">
<namePart type="given">Yusuke</namePart>
<namePart type="family">Miyao</namePart>
<role>
<roleTerm authority="marcrelator" type="text">editor</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Kenji</namePart>
<namePart type="family">Sagae</namePart>
<role>
<roleTerm authority="marcrelator" type="text">editor</roleTerm>
</role>
</name>
<originInfo>
<publisher>Association for Computational Linguistics</publisher>
<place>
<placeTerm type="text">Pisa, Italy</placeTerm>
</place>
</originInfo>
<genre authority="marcgt">conference publication</genre>
</relatedItem>
<abstract>Encoder-decoder neural networks have been used for many NLP tasks, such as neural machine translation. They have also been applied to constituent parsing by using bracketed tree structures as a target language, translating input sentences into syntactic trees. A more commonly used method to linearize syntactic trees is the shift-reduce system, which uses a sequence of transition-actions to build trees. We empirically investigate the effectiveness of applying the encoder-decoder network to transition-based parsing. On standard benchmarks, our system gives comparable results to the stack LSTM parser for dependency parsing, and significantly better results compared to the aforementioned parser for constituent parsing, which uses bracketed tree formats.</abstract>
<identifier type="citekey">liu-zhang-2017-encoder</identifier>
<location>
<url>https://aclanthology.org/W17-6315</url>
</location>
<part>
<date>2017-09</date>
<extent unit="page">
<start>105</start>
<end>114</end>
</extent>
</part>
</mods>
</modsCollection>
%0 Conference Proceedings
%T Encoder-Decoder Shift-Reduce Syntactic Parsing
%A Liu, Jiangming
%A Zhang, Yue
%Y Miyao, Yusuke
%Y Sagae, Kenji
%S Proceedings of the 15th International Conference on Parsing Technologies
%D 2017
%8 September
%I Association for Computational Linguistics
%C Pisa, Italy
%F liu-zhang-2017-encoder
%X Encoder-decoder neural networks have been used for many NLP tasks, such as neural machine translation. They have also been applied to constituent parsing by using bracketed tree structures as a target language, translating input sentences into syntactic trees. A more commonly used method to linearize syntactic trees is the shift-reduce system, which uses a sequence of transition-actions to build trees. We empirically investigate the effectiveness of applying the encoder-decoder network to transition-based parsing. On standard benchmarks, our system gives comparable results to the stack LSTM parser for dependency parsing, and significantly better results compared to the aforementioned parser for constituent parsing, which uses bracketed tree formats.
%U https://aclanthology.org/W17-6315
%P 105-114
Markdown (Informal)
[Encoder-Decoder Shift-Reduce Syntactic Parsing](https://aclanthology.org/W17-6315) (Liu & Zhang, IWPT 2017)
ACL
- Jiangming Liu and Yue Zhang. 2017. Encoder-Decoder Shift-Reduce Syntactic Parsing. In Proceedings of the 15th International Conference on Parsing Technologies, pages 105–114, Pisa, Italy. Association for Computational Linguistics.