Leveraging AMR Graph Structure for Better Sequence-to-Sequence AMR Parsing

Linyu Fan, Wu Wu Yiheng, Jun Xie, Junhui Li, Fang Kong, Guodong Zhou


Abstract
Thanks to the development of pre-trained sequence-to-sequence (seq2seq) models (e.g., BART), recent studies on AMR parsing often regard this task as a seq2seq translation problem by linearizing AMR graphs into AMR token sequences in pre-processing and recovering AMR graphs from sequences in post-processing. Seq2seq AMR parsing is a relatively simple paradigm but it unavoidably loses structural information among AMR tokens. To compensate for the loss of structural information, in this paper we explicitly leverage AMR structure in the decoding phase. Given an AMR graph, we first project the structure in the graph into an AMR token graph, i.e., structure among AMR tokens in the linearized sequence. The structures for an AMR token could be divided into two parts: structure in prediction history and structure in future. Then we propose to model structure in prediction history via a graph attention network (GAT) and learn structure in future via a multi-task scheme, respectively. Experimental results show that our approach significantly outperforms a strong baseline and achieves performance with 85.5 ±0.1 and 84.2 ±0.1 Smatch scores on AMR 2.0 and AMR 3.0, respectively
Anthology ID:
2024.lrec-main.903
Volume:
Proceedings of the 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC-COLING 2024)
Month:
May
Year:
2024
Address:
Torino, Italia
Editors:
Nicoletta Calzolari, Min-Yen Kan, Veronique Hoste, Alessandro Lenci, Sakriani Sakti, Nianwen Xue
Venues:
LREC | COLING
SIG:
Publisher:
ELRA and ICCL
Note:
Pages:
10336–10346
Language:
URL:
https://aclanthology.org/2024.lrec-main.903
DOI:
Bibkey:
Cite (ACL):
Linyu Fan, Wu Wu Yiheng, Jun Xie, Junhui Li, Fang Kong, and Guodong Zhou. 2024. Leveraging AMR Graph Structure for Better Sequence-to-Sequence AMR Parsing. In Proceedings of the 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC-COLING 2024), pages 10336–10346, Torino, Italia. ELRA and ICCL.
Cite (Informal):
Leveraging AMR Graph Structure for Better Sequence-to-Sequence AMR Parsing (Fan et al., LREC-COLING 2024)
Copy Citation:
PDF:
https://aclanthology.org/2024.lrec-main.903.pdf