From 90eeda900ab4c7a059c47643d3e2d6d37e8b9d1f Mon Sep 17 00:00:00 2001 From: Nathan Schneider Date: Sat, 12 Oct 2024 08:21:20 -0400 Subject: [PATCH] 2024.textgraphs-1.3: title, abstract spaces (#3950) --- data/xml/2024.textgraphs.xml | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/data/xml/2024.textgraphs.xml b/data/xml/2024.textgraphs.xml index d24ee99797..7b7b8f9dd9 100644 --- a/data/xml/2024.textgraphs.xml +++ b/data/xml/2024.textgraphs.xml @@ -51,11 +51,11 @@ brannon-etal-2024-congrat - A Pipeline Approach for Parsing Documents into Uniform Meaning Representation Graphs + Uniform Meaning Representation Parsing as a Pipelined Approach JayeolChunBrandeis University NianwenXueBrandeis University 40-52 - Uniform Meaning Representation (UMR) is the next phase of semantic formalism following Abstract Meaning Representation (AMR), with added focus on inter-sentential relations allowing the representational scope of UMR to cover a full document.This, in turn, greatly increases the complexity of its parsing task with the additional requirement of capturing document-level linguistic phenomena such as coreference, modal and temporal dependencies.In order to establish a strong baseline despite the small size of recently released UMR v1.0 corpus, we introduce a pipeline model that does not require any training.At the core of our method is a two-track strategy of obtaining UMR’s sentence and document graphs separately, with the document-level triples being compiled at the token level and the sentence graph being converted from AMR graphs.By leveraging alignment between AMR and its sentence, we are able to generate the first automatic English UMR parses. + Uniform Meaning Representation (UMR) is the next phase of semantic formalism following Abstract Meaning Representation (AMR), with added focus on inter-sentential relations allowing the representational scope of UMR to cover a full document. This, in turn, greatly increases the complexity of its parsing task with the additional requirement of capturing document-level linguistic phenomena such as coreference, modal and temporal dependencies. In order to establish a strong baseline despite the small size of recently released UMR v1.0 corpus, we introduce a pipeline model that does not require any training. At the core of our method is a two-track strategy of obtaining UMR’s sentence and document graphs separately, with the document-level triples being compiled at the token level and the sentence graph being converted from AMR graphs. By leveraging alignment between AMR and its sentence, we are able to generate the first automatic English UMR parses. 2024.textgraphs-1.3 chun-xue-2024-pipeline